Skip to content

The Whitespace tokenizer should support the max_token_length parameter #26643

@cbuescher

Description

@cbuescher

Other tokenizers (like Standard) support overriding the max_token_length parameter, but it seems Whitespace doesn't, while the underlying Lucene WhitespaceTokenizer seems to support this parameter. We should probably enable setting this parameter in Elasticsearch as well.

Metadata

Metadata

Assignees

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions