KeywordTokenizer interface
Emits the entire input as a single token. This tokenizer is implemented using Apache Lucene.
Properties
max |
The maximum token length. Default is 256. Tokens longer than the maximum length are split. The maximum token length that can be used is 300 characters. Default value: 256. |
name | The name of the tokenizer. It must only contain letters, digits, spaces, dashes or underscores, can only start and end with alphanumeric characters, and is limited to 128 characters. |
odatatype | Polymorphic Discriminator |
Property Details
maxTokenLength
The maximum token length. Default is 256. Tokens longer than the maximum length are split. The maximum token length that can be used is 300 characters. Default value: 256.
maxTokenLength?: number
Property Value
number
name
The name of the tokenizer. It must only contain letters, digits, spaces, dashes or underscores, can only start and end with alphanumeric characters, and is limited to 128 characters.
name: string
Property Value
string
odatatype
Polymorphic Discriminator
odatatype: "#Microsoft.Azure.Search.KeywordTokenizer" | "#Microsoft.Azure.Search.KeywordTokenizerV2"
Property Value
"#Microsoft.Azure.Search.KeywordTokenizer" | "#Microsoft.Azure.Search.KeywordTokenizerV2"