models Package
Classes
AnalyzeResult |
The result of testing an analyzer on text. All required parameters must be populated in order to send to server. |
AnalyzeTextOptions |
Specifies some text and analysis components used to break that text into tokens. All required parameters must be populated in order to send to Azure. |
AnalyzedTokenInfo |
Information about a token returned by an analyzer. Variables are only populated by the server, and will be ignored when sending a request. All required parameters must be populated in order to send to server. |
AsciiFoldingTokenFilter |
Converts alphabetic, numeric, and symbolic Unicode characters which are not in the first 127 ASCII characters (the "Basic Latin" Unicode block) into their ASCII equivalents, if such equivalents exist. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
AzureOpenAIEmbeddingSkill |
Allows you to generate a vector embedding for a given text input using the Azure OpenAI resource. All required parameters must be populated in order to send to server. |
AzureOpenAIVectorizer |
Specifies the Azure OpenAI resource used to vectorize a query string. All required parameters must be populated in order to send to server. |
AzureOpenAIVectorizerParameters |
Specifies the parameters for connecting to the Azure OpenAI resource. |
BM25SimilarityAlgorithm |
Ranking function based on the Okapi BM25 similarity algorithm. BM25 is a TF-IDF-like algorithm that includes length normalization (controlled by the 'b' parameter) as well as term frequency saturation (controlled by the 'k1' parameter). All required parameters must be populated in order to send to server. |
BinaryQuantizationCompression |
Contains configuration options specific to the binary quantization compression method used during indexing and querying. All required parameters must be populated in order to send to server. |
CharFilter |
Base type for character filters. You probably want to use the sub-classes and not this class directly. Known sub-classes are: MappingCharFilter, PatternReplaceCharFilter All required parameters must be populated in order to send to server. |
CjkBigramTokenFilter |
Forms bigrams of CJK terms that are generated from the standard tokenizer. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
ClassicSimilarityAlgorithm |
Legacy similarity algorithm which uses the Lucene TFIDFSimilarity implementation of TF-IDF. This variation of TF-IDF introduces static document length normalization as well as coordinating factors that penalize documents that only partially match the searched queries. All required parameters must be populated in order to send to server. |
ClassicTokenizer |
Grammar-based tokenizer that is suitable for processing most European-language documents. This tokenizer is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
CognitiveServicesAccount |
Base type for describing any Azure AI service resource attached to a skillset. You probably want to use the sub-classes and not this class directly. Known sub-classes are: CognitiveServicesAccountKey, DefaultCognitiveServicesAccount All required parameters must be populated in order to send to server. |
CognitiveServicesAccountKey |
The multi-region account key of an Azure AI service resource that's attached to a skillset. All required parameters must be populated in order to send to server. |
CommonGramTokenFilter |
Construct bigrams for frequently occurring terms while indexing. Single terms are still indexed too, with bigrams overlaid. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
ConditionalSkill |
A skill that enables scenarios that require a Boolean operation to determine the data to assign to an output. All required parameters must be populated in order to send to server. |
CorsOptions |
Defines options to control Cross-Origin Resource Sharing (CORS) for an index. All required parameters must be populated in order to send to server. |
CustomAnalyzer |
Allows you to take control over the process of converting text into indexable/searchable tokens. It's a user-defined configuration consisting of a single predefined tokenizer and one or more filters. The tokenizer is responsible for breaking text into tokens, and the filters for modifying tokens emitted by the tokenizer. All required parameters must be populated in order to send to Azure. |
CustomEntity |
An object that contains information about the matches that were found, and related metadata. All required parameters must be populated in order to send to server. |
CustomEntityAlias |
A complex object that can be used to specify alternative spellings or synonyms to the root entity name. All required parameters must be populated in order to send to server. |
CustomEntityLookupSkill |
A skill looks for text from a custom, user-defined list of words and phrases. All required parameters must be populated in order to send to server. |
DataChangeDetectionPolicy |
Base type for data change detection policies. You probably want to use the sub-classes and not this class directly. Known sub-classes are: HighWaterMarkChangeDetectionPolicy, SqlIntegratedChangeTrackingPolicy All required parameters must be populated in order to send to server. |
DataDeletionDetectionPolicy |
Base type for data deletion detection policies. You probably want to use the sub-classes and not this class directly. Known sub-classes are: SoftDeleteColumnDeletionDetectionPolicy All required parameters must be populated in order to send to server. |
DefaultCognitiveServicesAccount |
An empty object that represents the default Azure AI service resource for a skillset. All required parameters must be populated in order to send to server. |
DictionaryDecompounderTokenFilter |
Decomposes compound words found in many Germanic languages. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
DistanceScoringFunction |
Defines a function that boosts scores based on distance from a geographic location. All required parameters must be populated in order to send to server. |
DistanceScoringParameters |
Provides parameter values to a distance scoring function. All required parameters must be populated in order to send to server. |
DocumentExtractionSkill |
A skill that extracts content from a file within the enrichment pipeline. All required parameters must be populated in order to send to server. |
EdgeNGramTokenFilter |
Generates n-grams of the given size(s) starting from the front or the back of an input token. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
EdgeNGramTokenizer |
Tokenizes the input from an edge into n-grams of the given size(s). This tokenizer is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
ElisionTokenFilter |
Removes elisions. For example, "l'avion" (the plane) will be converted to "avion" (plane). This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
EntityLinkingSkill |
Using the Text Analytics API, extracts linked entities from text. All required parameters must be populated in order to send to server. |
EntityRecognitionSkill |
Using the Text Analytics API, extracts entities of different types from text. All required parameters must be populated in order to send to Azure. |
ExhaustiveKnnAlgorithmConfiguration |
Contains configuration options specific to the exhaustive KNN algorithm used during querying, which will perform brute-force search across the entire vector index. All required parameters must be populated in order to send to server. |
ExhaustiveKnnParameters |
Contains the parameters specific to exhaustive KNN algorithm. |
FieldMapping |
Defines a mapping between a field in a data source and a target field in an index. All required parameters must be populated in order to send to server. |
FieldMappingFunction |
Represents a function that transforms a value from a data source before indexing. All required parameters must be populated in order to send to server. |
FreshnessScoringFunction |
Defines a function that boosts scores based on the value of a date-time field. All required parameters must be populated in order to send to server. |
FreshnessScoringParameters |
Provides parameter values to a freshness scoring function. All required parameters must be populated in order to send to server. |
GetIndexStatisticsResult |
Statistics for a given index. Statistics are collected periodically and are not guaranteed to always be up-to-date. Variables are only populated by the server, and will be ignored when sending a request. All required parameters must be populated in order to send to server. |
HighWaterMarkChangeDetectionPolicy |
Defines a data change detection policy that captures changes based on the value of a high water mark column. All required parameters must be populated in order to send to server. |
HnswAlgorithmConfiguration |
Contains configuration options specific to the HNSW approximate nearest neighbors algorithm used during indexing and querying. The HNSW algorithm offers a tunable trade-off between search speed and accuracy. All required parameters must be populated in order to send to server. |
HnswParameters |
Contains the parameters specific to the HNSW algorithm. |
ImageAnalysisSkill |
A skill that analyzes image files. It extracts a rich set of visual features based on the image content. All required parameters must be populated in order to send to server. |
IndexerExecutionResult |
Represents the result of an individual indexer execution. Variables are only populated by the server, and will be ignored when sending a request. All required parameters must be populated in order to send to server. |
IndexingParameters |
Represents parameters for indexer execution. |
IndexingParametersConfiguration |
A dictionary of indexer-specific configuration properties. Each name is the name of a specific property. Each value must be of a primitive type. |
IndexingSchedule |
Represents a schedule for indexer execution. All required parameters must be populated in order to send to server. |
InputFieldMappingEntry |
Input field mapping for a skill. All required parameters must be populated in order to send to server. |
KeepTokenFilter |
A token filter that only keeps tokens with text contained in a specified list of words. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
KeyPhraseExtractionSkill |
A skill that uses text analytics for key phrase extraction. All required parameters must be populated in order to send to server. |
KeywordMarkerTokenFilter |
Marks terms as keywords. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
KeywordTokenizer | |
LanguageDetectionSkill |
A skill that detects the language of input text and reports a single language code for every document submitted on the request. The language code is paired with a score indicating the confidence of the analysis. All required parameters must be populated in order to send to server. |
LengthTokenFilter |
Removes words that are too long or too short. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
LexicalAnalyzer |
Base type for analyzers. You probably want to use the sub-classes and not this class directly. Known sub-classes are: CustomAnalyzer, PatternAnalyzer, LuceneStandardAnalyzer, StopAnalyzer All required parameters must be populated in order to send to server. |
LexicalTokenizer |
Base type for tokenizers. You probably want to use the sub-classes and not this class directly. Known sub-classes are: ClassicTokenizer, EdgeNGramTokenizer, KeywordTokenizer, KeywordTokenizerV2, MicrosoftLanguageStemmingTokenizer, MicrosoftLanguageTokenizer, NGramTokenizer, PathHierarchyTokenizerV2, PatternTokenizer, LuceneStandardTokenizer, LuceneStandardTokenizerV2, UaxUrlEmailTokenizer All required parameters must be populated in order to send to server. |
LimitTokenFilter |
Limits the number of tokens while indexing. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
LuceneStandardAnalyzer |
Standard Apache Lucene analyzer; Composed of the standard tokenizer, lowercase filter and stop filter. All required parameters must be populated in order to send to server. |
LuceneStandardTokenizer |
Breaks text following the Unicode Text Segmentation rules. This tokenizer is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
MagnitudeScoringFunction |
Defines a function that boosts scores based on the magnitude of a numeric field. All required parameters must be populated in order to send to server. |
MagnitudeScoringParameters |
Provides parameter values to a magnitude scoring function. All required parameters must be populated in order to send to server. |
MappingCharFilter |
A character filter that applies mappings defined with the mappings option. Matching is greedy (longest pattern matching at a given point wins). Replacement is allowed to be the empty string. This character filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
MergeSkill |
A skill for merging two or more strings into a single unified string, with an optional user-defined delimiter separating each component part. All required parameters must be populated in order to send to server. |
MicrosoftLanguageStemmingTokenizer |
Divides text using language-specific rules and reduces words to their base forms. All required parameters must be populated in order to send to server. |
MicrosoftLanguageTokenizer |
Divides text using language-specific rules. All required parameters must be populated in order to send to server. |
NGramTokenFilter |
Generates n-grams of the given size(s). This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
NGramTokenizer |
Tokenizes the input into n-grams of the given size(s). This tokenizer is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
OcrSkill |
A skill that extracts text from image files. All required parameters must be populated in order to send to server. |
OutputFieldMappingEntry |
Output field mapping for a skill. All required parameters must be populated in order to send to server. |
PIIDetectionSkill |
Using the Text Analytics API, extracts personal information from an input text and gives you the option of masking it. All required parameters must be populated in order to send to server. |
PathHierarchyTokenizer | |
PatternAnalyzer |
Flexibly separates text into terms via a regular expression. This analyzer is implemented using Apache Lucene. All required parameters must be populated in order to send to Azure. |
PatternCaptureTokenFilter |
Uses Java regexes to emit multiple tokens - one for each capture group in one or more patterns. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
PatternReplaceCharFilter |
A character filter that replaces characters in the input string. It uses a regular expression to identify character sequences to preserve and a replacement pattern to identify characters to replace. For example, given the input text "aa bb aa bb", pattern "(aa)s+(bb)", and replacement "$1#$2", the result would be "aa#bb aa#bb". This character filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
PatternReplaceTokenFilter |
A character filter that replaces characters in the input string. It uses a regular expression to identify character sequences to preserve and a replacement pattern to identify characters to replace. For example, given the input text "aa bb aa bb", pattern "(aa)s+(bb)", and replacement "$1#$2", the result would be "aa#bb aa#bb". This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
PatternTokenizer |
Tokenizer that uses regex pattern matching to construct distinct tokens. This tokenizer is implemented using Apache Lucene. All required parameters must be populated in order to send to Azure. |
PhoneticTokenFilter |
Create tokens for phonetic matches. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
ScalarQuantizationCompression |
Contains configuration options specific to the scalar quantization compression method used during indexing and querying. All required parameters must be populated in order to send to server. |
ScalarQuantizationParameters |
Contains the parameters specific to Scalar Quantization. |
ScoringFunction |
Base type for functions that can modify document scores during ranking. You probably want to use the sub-classes and not this class directly. Known sub-classes are: DistanceScoringFunction, FreshnessScoringFunction, MagnitudeScoringFunction, TagScoringFunction All required parameters must be populated in order to send to server. |
ScoringProfile |
Defines parameters for a search index that influence scoring in search queries. All required parameters must be populated in order to send to server. |
SearchField |
Represents a field in an index definition, which describes the name, data type, and search behavior of a field. All required parameters must be populated in order to send to Azure. |
SearchIndex |
Represents a search index definition, which describes the fields and search behavior of an index. All required parameters must be populated in order to send to Azure. |
SearchIndexer |
Represents an indexer. All required parameters must be populated in order to send to server. |
SearchIndexerDataContainer |
Represents information about the entity (such as Azure SQL table or CosmosDB collection) that will be indexed. All required parameters must be populated in order to send to server. |
SearchIndexerDataIdentity |
Abstract base type for data identities. You probably want to use the sub-classes and not this class directly. Known sub-classes are: SearchIndexerDataNoneIdentity, SearchIndexerDataUserAssignedIdentity All required parameters must be populated in order to send to server. |
SearchIndexerDataNoneIdentity |
Clears the identity property of a datasource. All required parameters must be populated in order to send to server. |
SearchIndexerDataSourceConnection |
Represents a datasource connection definition, which can be used to configure an indexer. All required parameters must be populated in order to send to Azure. |
SearchIndexerDataUserAssignedIdentity |
Specifies the identity for a datasource to use. All required parameters must be populated in order to send to server. |
SearchIndexerError |
Represents an item- or document-level indexing error. Variables are only populated by the server, and will be ignored when sending a request. All required parameters must be populated in order to send to server. |
SearchIndexerIndexProjection |
Definition of additional projections to secondary search indexes. All required parameters must be populated in order to send to server. |
SearchIndexerIndexProjectionSelector |
Description for what data to store in the designated search index. All required parameters must be populated in order to send to server. |
SearchIndexerIndexProjectionsParameters |
A dictionary of index projection-specific configuration properties. Each name is the name of a specific property. Each value must be of a primitive type. |
SearchIndexerKnowledgeStore |
Definition of additional projections to azure blob, table, or files, of enriched data. All required parameters must be populated in order to send to server. |
SearchIndexerKnowledgeStoreBlobProjectionSelector |
Abstract class to share properties between concrete selectors. All required parameters must be populated in order to send to server. |
SearchIndexerKnowledgeStoreFileProjectionSelector |
Projection definition for what data to store in Azure Files. All required parameters must be populated in order to send to server. |
SearchIndexerKnowledgeStoreObjectProjectionSelector |
Projection definition for what data to store in Azure Blob. All required parameters must be populated in order to send to server. |
SearchIndexerKnowledgeStoreProjection |
Container object for various projection selectors. |
SearchIndexerKnowledgeStoreProjectionSelector |
Abstract class to share properties between concrete selectors. |
SearchIndexerKnowledgeStoreTableProjectionSelector |
Description for what data to store in Azure Tables. All required parameters must be populated in order to send to server. |
SearchIndexerLimits |
SearchIndexerLimits. Variables are only populated by the server, and will be ignored when sending a request. |
SearchIndexerSkill |
Base type for skills. You probably want to use the sub-classes and not this class directly. Known sub-classes are: WebApiSkill, AzureOpenAIEmbeddingSkill, CustomEntityLookupSkill, EntityRecognitionSkill, KeyPhraseExtractionSkill, LanguageDetectionSkill, MergeSkill, PIIDetectionSkill, SentimentSkill, SplitSkill, TextTranslationSkill, EntityLinkingSkill, EntityRecognitionSkillV3, SentimentSkillV3, ConditionalSkill, DocumentExtractionSkill, ShaperSkill, ImageAnalysisSkill, OcrSkill All required parameters must be populated in order to send to server. |
SearchIndexerSkillset |
A list of skills. All required parameters must be populated in order to send to Azure. |
SearchIndexerStatus |
Represents the current status and execution history of an indexer. Variables are only populated by the server, and will be ignored when sending a request. All required parameters must be populated in order to send to server. |
SearchIndexerWarning |
Represents an item-level warning. Variables are only populated by the server, and will be ignored when sending a request. All required parameters must be populated in order to send to server. |
SearchResourceEncryptionKey |
A customer-managed encryption key in Azure Key Vault. Keys that you create and manage can be used to encrypt or decrypt data-at-rest in Azure Cognitive Search, such as indexes and synonym maps. All required parameters must be populated in order to send to Azure. |
SearchServiceCounters |
Represents service-level resource counters and quotas. All required parameters must be populated in order to send to server. |
SearchServiceLimits |
Represents various service level limits. |
SearchServiceStatistics |
Response from a get service statistics request. If successful, it includes service level counters and limits. All required parameters must be populated in order to send to server. |
SearchSuggester |
Defines how the Suggest API should apply to a group of fields in the index. Variables are only populated by the server, and will be ignored when sending a request. All required parameters must be populated in order to send to server. |
SemanticConfiguration |
Defines a specific configuration to be used in the context of semantic capabilities. All required parameters must be populated in order to send to server. |
SemanticField |
A field that is used as part of the semantic configuration. All required parameters must be populated in order to send to server. |
SemanticPrioritizedFields |
Describes the title, content, and keywords fields to be used for semantic ranking, captions, highlights, and answers. |
SemanticSearch |
Defines parameters for a search index that influence semantic capabilities. |
SentimentSkill |
V1: Text analytics positive-negative sentiment analysis, scored as a floating point value in a range of zero to 1. V3: Using the Text Analytics API, evaluates unstructured text and for each record, provides sentiment labels (such as "negative", "neutral" and "positive") based on the highest confidence score found by the service at a sentence and document-level. All required parameters must be populated in order to send to Azure. |
ShaperSkill |
A skill for reshaping the outputs. It creates a complex type to support composite fields (also known as multipart fields). All required parameters must be populated in order to send to server. |
ShingleTokenFilter |
Creates combinations of tokens as a single token. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
SimilarityAlgorithm |
Base type for similarity algorithms. Similarity algorithms are used to calculate scores that tie queries to documents. The higher the score, the more relevant the document is to that specific query. Those scores are used to rank the search results. You probably want to use the sub-classes and not this class directly. Known sub-classes are: BM25SimilarityAlgorithm, ClassicSimilarityAlgorithm All required parameters must be populated in order to send to server. |
SnowballTokenFilter |
A filter that stems words using a Snowball-generated stemmer. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
SoftDeleteColumnDeletionDetectionPolicy |
Defines a data deletion detection policy that implements a soft-deletion strategy. It determines whether an item should be deleted based on the value of a designated 'soft delete' column. All required parameters must be populated in order to send to server. |
SplitSkill |
A skill to split a string into chunks of text. All required parameters must be populated in order to send to server. |
SqlIntegratedChangeTrackingPolicy |
Defines a data change detection policy that captures changes using the Integrated Change Tracking feature of Azure SQL Database. All required parameters must be populated in order to send to server. |
StemmerOverrideTokenFilter |
Provides the ability to override other stemming filters with custom dictionary-based stemming. Any dictionary-stemmed terms will be marked as keywords so that they will not be stemmed with stemmers down the chain. Must be placed before any stemming filters. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
StemmerTokenFilter |
Language specific stemming filter. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
StopAnalyzer |
Divides text at non-letters; Applies the lowercase and stopword token filters. This analyzer is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
StopwordsTokenFilter |
Removes stop words from a token stream. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
SuggestOptions |
Parameter group. |
SynonymMap |
Represents a synonym map definition. Variables are only populated by the server, and will be ignored when sending a request. All required parameters must be populated in order to send to Azure. |
SynonymTokenFilter |
Matches single or multi-word synonyms in a token stream. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
TagScoringFunction |
Defines a function that boosts scores of documents with string values matching a given list of tags. All required parameters must be populated in order to send to server. |
TagScoringParameters |
Provides parameter values to a tag scoring function. All required parameters must be populated in order to send to server. |
TextTranslationSkill |
A skill to translate text from one language to another. All required parameters must be populated in order to send to server. |
TextWeights |
Defines weights on index fields for which matches should boost scoring in search queries. All required parameters must be populated in order to send to server. |
TokenFilter |
Base type for token filters. You probably want to use the sub-classes and not this class directly. Known sub-classes are: AsciiFoldingTokenFilter, CjkBigramTokenFilter, CommonGramTokenFilter, DictionaryDecompounderTokenFilter, EdgeNGramTokenFilter, EdgeNGramTokenFilterV2, ElisionTokenFilter, KeepTokenFilter, KeywordMarkerTokenFilter, LengthTokenFilter, LimitTokenFilter, NGramTokenFilter, NGramTokenFilterV2, PatternCaptureTokenFilter, PatternReplaceTokenFilter, PhoneticTokenFilter, ShingleTokenFilter, SnowballTokenFilter, StemmerOverrideTokenFilter, StemmerTokenFilter, StopwordsTokenFilter, SynonymTokenFilter, TruncateTokenFilter, UniqueTokenFilter, WordDelimiterTokenFilter All required parameters must be populated in order to send to server. |
TruncateTokenFilter |
Truncates the terms to a specific length. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
UaxUrlEmailTokenizer |
Tokenizes urls and emails as one token. This tokenizer is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
UniqueTokenFilter |
Filters out tokens with same text as the previous token. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
VectorSearch |
Contains configuration options related to vector search. |
VectorSearchAlgorithmConfiguration |
Contains configuration options specific to the algorithm used during indexing or querying. You probably want to use the sub-classes and not this class directly. Known sub-classes are: ExhaustiveKnnAlgorithmConfiguration, HnswAlgorithmConfiguration All required parameters must be populated in order to send to server. |
VectorSearchCompression |
Contains configuration options specific to the compression method used during indexing or querying. You probably want to use the sub-classes and not this class directly. Known sub-classes are: BinaryQuantizationCompression, ScalarQuantizationCompression All required parameters must be populated in order to send to server. |
VectorSearchProfile |
Defines a combination of configurations to use with vector search. All required parameters must be populated in order to send to server. |
VectorSearchVectorizer |
Specifies the vectorization method to be used during query time. You probably want to use the sub-classes and not this class directly. Known sub-classes are: AzureOpenAIVectorizer, WebApiVectorizer All required parameters must be populated in order to send to server. |
WebApiSkill |
A skill that can call a Web API endpoint, allowing you to extend a skillset by having it call your custom code. All required parameters must be populated in order to send to server. |
WebApiVectorizer |
Specifies a user-defined vectorizer for generating the vector embedding of a query string. Integration of an external vectorizer is achieved using the custom Web API interface of a skillset. All required parameters must be populated in order to send to server. |
WebApiVectorizerParameters |
Specifies the properties for connecting to a user-defined vectorizer. |
WordDelimiterTokenFilter |
Splits words into subwords and performs optional transformations on subword groups. This token filter is implemented using Apache Lucene. All required parameters must be populated in order to send to server. |
Enums
AzureOpenAIModelName |
The Azure Open AI model name that will be called. |
BlobIndexerDataToExtract |
Specifies the data to extract from Azure blob storage and tells the indexer which data to extract from image content when "imageAction" is set to a value other than "none". This applies to embedded image content in a .PDF or other application, or image files such as .jpg and .png, in Azure blobs. |
BlobIndexerImageAction |
Determines how to process embedded images and image files in Azure blob storage. Setting the "imageAction" configuration to any value other than "none" requires that a skillset also be attached to that indexer. |
BlobIndexerPDFTextRotationAlgorithm |
Determines algorithm for text extraction from PDF files in Azure blob storage. |
BlobIndexerParsingMode |
Represents the parsing mode for indexing from an Azure blob data source. |
CharFilterName |
Defines the names of all character filters supported by the search engine. |
CjkBigramTokenFilterScripts |
Scripts that can be ignored by CjkBigramTokenFilter. |
CustomEntityLookupSkillLanguage |
The language codes supported for input text by CustomEntityLookupSkill. |
EdgeNGramTokenFilterSide |
Specifies which side of the input an n-gram should be generated from. |
EntityCategory |
A string indicating what entity categories to return. |
EntityRecognitionSkillLanguage |
Deprecated. The language codes supported for input text by EntityRecognitionSkill. |
EntityRecognitionSkillVersion |
Specifies the Entity Recognition skill version to use. |
ImageAnalysisSkillLanguage |
The language codes supported for input by ImageAnalysisSkill. |
ImageDetail |
A string indicating which domain-specific details to return. |
IndexProjectionMode |
Defines behavior of the index projections in relation to the rest of the indexer. |
IndexerExecutionEnvironment |
Specifies the environment in which the indexer should execute. |
IndexerExecutionStatus |
Represents the status of an individual indexer execution. |
IndexerStatus |
Represents the overall indexer status. |
KeyPhraseExtractionSkillLanguage |
The language codes supported for input text by KeyPhraseExtractionSkill. |
LexicalAnalyzerName |
Defines the names of all text analyzers supported by the search engine. |
LexicalTokenizerName |
Defines the names of all tokenizers supported by the search engine. |
MicrosoftStemmingTokenizerLanguage |
Lists the languages supported by the Microsoft language stemming tokenizer. |
MicrosoftTokenizerLanguage |
Lists the languages supported by the Microsoft language tokenizer. |
OcrLineEnding |
Defines the sequence of characters to use between the lines of text recognized by the OCR skill. The default value is "space". |
OcrSkillLanguage |
The language codes supported for input by OcrSkill. |
PIIDetectionSkillMaskingMode |
A string indicating what maskingMode to use to mask the personal information detected in the input text. |
PhoneticEncoder |
Identifies the type of phonetic encoder to use with a PhoneticTokenFilter. |
RegexFlags |
Defines flags that can be combined to control how regular expressions are used in the pattern analyzer and pattern tokenizer. |
ScoringFunctionAggregation |
Defines the aggregation function used to combine the results of all the scoring functions in a scoring profile. |
ScoringFunctionInterpolation |
Defines the function used to interpolate score boosting across a range of documents. |
SearchIndexerDataSourceType |
Defines the type of a datasource. |
SentimentSkillLanguage |
Deprecated. The language codes supported for input text by SentimentSkill. |
SentimentSkillVersion |
Specifies the Sentiment Skill version to use. |
SnowballTokenFilterLanguage |
The language to use for a Snowball token filter. |
SplitSkillLanguage |
The language codes supported for input text by SplitSkill. |
StemmerTokenFilterLanguage |
The language to use for a stemmer token filter. |
StopwordsList |
Identifies a predefined list of language-specific stopwords. |
TextSplitMode |
A value indicating which split mode to perform. |
TextTranslationSkillLanguage |
The language codes supported for input text by TextTranslationSkill. |
TokenCharacterKind |
Represents classes of characters on which a token filter can operate. |
TokenFilterName |
Defines the names of all token filters supported by the search engine. |
VectorEncodingFormat |
The encoding format for interpreting vector field contents. |
VectorSearchAlgorithmKind |
The algorithm used for indexing and querying. |
VectorSearchAlgorithmMetric |
The similarity metric to use for vector comparisons. It is recommended to choose the same similarity metric as the embedding model was trained on. |
VectorSearchCompressionTarget |
The quantized data type of compressed vector values. |
VectorSearchVectorizerKind |
The vectorization method to be used during query time. |
VisualFeature |
The strings indicating what visual feature types to return. |
Functions
ComplexField
Configure a Complex or Complex collection field for an Azure Search Index
ComplexField(*, name: str, collection: bool = False, fields: List[SearchField] | None = None, **kw) -> SearchField
Keyword-Only Parameters
Name | Description |
---|---|
name
|
Required. The name of the field, which must be unique within the fields collection of the index or parent field. |
collection
|
Whether this complex field is a collection (default False) |
fields
|
A list of sub-fields |
Returns
Type | Description |
---|---|
The search field object. |
SearchableField
Configure a searchable text field for an Azure Search Index
SearchableField(*, name: str, collection: bool = False, key: bool = False, hidden: bool = False, searchable: bool = True, filterable: bool = False, sortable: bool = False, facetable: bool = False, analyzer_name: str | LexicalAnalyzerName | None = None, search_analyzer_name: str | LexicalAnalyzerName | None = None, index_analyzer_name: str | LexicalAnalyzerName | None = None, synonym_map_names: List[str] | None = None, **kw) -> SearchField
Keyword-Only Parameters
Name | Description |
---|---|
name
|
Required. The name of the field, which must be unique within the fields collection of the index or parent field. |
collection
|
Whether this search field is a collection (default False) |
key
|
A value indicating whether the field uniquely identifies documents in the index. Exactly one top-level field in each index must be chosen as the key field and it must be of type SearchFieldDataType.String. Key fields can be used to look up documents directly and update or delete specific documents. Default is False |
hidden
|
A value indicating whether the field can be returned in a search result. You can enable this option if you want to use a field (for example, margin) as a filter, sorting, or scoring mechanism but do not want the field to be visible to the end user. This property must be False for key fields. This property can be changed on existing fields. Enabling this property does not cause any increase in index storage requirements. Default is False. |
searchable
|
A value indicating whether the field is full-text searchable. This means it will undergo analysis such as word-breaking during indexing. If you set a searchable field to a value like "sunny day", internally it will be split into the individual tokens "sunny" and "day". This enables full-text searches for these terms. Note: searchable fields consume extra space in your index since Azure Cognitive Search will store an additional tokenized version of the field value for full-text searches. If you want to save space in your index and you don't need a field to be included in searches, set searchable to false. Default is True. Default value: True
|
filterable
|
A value indicating whether to enable the field to be referenced in $filter queries. filterable differs from searchable in how strings are handled. Fields that are filterable do not undergo word-breaking, so comparisons are for exact matches only. For example, if you set such a field f to "sunny day", $filter=f eq 'sunny' will find no matches, but $filter=f eq 'sunny day' will. Default is False. |
sortable
|
A value indicating whether to enable the field to be referenced in $orderby expressions. By default Azure Cognitive Search sorts results by score, but in many experiences users will want to sort by fields in the documents. The default is False. |
facetable
|
A value indicating whether to enable the field to be referenced in facet queries. Typically used in a presentation of search results that includes hit count by category (for example, search for digital cameras and see hits by brand, by megapixels, by price, and so on). Default is False. |
analyzer_name
|
The name of the analyzer to use for the field. This option can't be set together with either searchAnalyzer or indexAnalyzer. Once the analyzer is chosen, it cannot be changed for the field. Possible values include: 'ar.microsoft', 'ar.lucene', 'hy.lucene', 'bn.microsoft', 'eu.lucene', 'bg.microsoft', 'bg.lucene', 'ca.microsoft', 'ca.lucene', 'zh- Hans.microsoft', 'zh-Hans.lucene', 'zh-Hant.microsoft', 'zh-Hant.lucene', 'hr.microsoft', 'cs.microsoft', 'cs.lucene', 'da.microsoft', 'da.lucene', 'nl.microsoft', 'nl.lucene', 'en.microsoft', 'en.lucene', 'et.microsoft', 'fi.microsoft', 'fi.lucene', 'fr.microsoft', 'fr.lucene', 'gl.lucene', 'de.microsoft', 'de.lucene', 'el.microsoft', 'el.lucene', 'gu.microsoft', 'he.microsoft', 'hi.microsoft', 'hi.lucene', 'hu.microsoft', 'hu.lucene', 'is.microsoft', 'id.microsoft', 'id.lucene', 'ga.lucene', 'it.microsoft', 'it.lucene', 'ja.microsoft', 'ja.lucene', 'kn.microsoft', 'ko.microsoft', 'ko.lucene', 'lv.microsoft', 'lv.lucene', 'lt.microsoft', 'ml.microsoft', 'ms.microsoft', 'mr.microsoft', 'nb.microsoft', 'no.lucene', 'fa.lucene', 'pl.microsoft', 'pl.lucene', 'pt-BR.microsoft', 'pt-BR.lucene', 'pt- PT.microsoft', 'pt-PT.lucene', 'pa.microsoft', 'ro.microsoft', 'ro.lucene', 'ru.microsoft', 'ru.lucene', 'sr-cyrillic.microsoft', 'sr-latin.microsoft', 'sk.microsoft', 'sl.microsoft', 'es.microsoft', 'es.lucene', 'sv.microsoft', 'sv.lucene', 'ta.microsoft', 'te.microsoft', 'th.microsoft', 'th.lucene', 'tr.microsoft', 'tr.lucene', 'uk.microsoft', 'ur.microsoft', 'vi.microsoft', 'standard.lucene', 'standardasciifolding.lucene', 'keyword', 'pattern', 'simple', 'stop', 'whitespace'. |
search_analyzer_name
|
The name of the analyzer used at search time for the field. It must be set together with indexAnalyzer and it cannot be set together with the analyzer option. This property cannot be set to the name of a language analyzer; use the analyzer property instead if you need a language analyzer. This analyzer can be updated on an existing field. Possible values include: 'ar.microsoft', 'ar.lucene', 'hy.lucene', 'bn.microsoft', 'eu.lucene', 'bg.microsoft', 'bg.lucene', 'ca.microsoft', 'ca.lucene', 'zh-Hans.microsoft', 'zh-Hans.lucene', 'zh- Hant.microsoft', 'zh-Hant.lucene', 'hr.microsoft', 'cs.microsoft', 'cs.lucene', 'da.microsoft', 'da.lucene', 'nl.microsoft', 'nl.lucene', 'en.microsoft', 'en.lucene', 'et.microsoft', 'fi.microsoft', 'fi.lucene', 'fr.microsoft', 'fr.lucene', 'gl.lucene', 'de.microsoft', 'de.lucene', 'el.microsoft', 'el.lucene', 'gu.microsoft', 'he.microsoft', 'hi.microsoft', 'hi.lucene', 'hu.microsoft', 'hu.lucene', 'is.microsoft', 'id.microsoft', 'id.lucene', 'ga.lucene', 'it.microsoft', 'it.lucene', 'ja.microsoft', 'ja.lucene', 'kn.microsoft', 'ko.microsoft', 'ko.lucene', 'lv.microsoft', 'lv.lucene', 'lt.microsoft', 'ml.microsoft', 'ms.microsoft', 'mr.microsoft', 'nb.microsoft', 'no.lucene', 'fa.lucene', 'pl.microsoft', 'pl.lucene', 'pt-BR.microsoft', 'pt-BR.lucene', 'pt-PT.microsoft', 'pt-PT.lucene', 'pa.microsoft', 'ro.microsoft', 'ro.lucene', 'ru.microsoft', 'ru.lucene', 'sr- cyrillic.microsoft', 'sr-latin.microsoft', 'sk.microsoft', 'sl.microsoft', 'es.microsoft', 'es.lucene', 'sv.microsoft', 'sv.lucene', 'ta.microsoft', 'te.microsoft', 'th.microsoft', 'th.lucene', 'tr.microsoft', 'tr.lucene', 'uk.microsoft', 'ur.microsoft', 'vi.microsoft', 'standard.lucene', 'standardasciifolding.lucene', 'keyword', 'pattern', 'simple', 'stop', 'whitespace'. |
index_analyzer_name
|
The name of the analyzer used at indexing time for the field. It must be set together with searchAnalyzer and it cannot be set together with the analyzer option. This property cannot be set to the name of a language analyzer; use the analyzer property instead if you need a language analyzer. Once the analyzer is chosen, it cannot be changed for the field. Possible values include: 'ar.microsoft', 'ar.lucene', 'hy.lucene', 'bn.microsoft', 'eu.lucene', 'bg.microsoft', 'bg.lucene', 'ca.microsoft', 'ca.lucene', 'zh-Hans.microsoft', 'zh- Hans.lucene', 'zh-Hant.microsoft', 'zh-Hant.lucene', 'hr.microsoft', 'cs.microsoft', 'cs.lucene', 'da.microsoft', 'da.lucene', 'nl.microsoft', 'nl.lucene', 'en.microsoft', 'en.lucene', 'et.microsoft', 'fi.microsoft', 'fi.lucene', 'fr.microsoft', 'fr.lucene', 'gl.lucene', 'de.microsoft', 'de.lucene', 'el.microsoft', 'el.lucene', 'gu.microsoft', 'he.microsoft', 'hi.microsoft', 'hi.lucene', 'hu.microsoft', 'hu.lucene', 'is.microsoft', 'id.microsoft', 'id.lucene', 'ga.lucene', 'it.microsoft', 'it.lucene', 'ja.microsoft', 'ja.lucene', 'kn.microsoft', 'ko.microsoft', 'ko.lucene', 'lv.microsoft', 'lv.lucene', 'lt.microsoft', 'ml.microsoft', 'ms.microsoft', 'mr.microsoft', 'nb.microsoft', 'no.lucene', 'fa.lucene', 'pl.microsoft', 'pl.lucene', 'pt-BR.microsoft', 'pt-BR.lucene', 'pt-PT.microsoft', 'pt-PT.lucene', 'pa.microsoft', 'ro.microsoft', 'ro.lucene', 'ru.microsoft', 'ru.lucene', 'sr- cyrillic.microsoft', 'sr-latin.microsoft', 'sk.microsoft', 'sl.microsoft', 'es.microsoft', 'es.lucene', 'sv.microsoft', 'sv.lucene', 'ta.microsoft', 'te.microsoft', 'th.microsoft', 'th.lucene', 'tr.microsoft', 'tr.lucene', 'uk.microsoft', 'ur.microsoft', 'vi.microsoft', 'standard.lucene', 'standardasciifolding.lucene', 'keyword', 'pattern', 'simple', 'stop', 'whitespace'. |
synonym_map_names
|
A list of the names of synonym maps to associate with this field. Currently only one synonym map per field is supported. Assigning a synonym map to a field ensures that query terms targeting that field are expanded at query-time using the rules in the synonym map. This attribute can be changed on existing fields. |
Returns
Type | Description |
---|---|
The search field object. |
SimpleField
Configure a simple field for an Azure Search Index
SimpleField(*, name: str, type: str, key: bool = False, hidden: bool = False, filterable: bool = False, sortable: bool = False, facetable: bool = False, **kw) -> SearchField
Keyword-Only Parameters
Name | Description |
---|---|
name
|
Required. The name of the field, which must be unique within the fields collection of the index or parent field. |
type
|
Required. The data type of the field. Possible values include: SearchFieldDataType.String, SearchFieldDataType.Int32, SearchFieldDataType.Int64, SearchFieldDataType.Double, SearchFieldDataType.Boolean, SearchFieldDataType.DateTimeOffset, SearchFieldDataType.GeographyPoint, SearchFieldDataType.ComplexType, from azure.search.documents.SearchFieldDataType. |
key
|
A value indicating whether the field uniquely identifies documents in the index. Exactly one top-level field in each index must be chosen as the key field and it must be of type SearchFieldDataType.String. Key fields can be used to look up documents directly and update or delete specific documents. Default is False |
hidden
|
A value indicating whether the field can be returned in a search result. You can enable this option if you want to use a field (for example, margin) as a filter, sorting, or scoring mechanism but do not want the field to be visible to the end user. This property must be False for key fields. This property can be changed on existing fields. Enabling this property does not cause any increase in index storage requirements. Default is False. |
filterable
|
A value indicating whether to enable the field to be referenced in $filter queries. filterable differs from searchable in how strings are handled. Fields of type SearchFieldDataType.String or Collection(SearchFieldDataType.String) that are filterable do not undergo word-breaking, so comparisons are for exact matches only. For example, if you set such a field f to "sunny day", $filter=f eq 'sunny' will find no matches, but $filter=f eq 'sunny day' will. This property must be null for complex fields. Default is False |
sortable
|
A value indicating whether to enable the field to be referenced in $orderby expressions. By default Azure Cognitive Search sorts results by score, but in many experiences users will want to sort by fields in the documents. A simple field can be sortable only if it is single-valued (it has a single value in the scope of the parent document). Simple collection fields cannot be sortable, since they are multi-valued. Simple sub-fields of complex collections are also multi-valued, and therefore cannot be sortable. This is true whether it's an immediate parent field, or an ancestor field, that's the complex collection. The default is False. |
facetable
|
A value indicating whether to enable the field to be referenced in facet queries. Typically used in a presentation of search results that includes hit count by category (for example, search for digital cameras and see hits by brand, by megapixels, by price, and so on). Fields of type SearchFieldDataType.GeographyPoint or Collection(SearchFieldDataType.GeographyPoint) cannot be facetable. Default is False. |
Returns
Type | Description |
---|---|
The search field object. |
Azure SDK for Python