你当前正在访问 Microsoft Azure Global Edition 技术文档网站。 如果需要访问由世纪互联运营的 Microsoft Azure 中国技术文档网站,请访问 https://docs.azure.cn

models Package

Classes

AIServicesAccountIdentity

The multi-region account of an Azure AI service resource that's attached to a skillset.

AIServicesAccountKey

The account key of an Azure AI service resource that's attached to a skillset, to be used with the resource's subdomain.

AnalyzeResult

The result of testing an analyzer on text.

AnalyzeTextOptions

Specifies some text and analysis components used to break that text into tokens.

AnalyzedTokenInfo

Information about a token returned by an analyzer.

AsciiFoldingTokenFilter

Converts alphabetic, numeric, and symbolic Unicode characters which are not in the first 127 ASCII characters (the "Basic Latin" Unicode block) into their ASCII equivalents, if such equivalents exist. This token filter is implemented using Apache Lucene.

AzureActiveDirectoryApplicationCredentials

Credentials of a registered application created for your search service, used for authenticated access to the encryption keys stored in Azure Key Vault.

AzureBlobKnowledgeSource

Configuration for Azure Blob Storage knowledge source.

AzureBlobKnowledgeSourceParameters

Parameters for Azure Blob Storage knowledge source.

AzureMachineLearningParameters

Specifies the properties for connecting to an AML vectorizer.

AzureMachineLearningVectorizer

Specifies an Azure Machine Learning endpoint deployed via the Azure AI Foundry Model Catalog for generating the vector embedding of a query string.

AzureOpenAIEmbeddingSkill

Allows you to generate a vector embedding for a given text input using the Azure OpenAI resource.

AzureOpenAIVectorizer

Specifies the Azure OpenAI resource used to vectorize a query string.

AzureOpenAIVectorizerParameters

Specifies the parameters for connecting to the Azure OpenAI resource.

BM25SimilarityAlgorithm

Ranking function based on the Okapi BM25 similarity algorithm. BM25 is a TF-IDF-like algorithm that includes length normalization (controlled by the 'b' parameter) as well as term frequency saturation (controlled by the 'k1' parameter).

BinaryQuantizationCompression

Contains configuration options specific to the binary quantization compression method used during indexing and querying.

CharFilter

Base type for character filters.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: MappingCharFilter, PatternReplaceCharFilter

ChatCompletionCommonModelParameters

Common language model parameters for Chat Completions. If omitted, default values are used.

ChatCompletionResponseFormat

Determines how the language model's response should be serialized. Defaults to 'text'.

ChatCompletionSchema

Object defining the custom schema the model will use to structure its output.

ChatCompletionSchemaProperties

Properties for JSON schema response format.

ChatCompletionSkill

A skill that calls a language model via Azure AI Foundry's Chat Completions endpoint.

CjkBigramTokenFilter

Forms bigrams of CJK terms that are generated from the standard tokenizer. This token filter is implemented using Apache Lucene.

ClassicSimilarityAlgorithm

Legacy similarity algorithm which uses the Lucene TFIDFSimilarity implementation of TF-IDF. This variation of TF-IDF introduces static document length normalization as well as coordinating factors that penalize documents that only partially match the searched queries.

ClassicTokenizer

Grammar-based tokenizer that is suitable for processing most European-language documents. This tokenizer is implemented using Apache Lucene.

CognitiveServicesAccount

Base type for describing any Azure AI service resource attached to a skillset.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: AIServicesAccountIdentity, AIServicesAccountKey, CognitiveServicesAccountKey, DefaultCognitiveServicesAccount

CognitiveServicesAccountKey

The multi-region account key of an Azure AI service resource that's attached to a skillset.

CommonGramTokenFilter

Construct bigrams for frequently occurring terms while indexing. Single terms are still indexed too, with bigrams overlaid. This token filter is implemented using Apache Lucene.

ConditionalSkill

A skill that enables scenarios that require a Boolean operation to determine the data to assign to an output.

ContentUnderstandingSkill

A skill that leverages Azure AI Content Understanding to process and extract structured insights from documents, enabling enriched, searchable content for enhanced document indexing and retrieval.

ContentUnderstandingSkillChunkingProperties

Controls the cardinality for chunking the content.

CorsOptions

Defines options to control Cross-Origin Resource Sharing (CORS) for an index.

CreatedResources

Resources created by the knowledge source. Keys represent resource types (e.g., 'datasource', 'indexer', 'skillset', 'index') and values represent resource names.

CustomAnalyzer

Allows you to take control over the process of converting text into indexable/searchable tokens. It's a user-defined configuration consisting of a single predefined tokenizer and one or more filters. The tokenizer is responsible for breaking text into tokens, and the filters for modifying tokens emitted by the tokenizer.

CustomEntity

An object that contains information about the matches that were found, and related metadata.

CustomEntityAlias

A complex object that can be used to specify alternative spellings or synonyms to the root entity name.

CustomEntityLookupSkill

A skill looks for text from a custom, user-defined list of words and phrases.

CustomNormalizer

Allows you to configure normalization for filterable, sortable, and facetable fields, which by default operate with strict matching. This is a user-defined configuration consisting of at least one or more filters, which modify the token that is stored.

DataChangeDetectionPolicy

Base type for data change detection policies.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: HighWaterMarkChangeDetectionPolicy, SqlIntegratedChangeTrackingPolicy

DataDeletionDetectionPolicy

Base type for data deletion detection policies.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: NativeBlobSoftDeleteDeletionDetectionPolicy, SoftDeleteColumnDeletionDetectionPolicy

DataSourceCredentials

Represents credentials that can be used to connect to a datasource.

DefaultCognitiveServicesAccount

An empty object that represents the default Azure AI service resource for a skillset.

DictionaryDecompounderTokenFilter

Decomposes compound words found in many Germanic languages. This token filter is implemented using Apache Lucene.

DistanceScoringFunction

Defines a function that boosts scores based on distance from a geographic location.

DistanceScoringParameters

Provides parameter values to a distance scoring function.

DocumentExtractionSkill

A skill that extracts content from a file within the enrichment pipeline.

DocumentIntelligenceLayoutSkill

A skill that extracts content and layout information, via Azure AI Services, from files within the enrichment pipeline.

DocumentIntelligenceLayoutSkillChunkingProperties

Controls the cardinality for chunking the content.

DocumentKeysOrIds

The type of the keysOrIds.

EdgeNGramTokenFilter

Generates n-grams of the given size(s) starting from the front or the back of an input token. This token filter is implemented using Apache Lucene.

EdgeNGramTokenFilterV2

Generates n-grams of the given size(s) starting from the front or the back of an input token. This token filter is implemented using Apache Lucene.

EdgeNGramTokenizer

Tokenizes the input from an edge into n-grams of the given size(s). This tokenizer is implemented using Apache Lucene.

ElisionTokenFilter

Removes elisions. For example, "l'avion" (the plane) will be converted to "avion" (plane). This token filter is implemented using Apache Lucene.

EntityLinkingSkill

Using the Text Analytics API, extracts linked entities from text.

EntityRecognitionSkillV3

Using the Text Analytics API, extracts entities of different types from text.

ExhaustiveKnnAlgorithmConfiguration

Contains configuration options specific to the exhaustive KNN algorithm used during querying, which will perform brute-force search across the entire vector index.

ExhaustiveKnnParameters

Contains the parameters specific to exhaustive KNN algorithm.

FieldMapping

Defines a mapping between a field in a data source and a target field in an index.

FieldMappingFunction

Represents a function that transforms a value from a data source before indexing.

FreshnessScoringFunction

Defines a function that boosts scores based on the value of a date-time field.

FreshnessScoringParameters

Provides parameter values to a freshness scoring function.

GetIndexStatisticsResult

Statistics for a given index. Statistics are collected periodically and are not guaranteed to always be up-to-date.

HighWaterMarkChangeDetectionPolicy

Defines a data change detection policy that captures changes based on the value of a high water mark column.

HnswAlgorithmConfiguration

Contains configuration options specific to the HNSW approximate nearest neighbors algorithm used during indexing and querying. The HNSW algorithm offers a tunable trade-off between search speed and accuracy.

HnswParameters

Contains the parameters specific to the HNSW algorithm.

ImageAnalysisSkill

A skill that analyzes image files. It extracts a rich set of visual features based on the image content.

IndexedOneLakeKnowledgeSource

Configuration for OneLake knowledge source.

IndexedOneLakeKnowledgeSourceParameters

Parameters for OneLake knowledge source.

IndexerExecutionResult

Represents the result of an individual indexer execution.

IndexerResyncBody

Request body for resync indexer operation.

IndexingParameters

Represents parameters for indexer execution.

IndexingParametersConfiguration

A dictionary of indexer-specific configuration properties. Each name is the name of a specific property. Each value must be of a primitive type.

IndexingSchedule

Represents a schedule for indexer execution.

InputFieldMappingEntry

Input field mapping for a skill.

KeepTokenFilter

A token filter that only keeps tokens with text contained in a specified list of words. This token filter is implemented using Apache Lucene.

KeyPhraseExtractionSkill

A skill that uses text analytics for key phrase extraction.

KeywordMarkerTokenFilter

Marks terms as keywords. This token filter is implemented using Apache Lucene.

KeywordTokenizer

Emits the entire input as a single token. This tokenizer is implemented using Apache Lucene.

KeywordTokenizerV2

Emits the entire input as a single token. This tokenizer is implemented using Apache Lucene.

KnowledgeBase

Represents a knowledge base definition.

KnowledgeBaseAzureOpenAIModel

Specifies the Azure OpenAI resource used to do query planning.

KnowledgeBaseModel

Specifies the connection parameters for the model to use for query planning.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: KnowledgeBaseAzureOpenAIModel

KnowledgeSource

Represents a knowledge source definition.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: AzureBlobKnowledgeSource, IndexedOneLakeKnowledgeSource, SearchIndexKnowledgeSource, WebKnowledgeSource

KnowledgeSourceReference

Reference to a knowledge source.

LanguageDetectionSkill

A skill that detects the language of input text and reports a single language code for every document submitted on the request. The language code is paired with a score indicating the confidence of the analysis.

LengthTokenFilter

Removes words that are too long or too short. This token filter is implemented using Apache Lucene.

LexicalAnalyzer

Base type for analyzers.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: CustomAnalyzer, PatternAnalyzer, LuceneStandardAnalyzer, StopAnalyzer

LexicalNormalizer

Base type for normalizers.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: CustomNormalizer

LexicalTokenizer

Base type for tokenizers.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: ClassicTokenizer, EdgeNGramTokenizer, KeywordTokenizer, KeywordTokenizerV2, MicrosoftLanguageStemmingTokenizer, MicrosoftLanguageTokenizer, NGramTokenizer, PathHierarchyTokenizerV2, PatternTokenizer, LuceneStandardTokenizer, LuceneStandardTokenizerV2, UaxUrlEmailTokenizer

LimitTokenFilter

Limits the number of tokens while indexing. This token filter is implemented using Apache Lucene.

LuceneStandardAnalyzer

Standard Apache Lucene analyzer; Composed of the standard tokenizer, lowercase filter and stop filter.

LuceneStandardTokenizer

Breaks text following the Unicode Text Segmentation rules. This tokenizer is implemented using Apache Lucene.

LuceneStandardTokenizerV2

Breaks text following the Unicode Text Segmentation rules. This tokenizer is implemented using Apache Lucene.

MagnitudeScoringFunction

Defines a function that boosts scores based on the magnitude of a numeric field.

MagnitudeScoringParameters

Provides parameter values to a magnitude scoring function.

MappingCharFilter

A character filter that applies mappings defined with the mappings option. Matching is greedy (longest pattern matching at a given point wins). Replacement is allowed to be the empty string. This character filter is implemented using Apache Lucene.

MergeSkill

A skill for merging two or more strings into a single unified string, with an optional user-defined delimiter separating each component part.

MicrosoftLanguageStemmingTokenizer

Divides text using language-specific rules and reduces words to their base forms.

MicrosoftLanguageTokenizer

Divides text using language-specific rules.

NGramTokenFilter

Generates n-grams of the given size(s). This token filter is implemented using Apache Lucene.

NGramTokenFilterV2

Generates n-grams of the given size(s). This token filter is implemented using Apache Lucene.

NGramTokenizer

Tokenizes the input into n-grams of the given size(s). This tokenizer is implemented using Apache Lucene.

NativeBlobSoftDeleteDeletionDetectionPolicy

Defines a data deletion detection policy utilizing Azure Blob Storage's native soft delete feature for deletion detection.

OcrSkill

A skill that extracts text from image files.

OutputFieldMappingEntry

Output field mapping for a skill.

PIIDetectionSkill

Using the Text Analytics API, extracts personal information from an input text and gives you the option of masking it.

PathHierarchyTokenizerV2

Tokenizer for path-like hierarchies. This tokenizer is implemented using Apache Lucene.

PatternAnalyzer

Flexibly separates text into terms via a regular expression pattern. This analyzer is implemented using Apache Lucene.

PatternCaptureTokenFilter

Uses Java regexes to emit multiple tokens - one for each capture group in one or more patterns. This token filter is implemented using Apache Lucene.

PatternReplaceCharFilter

A character filter that replaces characters in the input string. It uses a regular expression to identify character sequences to preserve and a replacement pattern to identify characters to replace. For example, given the input text "aa bb aa bb", pattern "(aa)\s+(bb)", and replacement "$1#$2", the result would be "aa#bb aa#bb". This character filter is implemented using Apache Lucene.

PatternReplaceTokenFilter

A character filter that replaces characters in the input string. It uses a regular expression to identify character sequences to preserve and a replacement pattern to identify characters to replace. For example, given the input text "aa bb aa bb", pattern "(aa)\s+(bb)", and replacement "$1#$2", the result would be "aa#bb aa#bb". This token filter is implemented using Apache Lucene.

PatternTokenizer

Tokenizer that uses regex pattern matching to construct distinct tokens. This tokenizer is implemented using Apache Lucene.

PhoneticTokenFilter

Create tokens for phonetic matches. This token filter is implemented using Apache Lucene.

RescoringOptions

Contains the options for rescoring.

ResourceCounter

Represents a resource's usage and quota.

ScalarQuantizationCompression

Contains configuration options specific to the scalar quantization compression method used during indexing and querying.

ScalarQuantizationParameters

Contains the parameters specific to Scalar Quantization.

ScoringFunction

Base type for functions that can modify document scores during ranking.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: DistanceScoringFunction, FreshnessScoringFunction, MagnitudeScoringFunction, TagScoringFunction

ScoringProfile

Defines parameters for a search index that influence scoring in search queries.

SearchAlias

Represents an index alias, which describes a mapping from the alias name to an index. The alias name can be used in place of the index name for supported operations.

SearchField

Represents a field in an index definition, which describes the name, data type, and search behavior of a field.

This class adds backward compatibility support for the 'hidden' property, which is the inverse of 'retrievable'.

SearchIndex

Represents a search index definition, which describes the fields and search behavior of an index.

SearchIndexFieldReference

Field reference for a search index.

SearchIndexKnowledgeSource

Knowledge Source targeting a search index.

SearchIndexKnowledgeSourceParameters

Parameters for search index knowledge source.

SearchIndexer

Represents an indexer.

SearchIndexerDataContainer

Represents information about the entity (such as Azure SQL table or CosmosDB collection) that will be indexed.

SearchIndexerDataIdentity

Abstract base type for data identities.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: SearchIndexerDataNoneIdentity, SearchIndexerDataUserAssignedIdentity

SearchIndexerDataNoneIdentity

Clears the identity property of a datasource.

SearchIndexerDataSourceConnection

Represents a datasource definition, which can be used to configure an indexer.

This class adds an additional overload to support passing connection_string directly instead of credentials.

SearchIndexerDataUserAssignedIdentity

Specifies the identity for a datasource to use.

SearchIndexerError

Represents an item- or document-level indexing error.

SearchIndexerIndexProjection

Definition of additional projections to secondary search indexes.

SearchIndexerIndexProjectionSelector

Description for what data to store in the designated search index.

SearchIndexerIndexProjectionsParameters

A dictionary of index projection-specific configuration properties. Each name is the name of a specific property. Each value must be of a primitive type.

SearchIndexerKnowledgeStore

Definition of additional projections to azure blob, table, or files, of enriched data.

SearchIndexerKnowledgeStoreBlobProjectionSelector

Abstract class to share properties between concrete selectors.

SearchIndexerKnowledgeStoreFileProjectionSelector

Projection definition for what data to store in Azure Files.

SearchIndexerKnowledgeStoreObjectProjectionSelector

Projection definition for what data to store in Azure Blob.

SearchIndexerKnowledgeStoreProjection

Container object for various projection selectors.

SearchIndexerKnowledgeStoreProjectionSelector

Abstract class to share properties between concrete selectors.

SearchIndexerKnowledgeStoreTableProjectionSelector

Description for what data to store in Azure Tables.

SearchIndexerLimits

Represents the limits that can be applied to an indexer.

SearchIndexerSkill

Base type for skills.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: ChatCompletionSkill, WebApiSkill, AzureOpenAIEmbeddingSkill, CustomEntityLookupSkill, KeyPhraseExtractionSkill, LanguageDetectionSkill, MergeSkill, PIIDetectionSkill, SplitSkill, TextTranslationSkill, EntityLinkingSkill, EntityRecognitionSkillV3, SentimentSkillV3, ConditionalSkill, ContentUnderstandingSkill, DocumentExtractionSkill, DocumentIntelligenceLayoutSkill, ShaperSkill, ImageAnalysisSkill, OcrSkill

SearchIndexerSkillset

A list of skills.

SearchIndexerStatus

Represents the current status and execution history of an indexer.

SearchIndexerWarning

Represents an item-level warning.

SearchResourceEncryptionKey

A customer-managed encryption key in Azure Key Vault. Keys that you create and manage can be used to encrypt or decrypt data-at-rest, such as indexes and synonym maps.

SearchServiceCounters

Represents service-level resource counters and quotas.

SearchServiceLimits

Represents various service level limits.

SearchServiceStatistics

Response from a get service statistics request. If successful, it includes service level counters and limits.

SearchSuggester

Defines how the Suggest API should apply to a group of fields in the index.

SemanticConfiguration

Defines a specific configuration to be used in the context of semantic capabilities.

SemanticField

A field that is used as part of the semantic configuration.

SemanticPrioritizedFields

Describes the title, content, and keywords fields to be used for semantic ranking, captions, highlights, and answers.

SemanticSearch

Defines parameters for a search index that influence semantic capabilities.

SentimentSkillV3

Using the Text Analytics API, evaluates unstructured text and for each record, provides sentiment labels (such as "negative", "neutral" and "positive") based on the highest confidence score found by the service at a sentence and document-level.

ShaperSkill

A skill for reshaping the outputs. It creates a complex type to support composite fields (also known as multipart fields).

ShingleTokenFilter

Creates combinations of tokens as a single token. This token filter is implemented using Apache Lucene.

SimilarityAlgorithm

Base type for similarity algorithms. Similarity algorithms are used to calculate scores that tie queries to documents. The higher the score, the more relevant the document is to that specific query. Those scores are used to rank the search results.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: BM25SimilarityAlgorithm, ClassicSimilarityAlgorithm

SkillNames

The type of the skill names.

SnowballTokenFilter

A filter that stems words using a Snowball-generated stemmer. This token filter is implemented using Apache Lucene.

SoftDeleteColumnDeletionDetectionPolicy

Defines a data deletion detection policy that implements a soft-deletion strategy. It determines whether an item should be deleted based on the value of a designated 'soft delete' column.

SplitSkill

A skill to split a string into chunks of text.

SqlIntegratedChangeTrackingPolicy

Defines a data change detection policy that captures changes using the Integrated Change Tracking feature of Azure SQL Database.

StemmerOverrideTokenFilter

Provides the ability to override other stemming filters with custom dictionary-based stemming. Any dictionary-stemmed terms will be marked as keywords so that they will not be stemmed with stemmers down the chain. Must be placed before any stemming filters. This token filter is implemented using Apache Lucene. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/miscellaneous/StemmerOverrideFilter.html.

StemmerTokenFilter

Language specific stemming filter. This token filter is implemented using Apache Lucene. See https://learn.microsoft.com/rest/api/searchservice/Custom-analyzers-in-Azure-Search#TokenFilters.

StopAnalyzer

Divides text at non-letters; Applies the lowercase and stopword token filters. This analyzer is implemented using Apache Lucene.

StopwordsTokenFilter

Removes stop words from a token stream. This token filter is implemented using Apache Lucene. See http://lucene.apache.org/core/4_10_3/analyzers-common/org/apache/lucene/analysis/core/StopFilter.html.

SynonymMap

Represents a synonym map definition.

SynonymTokenFilter

Matches single or multi-word synonyms in a token stream. This token filter is implemented using Apache Lucene.

TagScoringFunction

Defines a function that boosts scores of documents with string values matching a given list of tags.

TagScoringParameters

Provides parameter values to a tag scoring function.

TextTranslationSkill

A skill to translate text from one language to another.

TextWeights

Defines weights on index fields for which matches should boost scoring in search queries.

TokenFilter

Base type for token filters.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: AsciiFoldingTokenFilter, CjkBigramTokenFilter, CommonGramTokenFilter, DictionaryDecompounderTokenFilter, EdgeNGramTokenFilter, EdgeNGramTokenFilterV2, ElisionTokenFilter, KeepTokenFilter, KeywordMarkerTokenFilter, LengthTokenFilter, LimitTokenFilter, NGramTokenFilter, NGramTokenFilterV2, PatternCaptureTokenFilter, PatternReplaceTokenFilter, PhoneticTokenFilter, ShingleTokenFilter, SnowballTokenFilter, StemmerOverrideTokenFilter, StemmerTokenFilter, StopwordsTokenFilter, SynonymTokenFilter, TruncateTokenFilter, UniqueTokenFilter, WordDelimiterTokenFilter

TruncateTokenFilter

Truncates the terms to a specific length. This token filter is implemented using Apache Lucene.

UaxUrlEmailTokenizer

Tokenizes urls and emails as one token. This tokenizer is implemented using Apache Lucene.

UniqueTokenFilter

Filters out tokens with same text as the previous token. This token filter is implemented using Apache Lucene.

VectorSearch

Contains configuration options related to vector search.

VectorSearchAlgorithmConfiguration

Contains configuration options specific to the algorithm used during indexing or querying.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: ExhaustiveKnnAlgorithmConfiguration, HnswAlgorithmConfiguration

VectorSearchCompression

Contains configuration options specific to the compression method used during indexing or querying.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: BinaryQuantizationCompression, ScalarQuantizationCompression

VectorSearchProfile

Defines a combination of configurations to use with vector search.

VectorSearchVectorizer

Specifies the vectorization method to be used during query time.

You probably want to use the sub-classes and not this class directly. Known sub-classes are: AzureMachineLearningVectorizer, AzureOpenAIVectorizer, WebApiVectorizer

WebApiHttpHeaders

A dictionary of http request headers.

WebApiSkill

A skill that can call a Web API endpoint, allowing you to extend a skillset by having it call your custom code.

WebApiVectorizer

Specifies a user-defined vectorizer for generating the vector embedding of a query string. Integration of an external vectorizer is achieved using the custom Web API interface of a skillset.

WebApiVectorizerParameters

Specifies the properties for connecting to a user-defined vectorizer.

WebKnowledgeSource

Knowledge Source targeting web results.

WebKnowledgeSourceDomain

Configuration for web knowledge source domain.

WebKnowledgeSourceDomains

Domain allow/block configuration for web knowledge source.

WebKnowledgeSourceParameters

Parameters for web knowledge source.

WordDelimiterTokenFilter

Splits words into subwords and performs optional transformations on subword groups. This token filter is implemented using Apache Lucene.

Enums

AIFoundryModelCatalogName

The name of the embedding model from the Azure AI Foundry Catalog that will be called.

AzureOpenAIModelName

The Azure Open AI model name that will be called.

BlobIndexerDataToExtract

Specifies the data to extract from Azure blob storage and tells the indexer which data to extract from image content when "imageAction" is set to a value other than "none". This applies to embedded image content in a .PDF or other application, or image files such as .jpg and .png, in Azure blobs.

BlobIndexerImageAction

Determines how to process embedded images and image files in Azure blob storage. Setting the "imageAction" configuration to any value other than "none" requires that a skillset also be attached to that indexer.

BlobIndexerPDFTextRotationAlgorithm

Determines algorithm for text extraction from PDF files in Azure blob storage.

BlobIndexerParsingMode

Represents the parsing mode for indexing from an Azure blob data source.

CharFilterName

Defines the names of all character filters supported by the search engine.

ChatCompletionExtraParametersBehavior

Specifies how 'extraParameters' should be handled by Azure AI Foundry. Defaults to 'error'.

ChatCompletionResponseFormatType

Specifies how the LLM should format the response.

CjkBigramTokenFilterScripts

Scripts that can be ignored by CjkBigramTokenFilter.

ContentUnderstandingSkillChunkingUnit

Controls the cardinality of the chunk unit. Default is 'characters'.

ContentUnderstandingSkillExtractionOptions

Controls the cardinality of the content extracted from the document by the skill.

CustomEntityLookupSkillLanguage

The language codes supported for input text by CustomEntityLookupSkill.

DocumentIntelligenceLayoutSkillChunkingUnit

Controls the cardinality of the chunk unit. Default is 'characters'.

DocumentIntelligenceLayoutSkillExtractionOptions

Controls the cardinality of the content extracted from the document by the skill.

DocumentIntelligenceLayoutSkillMarkdownHeaderDepth

The depth of headers in the markdown output. Default is h6.

DocumentIntelligenceLayoutSkillOutputFormat

Controls the cardinality of the output format. Default is 'markdown'.

DocumentIntelligenceLayoutSkillOutputMode

Controls the cardinality of the output produced by the skill. Default is 'oneToMany'.

EdgeNGramTokenFilterSide

Specifies which side of the input an n-gram should be generated from.

EntityCategory

A string indicating what entity categories to return.

EntityRecognitionSkillLanguage

The language codes supported for input text by EntityRecognitionSkill.

ImageAnalysisSkillLanguage

The language codes supported for input by ImageAnalysisSkill.

ImageDetail

A string indicating which domain-specific details to return.

IndexProjectionMode

Defines behavior of the index projections in relation to the rest of the indexer.

IndexerExecutionEnvironment

Specifies the environment in which the indexer should execute.

IndexerExecutionStatus

Represents the status of an individual indexer execution.

IndexerResyncOption

Options with various types of permission data to index.

IndexerStatus

Represents the overall indexer status.

KeyPhraseExtractionSkillLanguage

The language codes supported for input text by KeyPhraseExtractionSkill.

KnowledgeBaseModelKind

The AI model to be used for query planning.

KnowledgeSourceContentExtractionMode

Optional content extraction mode. Default is 'minimal'.

KnowledgeSourceKind

The kind of the knowledge source.

KnowledgeSourceSynchronizationStatus

The current synchronization status of the knowledge source.

LexicalAnalyzerName

Defines the names of all text analyzers supported by the search engine.

LexicalNormalizerName

Defines the names of all text normalizers supported by the search engine.

LexicalTokenizerName

Defines the names of all tokenizers supported by the search engine.

MarkdownHeaderDepth

Specifies the max header depth that will be considered while grouping markdown content. Default is h6.

MarkdownParsingSubmode

Specifies the submode that will determine whether a markdown file will be parsed into exactly one search document or multiple search documents. Default is oneToMany.

MicrosoftStemmingTokenizerLanguage

Lists the languages supported by the Microsoft language stemming tokenizer.

MicrosoftTokenizerLanguage

Lists the languages supported by the Microsoft language tokenizer.

OcrLineEnding

Defines the sequence of characters to use between the lines of text recognized by the OCR skill. The default value is "space".

OcrSkillLanguage

The language codes supported for input by OcrSkill.

PIIDetectionSkillMaskingMode

A string indicating what maskingMode to use to mask the personal information detected in the input text.

PhoneticEncoder

Identifies the type of phonetic encoder to use with a PhoneticTokenFilter.

RankingOrder

Represents score to use for sort order of documents.

RegexFlags

Defines a regular expression flag that can be used in the pattern analyzer and pattern tokenizer.

ScoringFunctionAggregation

Defines the aggregation function used to combine the results of all the scoring functions in a scoring profile.

ScoringFunctionInterpolation

Defines the function used to interpolate score boosting across a range of documents.

SearchFieldDataType

Defines the data type of a field in a search index.

SearchIndexerDataSourceType

Defines the type of a datasource.

SentimentSkillLanguage

The language codes supported for input text by SentimentSkill.

SnowballTokenFilterLanguage

The language to use for a Snowball token filter.

SplitSkillLanguage

The language codes supported for input text by SplitSkill.

StemmerTokenFilterLanguage

The language to use for a stemmer token filter.

StopwordsList

Identifies a predefined list of language-specific stopwords.

TextSplitMode

A value indicating which split mode to perform.

TextTranslationSkillLanguage

The language codes supported for input text by TextTranslationSkill.

TokenCharacterKind

Represents classes of characters on which a token filter can operate.

TokenFilterName

Defines the names of all token filters supported by the search engine.

VectorEncodingFormat

The encoding format for interpreting vector field contents.

VectorSearchAlgorithmKind

The algorithm used for indexing and querying.

VectorSearchAlgorithmMetric

The similarity metric to use for vector comparisons. It is recommended to choose the same similarity metric as the embedding model was trained on.

VectorSearchCompressionKind

The compression method used for indexing and querying.

VectorSearchCompressionRescoreStorageMethod

The storage method for the original full-precision vectors used for rescoring and internal index operations.

VectorSearchCompressionTarget

The quantized data type of compressed vector values.

VectorSearchVectorizerKind

The vectorization method to be used during query time.

VisualFeature

The strings indicating what visual feature types to return.

Functions

ComplexField

Configure a Complex or Complex collection field for an Azure Search Index

ComplexField(*, name: str, collection: bool = False, fields: List[SearchField] | None = None, **kw) -> SearchField

Keyword-Only Parameters

Name Description
name
str

Required. The name of the field, which must be unique within the fields collection of the index or parent field.

collection

Whether this complex field is a collection (default False)

Default value: False
fields

A list of sub-fields

Default value: None

Returns

Type Description

The search field object.

SearchableField

Configure a searchable text field for an Azure Search Index

SearchableField(*, name: str, collection: bool = False, key: bool = False, hidden: bool = False, searchable: bool = True, filterable: bool = False, sortable: bool = False, facetable: bool = False, analyzer_name: str | LexicalAnalyzerName | None = None, search_analyzer_name: str | LexicalAnalyzerName | None = None, index_analyzer_name: str | LexicalAnalyzerName | None = None, synonym_map_names: List[str] | None = None, **kw) -> SearchField

Keyword-Only Parameters

Name Description
name
str

Required. The name of the field, which must be unique within the fields collection of the index or parent field.

collection

Whether this search field is a collection (default False)

Default value: False
key

A value indicating whether the field uniquely identifies documents in the index. Exactly one top-level field in each index must be chosen as the key field and it must be of type SearchFieldDataType.STRING. Key fields can be used to look up documents directly and update or delete specific documents. Default is False

Default value: False
hidden

A value indicating whether the field will be returned in a search result. Setting this to True is equivalent to setting retrievable to False. You can enable this option if you want to use a field (for example, margin) as a filter, sorting, or scoring mechanism but do not want the field to be visible to the end user. This property must be False for key fields. Default is False.

Default value: False
searchable

A value indicating whether the field is full-text searchable. This means it will undergo analysis such as word-breaking during indexing. If you set a searchable field to a value like "sunny day", internally it will be split into the individual tokens "sunny" and "day". This enables full-text searches for these terms. Note: searchable fields consume extra space in your index since Azure Cognitive Search will store an additional tokenized version of the field value for full-text searches. If you want to save space in your index and you don't need a field to be included in searches, set searchable to false. Default is True.

Default value: True
filterable

A value indicating whether to enable the field to be referenced in $filter queries. filterable differs from searchable in how strings are handled. Fields that are filterable do not undergo word-breaking, so comparisons are for exact matches only. For example, if you set such a field f to "sunny day", $filter=f eq 'sunny' will find no matches, but $filter=f eq 'sunny day' will. Default is False.

Default value: False
sortable

A value indicating whether to enable the field to be referenced in $orderby expressions. By default Azure Cognitive Search sorts results by score, but in many experiences users will want to sort by fields in the documents. The default is False.

Default value: False
facetable

A value indicating whether to enable the field to be referenced in facet queries. Typically used in a presentation of search results that includes hit count by category (for example, search for digital cameras and see hits by brand, by megapixels, by price, and so on). Default is False.

Default value: False
analyzer_name

The name of the analyzer to use for the field. This option can't be set together with either searchAnalyzer or indexAnalyzer. Once the analyzer is chosen, it cannot be changed for the field. Possible values include: 'ar.microsoft', 'ar.lucene', 'hy.lucene', 'bn.microsoft', 'eu.lucene', 'bg.microsoft', 'bg.lucene', 'ca.microsoft', 'ca.lucene', 'zh- Hans.microsoft', 'zh-Hans.lucene', 'zh-Hant.microsoft', 'zh-Hant.lucene', 'hr.microsoft', 'cs.microsoft', 'cs.lucene', 'da.microsoft', 'da.lucene', 'nl.microsoft', 'nl.lucene', 'en.microsoft', 'en.lucene', 'et.microsoft', 'fi.microsoft', 'fi.lucene', 'fr.microsoft', 'fr.lucene', 'gl.lucene', 'de.microsoft', 'de.lucene', 'el.microsoft', 'el.lucene', 'gu.microsoft', 'he.microsoft', 'hi.microsoft', 'hi.lucene', 'hu.microsoft', 'hu.lucene', 'is.microsoft', 'id.microsoft', 'id.lucene', 'ga.lucene', 'it.microsoft', 'it.lucene', 'ja.microsoft', 'ja.lucene', 'kn.microsoft', 'ko.microsoft', 'ko.lucene', 'lv.microsoft', 'lv.lucene', 'lt.microsoft', 'ml.microsoft', 'ms.microsoft', 'mr.microsoft', 'nb.microsoft', 'no.lucene', 'fa.lucene', 'pl.microsoft', 'pl.lucene', 'pt-BR.microsoft', 'pt-BR.lucene', 'pt- PT.microsoft', 'pt-PT.lucene', 'pa.microsoft', 'ro.microsoft', 'ro.lucene', 'ru.microsoft', 'ru.lucene', 'sr-cyrillic.microsoft', 'sr-latin.microsoft', 'sk.microsoft', 'sl.microsoft', 'es.microsoft', 'es.lucene', 'sv.microsoft', 'sv.lucene', 'ta.microsoft', 'te.microsoft', 'th.microsoft', 'th.lucene', 'tr.microsoft', 'tr.lucene', 'uk.microsoft', 'ur.microsoft', 'vi.microsoft', 'standard.lucene', 'standardasciifolding.lucene', 'keyword', 'pattern', 'simple', 'stop', 'whitespace'.

Default value: None
search_analyzer_name

The name of the analyzer used at search time for the field. It must be set together with indexAnalyzer and it cannot be set together with the analyzer option. This property cannot be set to the name of a language analyzer; use the analyzer property instead if you need a language analyzer. This analyzer can be updated on an existing field. Possible values include: 'ar.microsoft', 'ar.lucene', 'hy.lucene', 'bn.microsoft', 'eu.lucene', 'bg.microsoft', 'bg.lucene', 'ca.microsoft', 'ca.lucene', 'zh-Hans.microsoft', 'zh-Hans.lucene', 'zh- Hant.microsoft', 'zh-Hant.lucene', 'hr.microsoft', 'cs.microsoft', 'cs.lucene', 'da.microsoft', 'da.lucene', 'nl.microsoft', 'nl.lucene', 'en.microsoft', 'en.lucene', 'et.microsoft', 'fi.microsoft', 'fi.lucene', 'fr.microsoft', 'fr.lucene', 'gl.lucene', 'de.microsoft', 'de.lucene', 'el.microsoft', 'el.lucene', 'gu.microsoft', 'he.microsoft', 'hi.microsoft', 'hi.lucene', 'hu.microsoft', 'hu.lucene', 'is.microsoft', 'id.microsoft', 'id.lucene', 'ga.lucene', 'it.microsoft', 'it.lucene', 'ja.microsoft', 'ja.lucene', 'kn.microsoft', 'ko.microsoft', 'ko.lucene', 'lv.microsoft', 'lv.lucene', 'lt.microsoft', 'ml.microsoft', 'ms.microsoft', 'mr.microsoft', 'nb.microsoft', 'no.lucene', 'fa.lucene', 'pl.microsoft', 'pl.lucene', 'pt-BR.microsoft', 'pt-BR.lucene', 'pt-PT.microsoft', 'pt-PT.lucene', 'pa.microsoft', 'ro.microsoft', 'ro.lucene', 'ru.microsoft', 'ru.lucene', 'sr- cyrillic.microsoft', 'sr-latin.microsoft', 'sk.microsoft', 'sl.microsoft', 'es.microsoft', 'es.lucene', 'sv.microsoft', 'sv.lucene', 'ta.microsoft', 'te.microsoft', 'th.microsoft', 'th.lucene', 'tr.microsoft', 'tr.lucene', 'uk.microsoft', 'ur.microsoft', 'vi.microsoft', 'standard.lucene', 'standardasciifolding.lucene', 'keyword', 'pattern', 'simple', 'stop', 'whitespace'.

Default value: None
index_analyzer_name

The name of the analyzer used at indexing time for the field. It must be set together with searchAnalyzer and it cannot be set together with the analyzer option. This property cannot be set to the name of a language analyzer; use the analyzer property instead if you need a language analyzer. Once the analyzer is chosen, it cannot be changed for the field. Possible values include: 'ar.microsoft', 'ar.lucene', 'hy.lucene', 'bn.microsoft', 'eu.lucene', 'bg.microsoft', 'bg.lucene', 'ca.microsoft', 'ca.lucene', 'zh-Hans.microsoft', 'zh- Hans.lucene', 'zh-Hant.microsoft', 'zh-Hant.lucene', 'hr.microsoft', 'cs.microsoft', 'cs.lucene', 'da.microsoft', 'da.lucene', 'nl.microsoft', 'nl.lucene', 'en.microsoft', 'en.lucene', 'et.microsoft', 'fi.microsoft', 'fi.lucene', 'fr.microsoft', 'fr.lucene', 'gl.lucene', 'de.microsoft', 'de.lucene', 'el.microsoft', 'el.lucene', 'gu.microsoft', 'he.microsoft', 'hi.microsoft', 'hi.lucene', 'hu.microsoft', 'hu.lucene', 'is.microsoft', 'id.microsoft', 'id.lucene', 'ga.lucene', 'it.microsoft', 'it.lucene', 'ja.microsoft', 'ja.lucene', 'kn.microsoft', 'ko.microsoft', 'ko.lucene', 'lv.microsoft', 'lv.lucene', 'lt.microsoft', 'ml.microsoft', 'ms.microsoft', 'mr.microsoft', 'nb.microsoft', 'no.lucene', 'fa.lucene', 'pl.microsoft', 'pl.lucene', 'pt-BR.microsoft', 'pt-BR.lucene', 'pt-PT.microsoft', 'pt-PT.lucene', 'pa.microsoft', 'ro.microsoft', 'ro.lucene', 'ru.microsoft', 'ru.lucene', 'sr- cyrillic.microsoft', 'sr-latin.microsoft', 'sk.microsoft', 'sl.microsoft', 'es.microsoft', 'es.lucene', 'sv.microsoft', 'sv.lucene', 'ta.microsoft', 'te.microsoft', 'th.microsoft', 'th.lucene', 'tr.microsoft', 'tr.lucene', 'uk.microsoft', 'ur.microsoft', 'vi.microsoft', 'standard.lucene', 'standardasciifolding.lucene', 'keyword', 'pattern', 'simple', 'stop', 'whitespace'.

Default value: None
synonym_map_names

A list of the names of synonym maps to associate with this field. Currently only one synonym map per field is supported. Assigning a synonym map to a field ensures that query terms targeting that field are expanded at query-time using the rules in the synonym map. This attribute can be changed on existing fields.

Default value: None

Returns

Type Description

The search field object.

SimpleField

Configure a simple field for an Azure Search Index

SimpleField(*, name: str, type: str | SearchFieldDataType, key: bool = False, hidden: bool = False, filterable: bool = False, sortable: bool = False, facetable: bool = False, **kw) -> SearchField

Keyword-Only Parameters

Name Description
name
str

Required. The name of the field, which must be unique within the fields collection of the index or parent field.

type

Required. The data type of the field. Possible values include: SearchFieldDataType.STRING, SearchFieldDataType.INT32, SearchFieldDataType.INT64, SearchFieldDataType.DOUBLE, SearchFieldDataType.BOOLEAN, SearchFieldDataType.DATETIMEOFFSET, SearchFieldDataType.GEOGRAPHY_POINT, SearchFieldDataType.COMPLEXTYPE, from azure.search.documents.SearchFieldDataType.

key

A value indicating whether the field uniquely identifies documents in the index. Exactly one top-level field in each index must be chosen as the key field and it must be of type SearchFieldDataType.STRING. Key fields can be used to look up documents directly and update or delete specific documents. Default is False

Default value: False
hidden

A value indicating whether the field will be returned in a search result. Setting this to True is equivalent to setting retrievable to False. You can enable this option if you want to use a field (for example, margin) as a filter, sorting, or scoring mechanism but do not want the field to be visible to the end user. This property must be False for key fields. Default is False.

Default value: False
filterable

A value indicating whether to enable the field to be referenced in $filter queries. filterable differs from searchable in how strings are handled. Fields of type SearchFieldDataType.STRING or Collection(SearchFieldDataType.STRING) that are filterable do not undergo word-breaking, so comparisons are for exact matches only. For example, if you set such a field f to "sunny day", $filter=f eq 'sunny' will find no matches, but $filter=f eq 'sunny day' will. This property must be null for complex fields. Default is False

Default value: False
sortable

A value indicating whether to enable the field to be referenced in $orderby expressions. By default Azure Cognitive Search sorts results by score, but in many experiences users will want to sort by fields in the documents. A simple field can be sortable only if it is single-valued (it has a single value in the scope of the parent document). Simple collection fields cannot be sortable, since they are multi-valued. Simple sub-fields of complex collections are also multi-valued, and therefore cannot be sortable. This is true whether it's an immediate parent field, or an ancestor field, that's the complex collection. The default is False.

Default value: False
facetable

A value indicating whether to enable the field to be referenced in facet queries. Typically used in a presentation of search results that includes hit count by category (for example, search for digital cameras and see hits by brand, by megapixels, by price, and so on). Fields of type SearchFieldDataType.GEOGRAPHY_POINT or Collection(SearchFieldDataType.GEOGRAPHY_POINT) cannot be facetable. Default is False.

Default value: False

Returns

Type Description

The search field object.