AI services in Fabric (preview)

Important

This feature is in preview.

Azure AI services help developers and organizations rapidly create intelligent, cutting-edge, market-ready, and responsible applications with prebuilt and customizable APIs and models. Formerly named Azure Cognitive Services, Azure AI services empower developers even when they don't have direct AI or data science skills or knowledge. The goal of Azure AI services is to help developers create applications that can see, hear, speak, understand, and even begin to reason.

Fabric provides two options to use Azure AI services:

  • Pre-built AI models in Fabric (preview)

    Fabric seamlessly integrates with Azure AI services, allowing you to enrich your data with prebuilt AI models without any prerequisite. We recommend this option because you can use your Fabric authentication to access AI services, and all usages are billed against your Fabric capacity. This option is currently in public preview, with limited AI services available.

    Fabric offers Azure OpenAI Service, Text Analytics, and Azure AI Translator by default, with support for both SynapseML and the RESTful API. You can also use the OpenAI Python Library to access Azure OpenAI service in Fabric. For more information about available models, visit prebuilt AI models in Fabric.

  • Bring your own key (BYOK)

    You can provision your AI services on Azure, and bring your own key to use them from Fabric. If the prebuilt AI models don't yet support the desired AI services, you can still use BYOK (Bring your own key).

    To learn more about how to use Azure AI services with BYOK, visit Azure AI services in SynapseML with bring your own key.

Prebuilt AI models in Fabric (preview)

Note

Prebuilt AI models are currently available in preview and offered for free, with a limit on the number of concurrent requests per user. For Open AI models, the limit is 20 requests per minute per user.

Azure OpenAI Service

REST API, Python SDK. SynapseML

  • GPT-35-turbo: GPT-3.5 models can understand and generate natural language or code. The most capable and cost effective model in the GPT-3.5 family is GPT-3. The 5 Turbo option, which is optimized for chat, works well for traditional completion tasks as well. The gpt-35-turbo-0125 model supports up to 16,385 input tokens and 4,096 output tokens.
  • gpt-4 family: gpt-4-32k is supported.
  • text-embedding-ada-002 (version 2), embedding model that can be used with embedding API requests. The maximum accepted request token is 8,191, and the returned vector has dimensions of 1,536.

Text Analytics

REST API, SynapseML

  • Language detection: detects language of the input text
  • Sentiment analysis: returns a score between 0 and 1, to indicate the sentiment in the input text
  • Key phrase extraction: identifies the key talking points in the input text
  • Personally Identifiable Information(PII) entity recognition: identify, categorize, and redact sensitive information in the input text
  • Named entity recognition: identifies known entities and general named entities in the input text
  • Entity linking: identifies and disambiguates the identity of entities found in text

Azure AI Translator

REST API, SynapseML

  • Translate: Translates text
  • Transliterate: Converts text in one language, in one script, to another script.

Available regions

Available regions for Azure OpenAI Service

For the list of Azure regions where prebuilt AI services in Fabric are now available, visit the Available regions section of the Overview of Copilot in Fabric and Power BI (preview) article.

Available regions for Text Analytics and Azure AI Translator

Prebuilt Text Analytics and the Azure AI Translator in Fabric are now available for public preview in the Azure regions listed in this article. If you don't find your Microsoft Fabric home region in this article, you can still create a Microsoft Fabric capacity in a supported region. For more information, visit Buy a Microsoft Fabric subscription. To determine your Fabric home region, visit Find your Fabric home region.

Asia Pacific Europe Americas Middle East and Africa
Australia East North Europe Brazil South South Africa North
Australia Southeast West Europe Canada Central UAE North
Central Indian France Central Canada East
East Asia Norway East East US
Japan East Switzerland North East US 2
Korea Central Switzerland West North Central US
Southeast Asia UK South South Central US
South India UK West West US
West US 2
West US 3

Consumption rate

Note

The billing for prebuilt AI services in Fabric became effective on November 1st, 2024, as part of your existing Power BI Premium or Fabric Capacity.

A request for prebuilt AI services consumes Fabric Capacity Units. This table defines how many capacity units (CU) are consumed when an AI service is used.

Consumption rate for OpenAI language models

Models Context Input (Per 1,000 Tokens) Output (Per 1,000 Tokens)
GPT-4o-2024-08-06 Global Deployment 128 K 84.03 CU seconds 336.13 CU seconds
GPT-4 32 K 2,016.81 CU seconds 4,033.61 CU seconds
GPT-3.5-Turbo-0125 16K 16.81 CU seconds 50.42 CU seconds

Consumption rate for OpenAI embedding models

Models Operation Unit of Measure Consumption rate
text-embedding-ada-002 1,000 Tokens 3.36 CU seconds

Consumption rate for Text Analytics

Operation Operation Unit of Measure Consumption rate
Language Detection 1,000 text records 33,613.45 CU seconds
Sentiment Analysis 1,000 text records 33,613.45 CU seconds
Key Phrase Extraction 1,000 text records 33,613.45 CU seconds
Personally Identifying Information Entity Recognition 1,000 text records 33,613.45 CU seconds
Named Entity Recognition 1,000 text records 33,613.45 CU seconds
Entity Linking 1,000 text records 33,613.45 CU seconds
Summarization 1,000 text records 67,226.89 CU seconds

Consumption rate for Text Translator

Operation Operation Unit of Measure Consumption rate
Translate 1M Characters 336,134.45 CU seconds
Transliterate 1M Characters 336,134.45 CU seconds

Changes to AI services in Fabric consumption rate

Consumption rates are subject to change at any time. Microsoft uses reasonable efforts to provide notice via email or through in-product notification. Changes shall be effective on the date stated in the Microsoft Release Notes or the Microsoft Fabric Blog. If any change to a AI service in Fabric Consumption Rate materially increases the Capacity Units (CU) required to use, customers can use the cancellation options available for the chosen payment method.

Monitor the Usage

The workload meter associated with the task determines the charges for prebuilt AI services in Fabric. For example, if AI service usage is derived from a Spark workload, the AI usage is grouped together and billed under the Spark billing meter on Fabric Capacity Metrics app.

Example

An online shop owner uses SynapseML and Spark to categorize millions of products into relevant categories. Currently, the shop owner applies hard-coded logic to clean and map the raw "product type" to categories. However, the owner plans to switch to use of the new native Fabric OpenAI LLM (Large Language Model) endpoints. This iteratively processes the data against an LLM for each row, and then categorizes the products based on their "product name," "description," "technical details," and so on.

The expected cost for Spark usage is 1000 CUs. The expected cost for OpenAI usage is about 300 CUs.

To test the new logic, first iterate it in a Spark notebook interactive run. For the operation name of the run, use "Notebook Interactive Run." The owner expects to see an all-up usage of 1300 CUs under "Notebook Interactive Run," with the Spark billing meter accounting for the entire usage.​

Once the shop owner validates the logic, the owner sets up the regular run and expects to see an all-up usage of 1300 CUs under the operation name "Spark Job Scheduled Run," with the Spark billing meter accounting for the entire usage.​