Issue with Ryan Multilingual Voice – Accent Changing from American to Australian
Experiencing an issue where the Ryan Multilingual Voice unexpectedly changes its accent from American to Australian during audio generation. Our bot language is set to English - US, and am are consistently using Ryan Multilingual Voice. We have tested…
How to fix destination container url folder in Azure Batch Transcription?
import os import datetime import json import time import requests from azure.storage.blob import BlobServiceClient ======================= Configuration Settings ======================= Provided container URL for audio files…
Japanese Voice-to-Text: Preventing Unwanted Kanji Transcription for Names
When using Azure Speech to Text for batch transcription of conversations in Japanese, there is an issue with person names being transcribed into incorrect Kanji characters. A custom speech model has been created to handle specific industry terms, but…
![](https://techprofile.blob.core.windows.net/images/4d92d0b699a24552b31553f4fb1d8d06.jpg)
How to solve problem of a failed speech model
I'm building a speech model on Azure. The data for training was successfully uploaded, but the model that uses the data failed with an "Internal error" message. Pls tell me what to do. Thanks! Elise
Features container Speech to text
Hello community, I would like to ask you some info about Speech to text containers. I would like to run the service on premise and I don't find any information in the documentation in these services are available in the version on-prem as well: …
How to get omissions and insertions from scripted assessments in streaming mode
Hi, I'm trying to use Pronunciation Assessment to assess audio files that are between 60 and 120 seconds. This is in csharp dotnet, with the Azure Speech Services SDK and en-GB. Because of the time limit on the non-streaming method…
Difference in results for Batch speech to text in speech studio and API
Hi, i was running a dual channel hindi audio file on speech studio as well as api on batch speech to text, but i am getting different results in both, and it is consistent. The speech studio version is the correct one as it recognises the phrases…
Which Azure Speech SDK Feature to Use for Real-Time Meeting Transcription with Speaker Diarization?
Hi , I am working on real-time meeting transcription using Azure Speech SDK and need: Accurate speaker diarization (identify who is speaking). Sentence-level segmentation (avoid merging multiple sentences into one recognition event). Improved…
Running Web Avatar Code Sample but with Audio-Only mode
I've been testing Microsoft's Web/Avatar code sample (cognitive-services-speech-sdk/samples/python/web/avatar at master · Azure-Samples/cognitive-services-speech-sdk) and successfully got it running. However, the pricing for the speech service with an…
Voice Activation/Assistant questions on Windows
Hi MSFT team, I want to ask, If a ATL Service(.exe) is a UWP App ? to use voice assistant features on Windows. As we know, only UWP app can use voice activation/assistant. when voice activation return keyword detected, how can I get SignalId or…
My webhook throws a BadHttpRequestException "Unexpected end of request content."
I have a webhook that receives updates from Azure AI Speech when a batch transcription is complete and I am regularly getting Microsoft.AspNetCore.Server.Kestrel.Core.BadHttpRequestException "Unexpected end of request content.". My function app…
How to add a custom id for a single audio file in a speech to text batch process?
Hi, i am processing multiple audio files using azure speech to text batch process using contentUrls, So i have lets have 5 public urls, and when i get the result of these 5 i get it as "contenturl_0.json", "contenturl_1.json",2,3,4.…
How to find the associated audio recording for a recognized sentence (RecongitionResult.ResultId?)
The audio recording has some meta data, but I am unable to correlate it with the RecognitionResult's ResultId. Would you be able to guide me on how to link the RecognitionResult with the audio logs? Thank you! { "self":…
Looking for a method in order to specify usage per user using 1 subscription key per organization for speech-to-text
Hi, We are right in the middle of developing a reporting application for health care use. Average use will be 20 concurrent users using AI Speech-to_text per customer/organization. Technically, everything seems to be working fine. However, when using 1…
Are TurboMultilingualNeural voices production ready?
Hi Microsofties, A conflict in voice status. These voices are listed as "public preview" in web…
WebSocket Authentication Error (401) When Using Private Endpoint for Azure Speech Services
Hello, I am experiencing an authentication issue when attempting to use Azure Cognitive Services Speech-to-Text with a Private Endpoint from an Azure App Service. Below are the details of my configuration and the error message I am receiving. Issue…
Speech to text costing $1/hr is crazy. No bulk available.
Hey, We are running an app that takes notes for doctors. Paying $1 per hr of speech to text is bankrupting our company. How come Azure doesn't offer any subscriptions or bulk discounts?
Azure Bot that joins Microsoft Teams Call and transfers speech to text using Azure Speech Service
I configured Azure Bot which is configured to join Microsoft Teams calls, and calling endpoint is provided. Now using .NET I have implementation for calling endpoint and also the speech service starts converting speech to text when the call is answered.…
Does Custom Neural Voice support Viseme output?
I could not find any documentation on the SSML supported by custom neural voice. I would like to know if Viseme events are supported.
[ISSUE] Pronunciation Assessment in Polish Fails
Hello, the Azure Team, I have noticed audio passed to the Azure Pronunciation Assessment in Polish doesn't work at all. Uploaded audio and even the live microphone in your Azure Speech dashboard fail in Polish. I haven't tested all the languages but the…