how to assign operation permissions a resources
Hello, I am new to Azure and I want to use it to convert text to speech. when I creat the resources -> enter the speech studio and try to start the service, the system raised an error and say "You don't have operation permissions to [New],…
Speech-to-Text batch transcribe API in germanycentralwest doesn't work
Last Friday (May 31 2024) we started getting the following errors on all transcripts sent to the batch transcription API on our speech resource in…
What are the HW or sound limitations for the echo cancellation algorithm in SpeechSDK
hi, I'm having some issues with the echo cancellation on my device, and I'm trying to use speech SDK, when I was analyzing the sounds that I record with microphone it seems that there are present higher harmonics which are 24dB less then primary…
create a basic voice-interactive dashboard
Hello Team, I need to create a basic voice-interactive dashboard using Azure Cognitive services like, Speech service, CLU(Conversational Language Understanding) & PowerBI.Also suggest if any other way to achieve this. It would be really helpful.
SpeakSsmlAsync is cancelled, but SpeakTextAsync is successfull
I am trying out the Azure AI service to convert text to speech from a C# WPF application. My calls through SpeakTextAsync are successfull, but my calls through SpeakSsmlAsync are returned with the Reason = Cancelled. I am on the free tier for South…
Azure Text To Speech docker container throws an exception with viseme
I'm using the Azure Text to Speech docker image (mcr.microsoft.com/azure-cognitive-services/speechservices/neural-text-to-speech:3.3.0-amd64-en-us-jennyneural). I'm passing it SSML through the dotnet SDK. When asking for viseme (via <mstts:viseme…
As a student how can I use Azure Speech resource
I have a student subscription and want to create an Azure Speech resource, but there's a problem. Is it because of the student subscription limitation or what I can do to use Azure speech service?
how can I set the permission to the resources
Hello, I want to upload a text file to Speech Studio, but the system raised an error Does anyone help how I can fix this and assign a proper role for myself? I already set my role as a Cognitive Services User.
Azure Cognitive Services Speech: Unable to get Custom Translator model results from speech translation code
In test C# code that I created based on the speech translation code in the following sample (“Using custom translation in speech translation”), I’m having trouble getting Custom Translator model translation results. The code just returns a cancellation…
Why does Viseme Callback no longer deliver blendshapes?
Hi there, A few months ago, I developed a project based on Speech Synthesis and Viseme Blendshapes. At that time, I had no problems receiving the blendshapes. Recently I realized that my code was no longer working. I receive Viseme with the corresponding…
"504.0 GatewayTimeout" On Deployed Flask Web Application's do_task() function
My Flask app is currently hosted on Azure and was deployed through docker hub. Once a form is submitted the do_task() function is called. This function takes a long time as it is using WhisperX. The program runs without problem locally but Azure's 240s…
How to create a dataset for Azure custom speech using spx (speechCLI)
I am using the following command for creating a custom speech dataset in my Azure Speech service: spx csr dataset create --api-version v3.1 --kind "Acoustic" --name "My Custom Speech" --description "My Acoustic Dataset…
Having issues when uploading daatset for custom STT model: Status: 400. The value of dataset.Locale is an invalid culture
I am uploading dataset in speech studio to train custom STT model. I have a .zip file which has audios in wav format and their trans. After uploading I am receiving this message: "Failed to upload data uch-ildiz. Error: Status: 400. The value of…
Is there a way for speech diarization to run across multiple files while keeping the speaker IDs the same for each speaker?
Let's say I have 5 large wav files of the same 4 speakers. The files are too large to concatenate into one wav file. Is there a way I can run diarization on these 5 files and keep the same speaker number for the respective voice across all files?
SPEECH DELAY, Length of created audio from SSML doesnt match my SSML file
Hello, i am using azure congnite services API where i try to transform SSML into audio which work well expect my sentences doesnt start at the right time for example my last sentence "Do hrnca dáme bravčovú masť?" should start at 35290ms but…
What is expected date to have AUDIO_INPUT_PROCESSING_ENABLE_V2 feature on Linux systems?
I'm having some issues with the current echo cancellation algorithm on Linux system, and I've noticed that there is enhanced version. Is there any date specified when this feature will be available on Linux systems? …
Currently, I know that STT is supported on mobile devices when the network is good. I would like to know if STT can still be used without a network connection.
Currently, I know that STT is supported on mobile devices when the network is good. I would like to know if STT can still be used without a network connection.
Transcription result returns Access denied due to invalid subscription key or wrong API endpoint?
I'm trying to run a batch transcription based on the example here. I have extracted my subscription key via the screenshot attached: and am trying to run the code via: import requests import json subscription_key = "<KEY VIA…
How to get audio file with the method recognizeOnceAsync of SpeechSDK.SpeechRecognizer?
I'm using the demo code from https://github.com/Azure-Samples/cognitive-services-speech-sdk/blob/master/quickstart/javascript/browser/from-microphone/index.html like this below recognizer.recognizing = (s, e) => { console.log(RECOGNIZING:…
Random Words Detected by Azure Speech Recognizer in Silence
Hello Azure Support Team, I am currently using the Azure Speech Service to recognize speech inputs in my application. The setup of my speech recognizer is as follows: export const createSpeechRecognizer = () => { const speechRecognitionConfig =…