Quenstion About CNV(Custom Neural Voice) Limit
Hi I have a question about the CNV concurrent request limit. (https://video2.skills-academy.com/ko-kr/azure/cognitive-services/speech-service/speech-services-quotas-and-limits) Is the limit for CNV concurrent request(10) is per Speech resource or…
Two times with a pause for each sentence
I am an ESL learner. I have a list of 400++ English sentences. I want to make it audible but two times for each sentence: One at a 70% speed and another at a 100% speed and with a certain time of pause between one after another . As for the creation of…
cognitive-services-speech-sdk-go/samples crashed
install Azure Speech go sdk with follow commands: yum -y update yum -y groupinstall "Development tools" yum -y install alsa-lib openssl wget export SPEECHSDK_ROOT="$HOME/speechsdk" mkdir -p…
Custom Voice Missing From Custom Command
In Speech Studio a month or two ago there was a section in settings to define the custom voice it would respond with when using the custom command interface. This has now disappeared in the UI. Is this still supported? …
Is this working right now ?
Hello, I just created a Custom commands project, train, test, and nothing work. No response at all. I also have a 401 error using the Windows Voice Assistant sample exe. Is it down right now ?
Custom Speech API 3.0 does not support Upload of new StructuredText Dataset
The API manual does not indicate what "kind" should be used when uploading structured text (in Markdown format), which is introduced in…
Realtime Speech To Text Performance
We are using the basicRecognizer.StartContinuousRecognitionAsync() and a PushAudioInputStream to push an audio input stream from a mic and then output the results to a text editor. I am looking to see if there is a setting to break the results into 2-4…
use speech to text sdk When using Bluetooth headset
1.During meetings using teams with Bluetooth headset。 2.Open my application, which uses speech to text SDK. 3.What I expect is that the voice I say and the voice I hear can be translated into text。 But the actual effect is that only my voice is…
Phoneme Support
Does Speech to Text Provide Phoneme Support @phoneme @Anonymous eg What is the first sound in CAT (Answer- /k/) 2nd. What is last sound in Bottom (Answer- /tum/) If yes please share me the resource so i can also check the same. …
Azure Speechsdk on Raspberry Pi
I've been working on a project for the better part of a month now, and to no end I keep hitting road blocks. I am trying to make an automatic translator on my Raspberry Pi 4 that uses Bluetooth headphones as the audio input and output, preferably in…
Audio content creation: access denied to existing speech resource
I want to create a model in speech.microsoft.com and fail with a speech resource, where I have contributor access only (no access to the hosting RG of that resource). I can only work with a resource I created in the speech portal. In the browser I am…
Azure TTS in webapp works in debug mode but not publish
I have created a basic webapp in which when a button is clicked, a TTS voice is playing. The app works fine in debug mode but not in browser. I know this is quite a common issue but I never find a way to solve it or another possibility to implement the…
Can we use audio content created in Speech Studio for public videos without restrictions?
Hello, We would like to create audio content with Speech Studio and embed it in our videos, which we distribute via social media and our homepage, for example. Can we do this without restrictions or do we have to purchase separate licences for this? …
How do I use the Nbest function?
Now I want to use Python to see the voice recognizer's expected words. However, just one word is being broadcast continually, and no other projected word can be observed at this time. So when I looked up the function, I saw NBest and wanted to…
The function enquiry of voice recognizer Nbest.
Hello, I'm a student that is interested in speech recognition. The "Nbest" function is explained, but I'm wondering because I haven't seen the specific example. In the case of Google, I am aware of various anticipated terms. Is there,…
Azure Speech SDK: How to specify the NBest length?
Hello, i'm using speech-to-test sdk to recognize audio to text. Current json result seems returning 5 NBest candidates, but how can i specify how many Nbest candidates to receive ? Or full NBest list ? Thank you.
speech to text (realtime)
I need help. I want to recognize real-time speech and see a list of predicted words. So, I want to apply a function called NBest to Python, but it doesn't work properly. I would appreciate it if someone could tell me the problem with the simple code…
Improving TTS latency
Hi, I'm using Azure TTS in my Android application. I use 2 objects in my implementation which are SpeechConfig and SpeechSynthesizer. My app is mainly based on a dynamic TTS (e.g. reading input back to the user, which happens very often) but the…
Speech Studio and service unavailable
In October 29th 2021 the portal speech.microsoft.com (custom commands) went unavailable also the service. Wondering which was the cause, this is something that happened before and it could be a problem when my project goes online. In the…
Speech recognition engine has difficulties recognizing separate letters (nl_NL)
I have noticed that when pronouncing separate letters in a test set, the engine has a lot of difficulty to recognize them (Dutch language). Of course this is not a big surprise, since there is no meaningful context, and the phoneme clusterse are…