speech-to-text

How to get Word Level Timestamps using Azure Speech to Text and the Python SDK?

柔情痞子 提交于 2020-05-15 08:25:07
问题 My code currently is able to read an audio file and transcribe it using Azure Speech to Text, with help from an example that I found on GitHub. However, I need to include the timestamps for all the words in the transcription. According to the documentation, this functionality was added in version 1.5.0, and is accessed through the method request_word_level_timestamps(). But even when I have called it, I get the same response as before. I cannot figure out how to use it from the documentation.

Using gcloud speech api for real-time speech recognition in dart, flutter

↘锁芯ラ 提交于 2020-05-13 08:00:28
问题 I want to use Google's real-time speech recognition api in a flutter project, written in dart. I've activated a gcloud account, created the api key (which should be the only necessary authentication method for google speech) and written a basic apk which ought to send an audio stream to Google cloud and display the response. I imported the googleapis/speech and googleapis_auth plugin. But I couldn't figure out how to set it up. They say you have to use gRPC, which makes sense as it should

Using gcloud speech api for real-time speech recognition in dart, flutter

给你一囗甜甜゛ 提交于 2020-05-13 07:59:05
问题 I want to use Google's real-time speech recognition api in a flutter project, written in dart. I've activated a gcloud account, created the api key (which should be the only necessary authentication method for google speech) and written a basic apk which ought to send an audio stream to Google cloud and display the response. I imported the googleapis/speech and googleapis_auth plugin. But I couldn't figure out how to set it up. They say you have to use gRPC, which makes sense as it should

com.google.api.gax.rpc.UnavailableException: io.grpc.StatusRuntimeException: UNAVAILABLE: Credentials failed to obtain metadata :cloud speech

雨燕双飞 提交于 2020-03-25 16:36:06
问题 i've done a speech to text configuration using google's cloud speech api in java, this works on other machine but the same setup does not work on my machine. i've installed google cloud platform tools from eclipse market place also i've set the environment credential variable for run time. can anyone please help ? 来源: https://stackoverflow.com/questions/59845869/com-google-api-gax-rpc-unavailableexception-io-grpc-statusruntimeexception-una

IBM Watson Speech-to-Text “unable to transcode data stream audio/webm -> audio/x-float-array” media MIME types

拜拜、爱过 提交于 2020-03-25 12:30:42
问题 I'm recording short audio files (a few seconds) in Chrome using mediaDevices.getUserMedia() , saving the file to Firebase Storage, and then trying to send the files to IBM Watson Speech-to-Text. I'm getting back this error message: unable to transcode data stream audio/webm -> audio/x-float-array In the browser I set up the microphone: navigator.mediaDevices.getUserMedia({ audio: true, video: false }) .then(stream => { var options = { audioBitsPerSecond : 128000, mimeType : 'audio/webm' };

Voice/Speech to text [closed]

断了今生、忘了曾经 提交于 2020-03-12 08:31:20
问题 Closed. This question is off-topic. It is not currently accepting answers. Want to improve this question? Update the question so it's on-topic for Stack Overflow. Closed 3 years ago . I need an API or library (preferably free) that will convert voice/speech through a microphone, into text (string). Additionally, I will need an API or library that can do text-to-speech. I'd like to use C# and .NET, but other languages will suffice. Thanks. 回答1: You can use CMU Sphinx as it is pretty open and

Android extras about speech recognition does not work

末鹿安然 提交于 2020-02-11 06:09:15
问题 I am currently busy with speech recognition, and I want speech recognizer to listen me continously, forever. The purpose of the app is, it is going to wait for speech of a specific keyword: when the user says this keyword, the service becomes ready to receive user voice commands After research I found that we can manupulate the time that the recognizer listens. By these constants: By the way the default recognizer service listens about 10 seconds. (The google api) String EXTRA_SPEECH_INPUT

Android extras about speech recognition does not work

那年仲夏 提交于 2020-02-11 06:08:19
问题 I am currently busy with speech recognition, and I want speech recognizer to listen me continously, forever. The purpose of the app is, it is going to wait for speech of a specific keyword: when the user says this keyword, the service becomes ready to receive user voice commands After research I found that we can manupulate the time that the recognizer listens. By these constants: By the way the default recognizer service listens about 10 seconds. (The google api) String EXTRA_SPEECH_INPUT

Android Custom Keyboard with SpeechRecognizer

自闭症网瘾萝莉.ら 提交于 2020-01-24 18:49:06
问题 I have a fully functional custom android keyboard in which i have to add speech recognition. Here are the relevant parts of the implementation i have public class CustomInputMethodService extends InputMethodService implements <random stuff> { private SpeechRecognizer mSpeechRecognizer; private RecognitionListener mSpeechlistener; public void onCreate() { super.onCreate(); mSpeechRecognizer = SpeechRecognizer.createSpeechRecognizer(this); mSpeechlistener = new CustomRecognitionListener();

How to setQuery from voice search

送分小仙女□ 提交于 2020-01-16 09:44:26
问题 I'm trying to inject the transcribed text in the search query in order to filter my adapter list, but, after trying every combination I could imagine, I come here to ask for help. For instance, if handle the intent after a new intent and set text to a TextView everything works fine and the voice recorder transcribes what is said correctly, but I'm not being able to get the same String and set the SearchView query successfully. The keyboard microphone, otherwise, captures the text and pass it