speech-to-text

Dialogflow email address from speech

a 夏天 提交于 2021-01-29 14:20:42
问题 Does anyone have any suggestions for obtaining a user’s email address through speech? Written is quite straight forward as email addresses follow a pattern to some degree, but using speech is quite difficult. Is it best to simply ask the user to read out the characters one by one? 回答1: Dialogflow provides System entites for most common user inputs. You can use sys.email entity for your purpose and then use it in your fulfillment. Getting the above email address in your webhook fulfillment :

The Application Default Credentials are not available

僤鯓⒐⒋嵵緔 提交于 2021-01-05 10:56:06
问题 I am trying to use speech-to-text API of Google Cloud Platform for my Android App. I have passed it a recorded audio file for conversion to text. I can't solve an IOException which is described as " The Application Default Credentials are not available. They are available if running in Google Compute Engine. Otherwise, the environment variable GOOGLE_APPLICATION_CREDENTIALS must be defined pointing to a file defining the credentials. See https://developers.google.com/accounts/docs/application

How to send live audio stream to IBM Watson Speech-To-Text immediately without saving audio file locally using C# and IBM Watson SDK?

本秂侑毒 提交于 2020-11-29 19:12:57
问题 I have a requirement to use IBM Watson SDK to record the audio using microphone and send it to IBM Watson speech-to-text using C#. I am able to achieve this functionality by saving the audio file locally and then sending it using NAudio library. But my requirement is to use streaming mode to send live audio to IBM Watson Speech-to-Text service without storing the audio file physically. I am not able to find RecognizeUsingWebSocket service in the SDK. I am able to find only Recognize service.

Stream audio to Azure speech api by node.js on browser

混江龙づ霸主 提交于 2020-07-22 10:15:05
问题 I'm making a demo of speech to text using Azure speech api on browser by node.js. According to API document here, it does specify that it need .wav or .ogg files. But the example down there does a api call through sending byte data to api. So I've already get my data from microphone in byte array form. Is it the right path to convert it to byte and send it to api? Or is it better for me to save it as a .wav file then send to the api? So below is my code. This is stream from microphone part.

Python SpeechRecognition word by word? continuous output?

ε祈祈猫儿з 提交于 2020-06-27 17:28:11
问题 I was wondering whether there is a way to output words as soon as possible. For example if I say "hello world" it should output: hello world Currently I'm using this code import speech_recognition as sr r = sr.Recognizer() with sr.Microphone() as source: while True: r.pause_threshold=0.1 ##i tried playing with these 3 but no luck r.phrase_threshold=0.5 r.non_speaking_duration=0.1 audio = r.listen(source) try: text = r.recognize_google(audio) print(text) except Exception as e: print("-") What

Long audio speech recognition on Android

一笑奈何 提交于 2020-06-11 06:02:08
问题 I want to develop a module which will use a speech to text support in Android. I found out many documentation and demos related to RecognizerIntent and such others. But I found that all of such demos just fetch the voice till 10 secs or so. But I want my demo to run for more than 5-10 minutes. I don't have any issue if that is not running offline, as my app is always working online. I have also looked in to Pocketsphinx on Android, but that didn't worked out well. Also, that gave support just

Are IBM Watson IAM tokens good for all services or specific to each service, e.g., Speech-to-Text?

走远了吗. 提交于 2020-05-17 03:33:04
问题 IBM's documentation says that the following Node back end code enables you to Use the API key to have the SDK manage the lifecycle of the token. The SDK requests an access token, ensures that the access token is valid, and refreshes it if necessary. const SpeechToTextV1 = require('ibm-watson/speech-to-text/v1'); const { IamAuthenticator } = require('ibm-watson/auth'); const speechToText = new SpeechToTextV1({ authenticator: new IamAuthenticator({ apikey: '{apikey}', }), url: '{url}', }); How