speech-recognition

TTS *and* Speech Input simultaneously?

倾然丶 夕夏残阳落幕 提交于 2019-12-24 03:25:23
问题 I noticed that as soon as a voice recognition activity starts, text-to-speech output stops. I understand the rational: TTS output could be "heard" by the voice recognition engine and interfere with its proper operation. My question: Is this behavior hard-coded into the system, or can it be modified by a setting or parameter (in the API)? 回答1: Must the activity simultaneously use recognition and TTS? If the recognition can wait (functionally speaking), force the event to spawn the

can speech recognizer take input from incoming call voice (Speaker)?

▼魔方 西西 提交于 2019-12-24 01:47:14
问题 During a phone call application listens to the incoming phone call i.e incoming voice from speaker not the mic. Converts the caller's(the person who has make a call to you) voice to text. Finally display that converted text into the text view. It does all this conversion while the phone call is happening. 来源: https://stackoverflow.com/questions/42521023/can-speech-recognizer-take-input-from-incoming-call-voice-speaker

How to execute both Speech Recognition and Audio recording at the same time?

扶醉桌前 提交于 2019-12-24 00:58:40
问题 THIS IS NOT A DUPLICATE, HERE I TALK ABOUT STOPPING AND STARTING THE RECORD PROCESS WHENEVER I WANT. BEFORE MARKING AS DUPLICATES, PLEASE READ THE OTHER ANSWER PROPERLY. I am developing a Phonegap plugin for Android . This plugin will basically support the Android Speech Recognition and recording the speech. I am capable of starting, stopping etc the speech recognition, but got serious issues with recording. First, the code is posted below. Below is how I start the speech, end it etc. public

CMU-Sphinx : NullPointerException at recognizer.allocate()

霸气de小男生 提交于 2019-12-24 00:48:27
问题 I've been trying to set up CMU-Sphinx4 on my PC in order to make an application using speech Recognition. I was able to setup most of the parts and configuration and trying to run the HelloWorld application provided by Sphinx, but stuck with NullPointerExcetption at recognizer.allocate(); . Complete Stack Trace : 12:34:45.501 WARNING dictionary Missing word: <sil> 12:34:45.517 WARNING jsgfGrammar Can't find pronunciation for <sil> 12:34:45.517 WARNING dictionary Missing word: <sil> 12:34:45

SpeechRecognition and SpeechSynthesis in TypeScript

孤人 提交于 2019-12-23 19:23:39
问题 I was able to run SpeechRecognition in TypeScript by creating interface as below, and it is working fine: namespace CORE{ export interface IWindow extends Window{ webkitSpeechRecognition: any; } } I tried to use the same way for SpeechSynthesis, but field, and the below code did not work: namespace CORE{ export interface IWindow extends Window{ SpeechSynthesisUtterance: any; speechSynthesis: any; } } my questions are: Is the way i used to define the SpeechRecognition is the best practice to

SpeechRecognizer: no selected voice recognition service

一笑奈何 提交于 2019-12-23 12:53:58
问题 This is how i start my RecogniseListener intent: Intent intent = new Intent(RecognizerIntent.ACTION_RECOGNIZE_SPEECH); intent.putExtra(RecognizerIntent.EXTRA_LANGUAGE_MODEL,RecognizerIntent.LANGUAGE_MODEL_FREE_FORM); intent.putExtra("android.speech.extra.DICTATION_MODE", true); intent.putExtra(RecognizerIntent.EXTRA_CALLING_PACKAGE,this.getPackageName()); intent.putExtra(RecognizerIntent.EXTRA_PARTIAL_RESULTS, true); intent.putExtra(RecognizerIntent.EXTRA_MAX_RESULTS,1); sr.startListening

Speech Recognition on Unity Game Engine

99封情书 提交于 2019-12-23 12:47:06
问题 I've succesfully achieved the goal of developing an app on android that use speech recognition from Google API. But the fact is that I must use a real Game Engine because my app is going to be crossplatform and with a very rich User Interfacea and gestures. I'm thinking on using Unity Game Engine and I'm going to deploy my app to iOS and Android platforms. So... searching on web... I figured out that I can't use Google Speech Recognition API on Unity Game Engine becuase its belong to Google..

Client ID for Project Oxford Speech API

好久不见. 提交于 2019-12-23 10:08:47
问题 I'd like to test Microsoft Project Oxford Speech recognition API, but I don't know where or how to get the client_id. I'm trying to test Search API. First, I'm trying to get a token to use the service: curl -d -i "https://oxford-speech.cloudapp.net/token/issueToken?grant_type=client_credentials&client_id=123&client_secret=456scope=https%3A%2F%2Fspeech.platform.bing.com" { "statusCode": 401, "message": "Access denied due to invalid subscription key. Make sure you are subscribed to an API you

OS X Yosemite (10.10) API for continuous speech recognition

*爱你&永不变心* 提交于 2019-12-23 10:04:05
问题 OSX has now had continuous speech recognition built in for several versions. It is remarkably good now. Does OSX 10.10 have an API for this continuous speech recognition engine? Googling tells me that circa 2009 there was only NSSpeechRecogniser (MS SAPI sdk equivalent on OSX) Can anyone confirm whether this is still the case? I'm asking because I am looking at building my own integrated UI, which would be a mix of Leap controller and speech input. If only I could use OSX's engine, this would

Python Speech Compare

人盡茶涼 提交于 2019-12-23 09:57:02
问题 I have two .wav files that I need to compare and decide if they contain the same words (same order too). I have been searching for the best method for a while now. I can't figure out how to have pyspeech use a file as input. I've tried getting the CMU sphinx project working but I cant seem to get GStreamer to work with Python 27 let alone their project. I've messed around with DragonFly as well with no luck. I am using Win7 64bit with Python27. Does anyone have any ideas? Any help is greatly