speech-recognition

AndroidPocketSphinx: How does the system know which recognizer is invoked?

爷,独闯天下 提交于 2019-12-10 17:34:37
问题 I am studying the source code of TestPocketSphinxAndAndroidASR.java and the first thing that is not so clear to me is how the system knows which recognizer (i.e. Google or CMUSphinx) to invoke. I can see that the recognition activity is started by: Intent intent = new Intent(RecognizerIntent.ACTION_RECOGNIZE_SPEECH); intent.putExtra(RecognizerIntent.EXTRA_LANGUAGE_MODEL, RecognizerIntent.LANGUAGE_MODEL_FREE_FORM); startActivityForResult(intent, VOICE_RECOGNITION_REQUEST_CODE); but as far as I

Improve Android speech recognition with additional context

喜夏-厌秋 提交于 2019-12-10 17:00:51
问题 As I understand Android API uses google speech recognition service for speech to text. I've learned API and I see it's pretty simple and just converts voice to words array. Is any way to improve the recognition, I mean, if I know the context can I send some parameters to the service in order to improve the recognition? Or alternatively is any other speech recognition service which can be used for this purpose? Thanks in advance. 回答1: Is any way to improve the recognition, I mean, if I know

How to disable sentence-level auto correction in Google Cloud Speech-to-Text API

天大地大妈咪最大 提交于 2019-12-10 16:59:59
问题 I am working on a speech recognition task, which involves the detection of children's speaking capability, improvement over time ... I'd like to use the Google Cloud Speech to Text API for the ASR part of the detection. Then I would use the transcripts of different measurements to estimate the advancement. But! The sentence level autocorrect of Google Speech API consistently rewrites the previous limb of the spoken sentence... Is there a way to disable the autocorrect of this ASR? I can't

Windows Universal App Continuous Dictation Without Network

泄露秘密 提交于 2019-12-10 16:20:09
问题 Following the samples provided here: https://github.com/Microsoft/Windows-universal-samples provides a great overview of some of the capabilities for UWP apps. But, the speech example seems to require an active connection to the internet. Does anyone know if this capability is possible on a disconnected corporate network? Thanks, JRF 回答1: Yes, Examples 1, 2, 5, 6 and 9 work Offline since they don't use the predefined SRGS scenarios. Look in the Folder: SRGS to see the SRGS. You will have to

RecognizerIntent not working; “missing extra calling_package”

帅比萌擦擦* 提交于 2019-12-10 15:47:21
问题 I'm having problems using the RecognizerIntent API on Android 2.2. When I call the API using this code: Intent intent = new Intent(RecognizerIntent.ACTION_RECOGNIZE_SPEECH); intent.putExtra(RecognizerIntent.EXTRA_LANGUAGE_MODEL, RecognizerIntent.LANGUAGE_MODEL_WEB_SEARCH); startActivityForResult(intent, REQUEST_CODE_VOICE_SEARCH); which looks like it should work, the search popup says "Unknown problem" on the device and in the logs it states: 01-17 14:25:30.433: ERROR/RecognitionActivity(9118

Configuring the length of utterance and pauses in Android's speech recognizer

左心房为你撑大大i 提交于 2019-12-10 11:47:14
问题 I have android's Speech To Text API to speak something to the phone and convert it into text. By default, if one stops speaking to the microphone, the API assumes that the user is done talking and returns the text from the input speech. For my application, the user might have long pauses between her consecutive sentences. How can I configure Android's speech to text API to consider the end of the speech only when I ask it to and not do that as soon as the speaker takes a small pause between

How to get audio amplitude with speech recognizer?

孤街醉人 提交于 2019-12-10 11:33:37
问题 I'm trying to make a custom dialog while speech recognizing instead of using the official one. I got that part, but when then I decided to show the amplitude of the sound while recognizing, in order to make it more fancy, like Google Now search bar does (its a circle around the microphone that grows if voice its louder): Then I started to code how to obtain the amplitude of the sound, and finally I got it with AudioRecord Class. The problem comes when I try to mix both (SpeechRecognizer and

Query supported languages for SpeechRecognizer using BroadcastReceiver intent

心不动则不痛 提交于 2019-12-10 11:24:52
问题 I'm having trouble querying for supported languages using the SpeechRecognizer.ACTION_GET_SUPPORTED_LANGUAGES. private void queryLanguages() { Intent i = new Intent(RecognizerIntent.ACTION_GET_LANGUAGE_DETAILS); sendOrderedBroadcast(i, null); } Now, I know it says the BroadcastReceiver is specified in RecognizerIntent.DETAILS_META_DATA, but I'm not sure as to how I can access that. So basically what I'm asking is how do I create an Intent to retrieve the available languages data? 回答1: This is

Getting WAV file transcription to work with Sphinx4

蓝咒 提交于 2019-12-10 11:17:37
问题 I've got Sphinx-4 installed on my windows XP system and JSAPI set up. I'd like to transcribe an English spoken WAV (or MP3) file to text. When I run the "WavFile" demo - it runs successfully. java -jar WavFile.jar But, when I pass my own wav file like this: java -jar WavFile.jar c:\test.wav I get: Loading Recognizer as defined in 'jar:file:/C:/sphinx4-1.0beta3-bin/sphinx4-1.0beta3/bin/WavFile.jar!/edu/cmu/sphinx/demo/wavfile/config.xml'... Decoding jar:file:/C:/sphinx4-1.0beta3-bin/sphinx4-1

speech recognition and sound comparation witth musicg

╄→гoц情女王★ 提交于 2019-12-10 11:04:57
问题 I'm trying to make an Android application with speech recognition, but unfortunately google doesn't support my language (MACEDONIAN) and i'm trying to compare two recording sounds. I'm using http://code.google.com/p/musicg/ to record and compare speech, and i'm stack on initialization the settings for detecting the speech .Some one can tell me how to rewrite this init function for speech detection it's very important to me.. or some other idea how to do that. this is the initialization for