speech-recognition

Has anyone created a MonoTouch binding for the Nuance Dragon Mobile Speech SDK for iOS?

北城以北 提交于 2019-11-29 08:06:55
I have the Dragon Mobile SDK running nicely on Windows Phone 7 and I would like to get the equivalent functionality working for iOS. Since the SDK wraps the microphone, it's not really possible to use the .NET assemblies in my MonoTouch project (even if I did have the source). It appears that the best way to do this is to create a binding library (as Miguel describes here ). It sure seems like a lot of work though, and I would love to reuse as opposed to reinventing the wheel if someone's done it already... Here are some more details for how I got this to work. I downloaded the binding sample

Listening for keywords at all times, like “Ok google” on 4.4 [closed]

北战南征 提交于 2019-11-29 07:53:13
问题 Closed . This question needs to be more focused. It is not currently accepting answers. Want to improve this question? Update the question so it focuses on one problem only by editing this post. Closed 5 years ago . I've checked the new Kitkat libraries to make sure that I'm not missing something. I want my application to always listen to a keyword before performing an action, just like how Google Now always listens for the keywords "Ok google" on the homescreen. How did Google accomplish

How to recognize a phrase from a voice file

限于喜欢 提交于 2019-11-29 07:45:32
How to get the engine to successfully recognize a phrase from a voice file (wav/mp3/etc..)? For example, if I'll have a voice file and a written text of the context of the same file, so to make it recognize the written words in the voice file. I tried to play around with the SpeechRecognitionEngine, but without success so far. I'll appreciate ideas, since this is my first time dealing with Speech Recognition techniques. I've seen examples of speech-to-text using dictionaries, but I'm not sure how it can be useful here. I was thinking of maybe converting the all voice file to text, and then

Google-speech-api transcribing spoken numbers incorrectly

半世苍凉 提交于 2019-11-29 07:31:56
I started using google speech api to transcribe audio. The audio being transcribed contains many numbers spoken one after the other. E.g. 273 298 But the transcription comes back 270-3298 My guess is that it is interpreting it as some sort of phone number. What i want is unparsed output e.g. "two seventy three two ninety eight' which i can deal with and parse on my own. Is there a setting or support for this kind of thing? thanks Try passing a speech context with some phrase hints. How to use it is documented here: https://cloud.google.com/speech/docs/basics#phrase-hints Give it the spelled

Android Wear Custom Voice Actions

我与影子孤独终老i 提交于 2019-11-29 06:31:09
I'm making an app that will work with Android Wear, And I wanted to implement a command into Google's "Ok Google" option. I saw this page: http://developer.android.com/training/wearables/apps/voice.html But it's related only to apps that include Activities in the Android Wear. I wanted to ask: Can I add custom commands? I mean, those who does not start with the word "Start"? Can I add commands that will do another thing than just opening the app? Like running a method? If it's not the place to ask this, can you give me an email/link to Google Developers help/support? thanks. For apps that run

Keyword Spotting in Speech on Android?

拥有回忆 提交于 2019-11-29 05:17:05
I need to use speech input to insert text. How can i detect keyword when I'm speaking ? Can i do this with Android Speech Input or I need external library ? Any ideas ? Thanks Nikolay Shmyrev Keyword detection task is different from a speech recognition task. While second tries to understand the text being spoken and check all possible word combinations, keyword spotting usually check two hypothesis - word is here or garbage is here. Its way more efficient to check keyword presence but it requires custom algorithm. You can implement one with the open source speech recognition toolkit like

SAPI: Speech to Text example

守給你的承諾、 提交于 2019-11-29 05:16:12
I am new to SAPI, and I would really appreciate if any of you can provide me a speech to text Hello World example in SAPI. I know MS got some examples like "Dictation" etc, but I would like to start with a very small one. Glad if you can help. I played a bit with Windows Voice Recognition using SAPI, it really isn't user friendly. Here is an example of code I wrote (in C++) : #include <sphelper.h> #include <sapi.h> #include <iostream> #include <string> const ULONGLONG grammarId = 0; const wchar_t* ruleName1 = L"ruleName1"; int start_listening(const std::string& word); ISpRecoGrammar* init

Android extras about speech recognition does not work

走远了吗. 提交于 2019-11-29 02:13:10
I am currently busy with speech recognition, and I want speech recognizer to listen me continously, forever. The purpose of the app is, it is going to wait for speech of a specific keyword: when the user says this keyword, the service becomes ready to receive user voice commands After research I found that we can manupulate the time that the recognizer listens. By these constants: By the way the default recognizer service listens about 10 seconds. (The google api) String EXTRA_SPEECH_INPUT_COMPLETE_SILENCE_LENGTH_MILLIS The amount of time that it should take after we stop hearing speech to

ios speech recognition Error Domain=kAFAssistantErrorDomain Code=216 “(null)”

前提是你 提交于 2019-11-29 01:00:21
问题 Basically I am learning ios speech recognition module following this tutorial: https://medium.com/ios-os-x-development/speech-recognition-with-swift-in-ios-10-50d5f4e59c48 But when I test it on my iphone6, I always got this error: Error Domain=kAFAssistantErrorDomain Code=216 "(null)" I searched it on the internet, but find very rare info about this. Here is my code: // // ViewController.swift // speech_sample // // Created by Peizheng Ma on 6/22/17. // Copyright © 2017 Peizheng Ma. All

Android: Accessing the microphone simultaneously (RecognizerIntent + own app)

一笑奈何 提交于 2019-11-29 00:55:01
问题 my app uses the RecognizerIntent to record the user's voice and doing some speech recognition. Now, I'd like to compare the results to some open source speech recognition engines. Most of them take an audio file as input. My thought was, to capture the sound from the Android's microphone, and start the RecognizerIntent at the same time. But it seems, that accessing the microphone is exclusive. Is it possible to use the RecognizerIntent with a recorded audio stream? Is it possible to access