speech-recognition

C# - Capture RTP Stream and send to speech recognition

流过昼夜 提交于 2019-12-03 16:32:58
What I am trying to accomplish: Capture RTP Stream in C# Forward that stream to the System.Speech.SpeechRecognitionEngine I am creating a Linux-based robot which will take microphone input, send it Windows machine which will process the audio using Microsoft Speech Recognition and send the response back to the robot. The robot might be hundreds of miles from the server, so I would like to do this over the Internet. What I have done so far: Have the robot generate an RTP stream encoded in MP3 format (other formats available) using FFmpeg (the robot is running on a Raspberry Pi running Arch

Speech to text in emulator: No Activity found to handle Intent

时光毁灭记忆、已成空白 提交于 2019-12-03 15:53:43
I want to ask how I can use speech to text code on my emulator. My codes work on real device but not work on emulator. The error said : No Activity found to handle Intent { act=android.speech.action.RECOGNIZE_SPEECH (has extras) } What can I do? package net.viralpatel.android.speechtotextdemo; import java.util.ArrayList; import android.app.Activity; import android.content.ActivityNotFoundException; import android.content.Intent; import android.os.Bundle; import android.speech.RecognizerIntent; import android.view.Menu; import android.view.View; import android.widget.ImageButton; import android

Non-Speech Noise or Sound Recognition Software?

南楼画角 提交于 2019-12-03 15:11:47
I'm working on some software for children, and looking to add the ability for the software to respond to a number of non-speech sounds. For instance, clapping, barking, whistling, fart noises, etc. I've used CMU Sphinx and the Windows Speech API in the past, however, as far as I can tell neither of these have any support for non-speech noises, and in fact I believe actively filter them out. In general I'm looking for "How do I get this functionality" but I suspect it may help if I break it down into three questions that are my guesses for what to search for next: Is there a way to use one of

Python Speech Recognition: 'module' object has no attribute 'microphone'

只谈情不闲聊 提交于 2019-12-03 14:57:31
Running the following code on a macbook air 64 bit, testing the code on python 2.7 and python 3.4 import speech_recognition as sr r = sr.Recognizer() with sr.microphone() as source: audio = r.listen(source) try: print("You said " + r.recognize(audio)) except LookupError: print("Could not understand audio") When i try python 2.7, I keep getting the error of : Traceback (most recent call last): File "star.py", line 3, in <module> with sr.microphone() as source: AttributeError: 'module' object has no attribute 'microphone' (star.py is the name of the file I am testing it on) When I try python 3.4

Microsoft Speech Recognition Speed

此生再无相见时 提交于 2019-12-03 14:40:12
I am working on a small practice app using Microsoft's speech recognizer. I cannot seem to get it to recognize single words fast enough for what I am doing. I would like to be able to speak normally and the system will grab "keywords" from what I am saying and produce a string. Currently I am using a custom grammar of 5 words. Red, Blue, Yellow, Green, Orange. For example if I say Red blue yellow. The engine will grab "Red" and the rest goes into the void. I have tried using the following: sr.RecognizeAsync(RecognizeMode.Single) Then reset it back to single mode after the AudioState changes to

Pocket Sphinx on Android

时间秒杀一切 提交于 2019-12-03 13:29:19
问题 What are the steps required to use PocketSphinx on Android? I have found various hints around the web that it is possible, but not real answers and hence I ask my question here. 回答1: This is the most comprehensive resource I've found to date. I still have some compilation errors, though. The blog gives you a tutorial on how to build the PocketSphinxAndroidDemo app, which was developed (and discontinued) on GitHub. Personally, I'm having similar compilation errors to the Stack Overflow

Using Mac OSX Dictation with Speech API

别等时光非礼了梦想. 提交于 2019-12-03 13:04:16
问题 In OSX Mavericks, speech dictation is now included, and is very useful. I am trying to use the dictation capability to create my own digital life assistant, but I can't find how to use the recognition functionality to get the speech in an application rather than a text box. I have looked into NSSpeechRecognizer, but that seems to be geared toward programming speakable commands with a pre-defined grammar rather than dictation. It doesn't matter what programming language I use, but Python or

How to plot MFCC in Python?

烂漫一生 提交于 2019-12-03 12:57:47
问题 I'm just a beginner here in signal processing. Here is my code so far on extracting MFCC feature from an audio file (.WAV): from python_speech_features import mfcc import scipy.io.wavfile as wav (rate,sig) = wav.read("AudioFile.wav") mfcc_feat = mfcc(sig,rate) print(mfcc_feat) I just wanted to plot the mfcc features to know what it looks like. 回答1: from python_speech_features import mfcc import scipy.io.wavfile as wav import matplotlib.pyplot as plt (rate,sig) = wav.read("AudioFile.wav") mfcc

RecognitionListener in JellyBean Freezes if not spoken to immediately

ぐ巨炮叔叔 提交于 2019-12-03 12:19:51
A speech-recognition based app I am working on works well on all versions of Android starting from API 8 (Android 2.2). But on a Nexus S 4G (Android 4.1.1), RecognitionListener will simply halt for about 1 minute , then issue an ERROR_SERVER via its onError() callback. If spoken to within 1-2 seconds (of that onReadyForSpeech bleep), it will behave properly as expected. What changed in JellyBean that could explain this behavior? More importantly, is there a way to make it behave like in the older versions of Android? (i.e. keep listening, then issue ERROR_SPEECH_TIMEOUT if not spoken to within

Add iOS speech recognition support for web app?

白昼怎懂夜的黑 提交于 2019-12-03 11:51:28
Currently, the HTML5 web speech api works great on google chrome for all devices except mobile iOS. Text-to-speech works, but speech-to-text is not supported. webkitSpeechRecognition is not supported. See: Chrome iOS webkit speech-recognition I am unable to find a workaround. I would like to add speech recognition support for iOS to my current web app that uses speech recognition and speech synthesis. Any suggestions? Thank you. Try something like this recognition = new (window.SpeechRecognition || window.webkitSpeechRecognition || window.mozSpeechRecognition || window.msSpeechRecognition)();