speech-recognition

Applying neural network to MFCCs for variable-length speech segments

醉酒当歌 提交于 2019-12-23 04:07:31
问题 I'm currently trying to create and train a neural network to perform simple speech classification using MFCCs. At the moment, I'm using 26 coefficients for each sample, and a total of 5 different classes - these are five different words with varying numbers of syllables. While each sample is 2 seconds long, I am unsure how to handle cases where the user can pronounce words either very slowly or very quickly. E.g., the word 'television' spoken within 1 second yields different coefficients than

Training acoustic model for CMU Sphinx using Finite state grammar (JSGF)

随声附和 提交于 2019-12-23 02:21:16
问题 I need to train my own acoustic model. On the official wiki, this process is described for ngramm models in DMP format, but I need to use a (grammar) language model provided in JSGF format. Is there a such possibility? And if so, could you give a tutorial on this subject? 回答1: Modifications are minimial. Open psdecode.pl script in the installed location or in scripts and change -lm => $ST::DEC_CFG_LANGUAGEMODEL, to -jsgf => "/full/path/to/jsgf/file", 来源: https://stackoverflow.com/questions

SpeechRecognizer not work, COMException: Class not registered/ UWP App Windows IoT (10.0.10586) and Visual Studio 2015 Update 1

两盒软妹~` 提交于 2019-12-22 17:58:19
问题 After I have installed Windows IoT (10.0.10586) and Visual Studio 2015 with Update 1, I got COM-Exception when I use the SpeechRecognizer in Universal App on my Raspberry Pi 2 (with Windows IoT 10.0.10586). If I run the SpeechRecognizer UWP App on Windows 10 it works without any problems, the COM-Exception occurs only in Windows IoT (10.0.10586). With older version of Windows IoT and Visual Studio 2015 without Update 1 it works, too. Has anyone a solution for the problem? var speechRecognizer

I want to incorporate continuous (hands free) voice command recognition in my app for home automation

情到浓时终转凉″ 提交于 2019-12-22 17:52:19
问题 I have created a simple android app for controlling a relay connected to my Raspberry Pi. I have used buttons as well as basic voice recognition to trigger those buttons and switch on/off the corresponding relay channel. As of now the voice recognition part is handled by a RecognizerIntent, wherein I need to press a button on my app to open a Google voice prompt which listens to my voice command and activates/deactivates the corresponding button which controls the relay switches. I want to do

Recognize Speech To Text Swift

穿精又带淫゛_ 提交于 2019-12-22 17:16:49
问题 Is it possible to recognize speech and then convert it into text with custom keyboard. Like by default message app in iphone. Screen Shot 1. Default recognize speech in iphone keyboard. 2. Speech to text Any help would be greatly appreciated. Thanks in advance. 回答1: I have following code which are used in my sample application to convert speech-to-text. import UIKit import Speech import AVKit class ViewController: UIViewController { //----------------------------------------------------------

Nist Sphere format files

空扰寡人 提交于 2019-12-22 14:00:13
问题 In order to read NIST sphere format files, I'm trying to install NIST SPHERE software downloaded from here, but I encountered some errors: make[2]: Entering directory `/home/ibtissem/tools/nist/src/bin' gcc -I/home/ibtissem/tools/nist/include -L/home/ibtissem/tools/nist/lib -g -g -DNARCH_linux h_add.c -lm -o h_add h_add.c:31: error: undefined reference to 'sp_verbose' h_add.c:31: error: undefined reference to 'sp_verbose' h_add.c:28: error: undefined reference to 'hs_getopt' h_add.c:42: error

how to pass language in speech recognition on android apps?

旧时模样 提交于 2019-12-22 13:59:34
问题 I've been working on speech Recognition API in android and found out that the speech results vary allot when the language settings are changed , is there a way to set it programmatically ? or is there an intent to lunch the speech language settings screen ? or what else ? note: I tried to use this intent extra: intent.putExtra(RecognizerIntent.EXTRA_LANGUAGE_PREFERENCE, "en-US"); and Intent detailsIntent = new Intent(RecognizerIntent.ACTION_GET_LANGUAGE_DETAILS); sendOrderedBroadcast

Speech to Text Include speech engine in apk

心不动则不痛 提交于 2019-12-22 11:08:57
问题 I need a speech to text facility in my app for Really non-Geeks. Is there a way to include the speech engine at compile time for all sorts of tablets? 回答1: You could include your own engine, like the CMU Sphinx Engine. This has also the advantage of working without a data connection. See this post on how to install it on Android - however you have to change the Android.mk : Reverse the order of LOCAL_STATIC_LIBRARIES so the line looks like this: LOCAL_STATIC_LIBRARIES := pocketsphinx sphinxlm

Using c++ to call and use Windows Speech Recognition [closed]

那年仲夏 提交于 2019-12-22 10:48:25
问题 Closed. This question is off-topic. It is not currently accepting answers. Want to improve this question? Update the question so it's on-topic for Stack Overflow. Closed 6 years ago . I am making an application that involves the use of windows speech recognition. I am thinking of using c++ to do this since i have some experience with this language. The way i want to use the speech recognition is so that it works internally. If i upload an audio file into my program, i want speech recognition

How to connect SpeechRecognizer to RecognizerIntent with Extras

為{幸葍}努か 提交于 2019-12-22 10:45:05
问题 I am trying to wrap my mind around the SpeechRecognizer. I have a SpeechRecognizer with my own Recognition Listener: rec = SpeechRecognizer.createSpeechRecognizer(this); rec.setRecognitionListener(new RecognitionListener() { //Lots of overrides that work perfectly fine }); Wich works fine when I launch it by using rec.startListening(intent); But my intent happens to have some Extras: intent.putExtra(RecognizerIntent.EXTRA_PARTIAL_RESULTS, true); intent.putExtra(RecognizerIntent.EXTRA_RESULTS