speech

Microsoft Speech Recognition Custom Training

核能气质少年 提交于 2019-12-04 09:34:54
I have been wanting to create an application using the Microsoft Speech Recognition. My application's users are expected to often say abbreviated things, such as 'LHC' for 'Large Hadron Collider' or 'CERN'. Given that exact order, my application will return You said: At age C. You said: Cern While it did work for 'CERN', it failed very badly for 'LHC'. However, if I could make my own custom training files, I could easily place the term 'LHC' somewhere in there. Then, I could make the user access the Speech Control Panel and run my training file. All the links I have found for this have been

Speech training files and registry locations

醉酒当歌 提交于 2019-12-04 08:09:54
I have a speech project that requires acoustic training to be done in code. I a successfully able to create training files with transcripts and their associated registry entries under Windows 7 using SAPI. However, I am unable to determine if the Recognition Engine is successfully using these files and adapting its model. My questions are as follows: When performing training through the Control Panel training UI, the system stores the training files in "{AppData}\Local\Microsoft\Speech\Files\TrainingAudio". Do the audio training files HAVE to be stored in this location, or can I store them

How to convert speech to text during call with different text colors for caller and call receiver?

感情迁移 提交于 2019-12-04 06:50:43
I want to convert speech to text during call. I also want the text to display in different colors: the call initiator's in red and the call receiver's green. During my tests, I converted speech to text during call but was unable to distinguish between the voice of the call initiator and that of the call receiver. Thanks in advance Please Help me out... 来源: https://stackoverflow.com/questions/20964359/how-to-convert-speech-to-text-during-call-with-different-text-colors-for-caller

Add another voice into .NET Speech

隐身守侯 提交于 2019-12-04 06:14:00
How can I add another voice into .NET Speech? I'd like to use Czech voice. I've found some *.sis files Eliska22k.sis but I don't know how to use it. SpeechSynthesizer synth = new SpeechSynthesizer(); foreach (var voice in synth.GetInstalledVoices()) Console.WriteLine(voice.VoiceInfo.Name); Gives me only Microsoft Anna . Smileek Well, There aren't good news for you. 1. Language Support - Czech language isn't supported by MS TTS. 2. sis is Symbian installer file and is useless in Windows. 3. Other languages can be found there: Download Microsoft Speech Platform - Runtime Languages (Version 11)

How to control speech output on Bixby

杀马特。学长 韩版系。学妹 提交于 2019-12-04 05:03:14
问题 I'm using Bixby Developer studio and I have my UI output created and it works. How do I create or control speech output? 回答1: Bixby will automatically read the text in the UI - if that is sufficient, you are set. (Be careful here - In good VUX design, UI text and speech text are typically different. UI text is often longer, speech text often shorter and more expressive). To have speech different from the UI text, you should set the speech key of a template in a dialog. You can use all of the

How to add custom SR (Speech Recognition) to Microsoft SAPI

社会主义新天地 提交于 2019-12-04 02:19:42
问题 I have a Speech Recognition application for a Language X. This language is not supported by Microsoft SAPI. How can I add this SR for Language X to Microsoft SAPI so that I can use it via Microsoft SAPI ? So What I wonder is Can I add a custom SR to Microsoft SAPI? If so How? If It is a complex process, where to find details? 回答1: SAPI is just the application programming interface, it only provides the connector for the speech recognition engine. There could be multiple implementations of the

Can Web Speech API used in conjunction with Web Audio API?

浪子不回头ぞ 提交于 2019-12-03 23:53:10
问题 Is it possible to use the synthesised speech from Web Speech API as a SourceNode inside Web Audio API's audio context? 回答1: I actually asked about adding this on the Web Speech mailing list, and was basically told "no". To be fair to people on that mailing list, I was unable to think of more than one or two specific use cases when prompted. So unless they've changed something in the past month or so, it sounds like this isn't a planned feature. 回答2: You can use Google's Web Speech API, you

Matlab: Finding dominant frequencies in a frame of audio data

北战南征 提交于 2019-12-03 21:11:37
I am pretty new to Matlab and I am trying to write a simple frequency based speech detection algorithm. The end goal is to run the script on a wav file, and have it output start/end times for each speech segment. If use the code: fr = 128; [ audio, fs, nbits ] = wavread(audioPath); spectrogram(audio,fr,120,fr,fs,'yaxis') I get a useful frequency intensity vs. time graph like this: By looking at it, it is very easy to see when speech occurs. I could write an algorithm to automate the detection process by looking at each x-axis frame, figuring out which frequencies are dominant (have the highest

How can you make the android emulator support speech recognition

与世无争的帅哥 提交于 2019-12-03 21:00:08
问题 I am writing an app which uses android's speech recognition. However my app doesnt run well on the emulator - runs fine on the mobile itself however. On the emulator I get a red error message through logcat which says: 05-06 20:19:38.527: ERROR/SpeechRecognizer(1745): no selected voice recognition service Does anybody know how to enable an emulator to support speech recognition services? Cheers, 回答1: In your emulator image, can you install Google Voice Search as in Converting speech to text.

SpeechRecognizer not Hearing After First Result

人走茶凉 提交于 2019-12-03 17:34:37
I am using SpeechRecognizer and RecognizerIntent in Android to implement speech recognition. My aim is to restart listening to speech after my speech recognizer displays the results on the screen. For that purpose, I am using the following code. The problem is, the first time runs fine and displays the results but after it starts listening for the second time (called from onResults method), it does not hear what is being spoken for some reason. Then it gives a ERROR_SPEECH_TIMEOUT error, which means there was no speech input. On Logcat, I can see it entering onReadyForSpeech() but somehow, it