speech

Python speech recognition slowing down

[亡魂溺海] 提交于 2021-02-16 21:20:50
问题 I have developed a new program to speak to my chatbot. It works very well, but there is one strange issue I can't seem to figure out. Every time the process repeats (the console outputs listening and does speech recognition), it slows down. The 1st go is quick, the 2nd a little slower, the 3rd slow, and then it just gets too slow to respond from there on. Please help me figure out what syntax could be causing this. import speech_recognition as sr r = sr.Recognizer() with sr.Microphone() as

Speech recognition intent starts before TTS ends speaking

99封情书 提交于 2021-02-10 05:28:48
问题 I've developed an Android App which give commands using tts (Text to Speech) and then listen to the vocal answer of the user using Speech Recognition Intent. Obviously, only when tts finishes to speak, then the intent of speech recognition is thrown. Sometimes, mostly when I swich from an activity to another, or when I activate this process through a button, THE SPEECH RECOGNIZER INTENT STARTS BEFORE TTS FINISHES TO TALK! The problem is that the same text of the command is then acquired and

Using Delphi, I'm trying to change the volume of the AndroidTTS, but I seem to keep getting the error “External exeception 1”

一曲冷凌霜 提交于 2021-01-28 06:12:54
问题 I'm using these libraries: https://github.com/jimmckeeth/FireMonkey-Android-Voice https://github.com/FMXExpress/android-object-pascal-wrapper/tree/master/android-25 Here is my source code: unit Unit1; interface uses System.SysUtils, System.Types, System.UITypes, System.Classes, System.Variants, FMX.Types, FMX.Controls, FMX.Forms, FMX.Graphics, FMX.Dialogs, SpeechRecognition, FMX.Controls.Presentation, FMX.ScrollBox, FMX.Memo, FMX.StdCtrls, FMX.Memo.Types, AndroidTTS, FMX.Media, System.IOUtils

Is there Web Speech API Limitation?

▼魔方 西西 提交于 2021-01-21 08:11:06
问题 I'm using W3C Browser Web Speech Api. I search everywhere but i cant find it answer. Is there any limitation for usage ? Demostration: https://www.google.com/intl/en/chrome/demos/speech.html 回答1: Yes, with a bit of a catch. I finally hit a limit recently after a lot of experimentation; as soon as a request starts, it's cut short. On a desktop web browser (Chrome), there's a limit for each user (Primiano Tucci works at Google, so I'd trust him!), and you can't buy more time with the API.

Is there Web Speech API Limitation?

你。 提交于 2021-01-21 08:09:24
问题 I'm using W3C Browser Web Speech Api. I search everywhere but i cant find it answer. Is there any limitation for usage ? Demostration: https://www.google.com/intl/en/chrome/demos/speech.html 回答1: Yes, with a bit of a catch. I finally hit a limit recently after a lot of experimentation; as soon as a request starts, it's cut short. On a desktop web browser (Chrome), there's a limit for each user (Primiano Tucci works at Google, so I'd trust him!), and you can't buy more time with the API.

Why 128 mel bands are used in mel spectrograms?

六月ゝ 毕业季﹏ 提交于 2020-06-29 06:42:19
问题 I am using the mel spectrogram function which can be found here:Mel Spectrogram Librosa I use it as follows: signal = librosa.feature.melspectrogram(y=waveform, sr=sample_rate, n_fft=512, n_mels=128) Why is 128 mel bands use? I understand that the mel filterbank is used to simulate the "filterbank" in human ears, that's why it discriminates higher frequencies. I am designing and implementing a Speech-to-Text with Deep Learning and when I used n_mels=64, it didn't work at all, it only works

How to get a list of ALL voices on iOS 9?

為{幸葍}努か 提交于 2020-06-24 08:46:32
问题 The standard way to get a list of 'all' AVSpeechSynthesisVoice objects, is to ask for it: [AVSpeechSynthesisVoice speechVoices] However, this only shows a unique default voice per language-region pair. e.g. for en-US: "Language: en-US, Name: Samantha, Quality: Default", If I look in the Settings app under General → Accessibility → Speech → Voices → English I see that I have 'Siri Female (Enhanced)' selected, but that voice is not selectable from code. The full list is So the big question is:

How to get a list of ALL voices on iOS 9?

限于喜欢 提交于 2020-06-24 08:45:53
问题 The standard way to get a list of 'all' AVSpeechSynthesisVoice objects, is to ask for it: [AVSpeechSynthesisVoice speechVoices] However, this only shows a unique default voice per language-region pair. e.g. for en-US: "Language: en-US, Name: Samantha, Quality: Default", If I look in the Settings app under General → Accessibility → Speech → Voices → English I see that I have 'Siri Female (Enhanced)' selected, but that voice is not selectable from code. The full list is So the big question is:

Web Speech API not properly loading voices in Chrome for Android

筅森魡賤 提交于 2020-05-17 08:49:41
问题 I have a simple app that should read out text entered into an input field in a selected language: https://speech-synthesis-demo.glitch.me/ This seems to work well on desktop in multiple browsers. However, when I try to run it in Chrome for Android, it seems that changing the language has no effect, and only the default language is used (in my case, English). For testing purposes, what I'm trying to do is test out counting in different languages. If you enter in the word 'one' for example in

Speech Recognition with free speech

淺唱寂寞╮ 提交于 2020-01-16 03:30:06
问题 I need help. I want to make an application that will recognize what I am saying and do stuff that I say. For example: If I say open [notepad] , where [notepad] can be any application name, it needs to open notepad. I think I need to use both Grammar and DictationGrammar , but I don't know how. Please help me. Thanks. My code now looks like this: string WelcomeSentence = "Hello sir, how are you today"; SpeechSynthesizer sSynth = new SpeechSynthesizer(); PromptBuilder pBuilder = new