How to use Speech Recognition inside the iOS SDK? [closed]

谁说胖子不能爱 提交于 2019-11-26 15:13:43

问题


I know that there is no public API for the SIRI-Services, but is there an API for simple Speech-Recognition? So if I have a textfield and the user taps onto that textfield, a keyboard with the typically microphone button appears and if he pressed it the speech get recognized and transformed into a string object? Or is this button maybe presented by default?


回答1:


There are many libraries availble. You can use any of them.

  1. openears // This is the best library

  2. VocalKit (Deprecated for open ears)

  3. TTS
  4. ispeech (Not free)

Hope it helps you.

NOTE:

if you download openears (which contains a sample project called "OpenEarsSampleApp") @efimovD mentions this

Check the code in view controller and you will see an array with possible commands. This thing detects commands! Not some talk. It listens and tries to compare what you've said with the words from array




回答2:


You can use the MindMeld iOS SDK provided by Expect Labs. It supports continuous (streaming) speech recognition on iOS. It also supports natural language understanding and content discovery for any custom content domain.

Try it out at: https://expectlabs.com/docs/sdks/ios/gettingStarted



来源:https://stackoverflow.com/questions/16812399/how-to-use-speech-recognition-inside-the-ios-sdk

易学教程内所有资源均来自网络或用户发布的内容,如有违反法律规定的内容欢迎反馈
该文章没有解决你所遇到的问题?点击提问,说说你的问题,让更多的人一起探讨吧!