I want to know that how voice recognition in android works? Which library it uses for voice recognition? Does it perform voice recognition process on device/mobile it self or it sends all voice to google servers and receives text in responce?
Thanks,
Sunny.
The 4 Feets.com answer is now very misleading, as the link contains quite a bit of speculative information that turned out to be inaccurate.
Please checkout the VoiceRecognition.java demo in ApiDemos, and the RecognizerIntent reference. Android speech recognition requires an internet connection as the data is sent off to google and you receive a list of possible text translations back.
Have a look at 4 Feets.com 4 Feets.com
They have a nice overview with a little example regarding voice reg in SDK 1.5.
Related
I am developing an android app which needs Speech recognition functionality, I need app should work same as Google Translator app in which it Google translate auto detect language spoken. I checked with Microsoft Azure Speech recognition and Google Speech recognition API in both the API we need to provide language code. I searched all over but did not get any workaround or solution to achieve similar functionality. Please help.
May be I am answering late but it may help
I think problem is very large than it seems. As far as I understood you want to achieve speech-to-text and then translate text to required language.
I think using paid services of microsoft or google or amazon would be helful.
https://azure.microsoft.com/en-in/services/cognitive-services/text-analytics/
https://aws.amazon.com/comprehend/
Is it possible to just have our Android app answer the question to only our Alexa custom skill and not have the entire default behavior of Echo? For example: I created a custom skill, say, calculate Can I make an android app which uses Alexa Voice service API to answer the questions only related to calculate and no other questions? (i.e. no default behavior like weather, music)
Why does the example on developer documentation app say "companion app"? Do I need Echo to use it? Can I not make an app which will answer questions but does not require echo?
Is it possible to get text and audio both as output using Alexa API?
I appreciate any input. Any links and references are welcome.
The benefit of Alexa is it's voice recognition abilities, and the ability to choose an appropriate intent based on a voice interaction. If the skill is written with clearly defined intents Alexa will be able to respond as you wanted. It may be that "Calculate..." might be too vague an intent for Alexa to differentiate.
Also, the useful bit is the skill you build. You define how things are calculated, and what answer to give. Unless you are trying to leverage the voice recognition and AI you might be better off going with some other technology (and if you need those things, then maybe WitAI might be more useful to you: https://wit.ai/ it's a little more roll-your-own than Alexa).
Alexa Voice Services (AVS) is available in the US, but not yet the UK or Germany until 2017 (and who know's when for other markets). AVS can be added physical devices that have a speaker and microphone, so it is possible to use Alexa without using an Echo or Echo Dot.
At it's core, the input and output of Alexa apps are JSON (so text). Alexa parses the text response and speaks the appropriate part. I'm not sure that you can route this response in some other way than having it spoken. However, in between the request and response is the Lambda function (or native device function), so in addition to generating the response to Alexa, you could dump the response somewhere else at the same time that would be available outside of Alexa.
Is it possible to just have our Android app answer the question to only our Alexa custom skill and not have the entire default behavior of Echo? For example: I created a custom skill, say, calculate Can I make an android app which uses Alexa Voice service API to answer the questions only related to calculate and no other questions? (i.e. no default behavior like weather, music)
Yes, it's possible to override the commands. First of all, create your custom skills using Amazon Skill Kit, then use android or iOS Alexa Application for
Android.
In "Settings", go to your product if echo/dot or your android/iOS application and enable your skill.
Why does the example on developer documentation app say "companion app"? Do I need Echo to use it? Can I not make an app which will answer questions but does not require echo?
The documentation context of companion app is only to use your hardware as an Alexa device. So using the Amazon login with Amazon SDK library the developer has the to authorize the user and get token from Amazon server for your hardware to communicate with the Alexa server.
Yes, you can make an android or iOS app for talking to the Alexa server. The link below is to a well-developed library for the same.
https://github.com/willblaschko/AlexaAndroid
Is it possible to get text and audio both as output using Alexa API?
No you will never get the text intepretation you will only get the response from Alexa in the form of JSON.
hello i wanted to ask for some knowledge on using voice/speech recognition to carry out instructions in my game.
I'm fairly new to game development and i am using libgdx to build my game. i know speech recognition api's exist. i would like to know how they work as in how i can integrate them to my game. i would like for the user to be able to say jump and the (player) to jump, the speech commands i want to use are very basic e.g. shoot and the player should shoot a bullet.
If anyone seems to have some knowledge with speech/voice recognition api's i want to know if there's a simple way i could set the speech recognition api i will be using to carry out specific action upon hearing specific keywords such as "jump".
Any answer will be helpful because my knowledge is very limited with using api's and speech recognition.
A service would be required to do this. This is basically a class that will run in background and hence will be able satisfy your needs above.
Here is a useful thread that uses a voice recognition service:
Android Speech Recognition Continuous Service
Can someone please help me?
I am developing an application with voice recognition via RecognizerIntent.
Which Android version brought in officially the offline recognition available to apps by API? Is there any statement about it?
By what I read until know it is not a choice of the developer if the voice recognition will be done via online service or the offline dictionaries. Am I right? Or are there any documented API to set offline?
Thanks
Now there is offline voice recognition is possible for android 4.1. check this article
article
and also check pochetsphinx
Google has speech recognition services available for use from mobile phones (Android has it built in, iPhone users can use the Google application) - http://www.google.com/mobile/. We've found one article where someone tried to reverse engineer the service at http://waxy.org/2008/11/deconstructing_google_mobiles_voice_search_on_the_iphone/.
We want to better understand what is happening over the network when we use Android's RecognizerIntent. Does anyone have any experience using this service over the web or know of other articles that may explain its workings?
I read this presentation few weeks ago- http://www.abelski.com/courses/android/speechinput.pdf
The following link is a 3 mile high review of the Google Voice Server ....
http://www.google.co.jp/events/developerday/2010/tokyo/pdf/tt1-gruenstein.pdf
Answer: just move your .apk file to your android phone it will work. Error is only occuring because we are trying to do in emulator