Google has speech recognition services available for use from mobile phones (Android has it built in, iPhone users can use the Google application) - http://www.google.com/mobile/. We've found one article where someone tried to reverse engineer the service at http://waxy.org/2008/11/deconstructing_google_mobiles_voice_search_on_the_iphone/.
We want to better understand what is happening over the network when we use Android's RecognizerIntent. Does anyone have any experience using this service over the web or know of other articles that may explain its workings?
I read this presentation few weeks ago- http://www.abelski.com/courses/android/speechinput.pdf
The following link is a 3 mile high review of the Google Voice Server ....
http://www.google.co.jp/events/developerday/2010/tokyo/pdf/tt1-gruenstein.pdf
Answer: just move your .apk file to your android phone it will work. Error is only occuring because we are trying to do in emulator
Related
I have this Samsung-watch
I implemented a mobile application for my graduation project, it is a triaging system that will serve the people who has an injury or illness,
In some cases we need to ask the patient about his heart rate, we want to take this data from the watch instead of asking him about it.
I searched for this and founded that I need to install Tizen extension to the visual studio, I want to write a code that take heart rate, then connect this code to my android application,
Does anyone know or can help me to do this?
Does anyone know the steps that i would follow to complete this process ?
You will find a lot of help from Samsung Developer Site.
Check below link:
Samsung Developers - Creating Your First App
May the force be with you.
How to create Pulsometer application for wearable:
http://tizenschool.org/tutorial/166/contents/1
then, for example, You can use Samsung Accessory Protocol to communicate wearable with android device:
https://developer.samsung.com/galaxy/accessory
https://developer.samsung.com/galaxy/accessory/guide
or implement other way of data exchange, for example uploading data to remote server:
How to make HTTP POST web request
Is it possible to just have our Android app answer the question to only our Alexa custom skill and not have the entire default behavior of Echo? For example: I created a custom skill, say, calculate Can I make an android app which uses Alexa Voice service API to answer the questions only related to calculate and no other questions? (i.e. no default behavior like weather, music)
Why does the example on developer documentation app say "companion app"? Do I need Echo to use it? Can I not make an app which will answer questions but does not require echo?
Is it possible to get text and audio both as output using Alexa API?
I appreciate any input. Any links and references are welcome.
The benefit of Alexa is it's voice recognition abilities, and the ability to choose an appropriate intent based on a voice interaction. If the skill is written with clearly defined intents Alexa will be able to respond as you wanted. It may be that "Calculate..." might be too vague an intent for Alexa to differentiate.
Also, the useful bit is the skill you build. You define how things are calculated, and what answer to give. Unless you are trying to leverage the voice recognition and AI you might be better off going with some other technology (and if you need those things, then maybe WitAI might be more useful to you: https://wit.ai/ it's a little more roll-your-own than Alexa).
Alexa Voice Services (AVS) is available in the US, but not yet the UK or Germany until 2017 (and who know's when for other markets). AVS can be added physical devices that have a speaker and microphone, so it is possible to use Alexa without using an Echo or Echo Dot.
At it's core, the input and output of Alexa apps are JSON (so text). Alexa parses the text response and speaks the appropriate part. I'm not sure that you can route this response in some other way than having it spoken. However, in between the request and response is the Lambda function (or native device function), so in addition to generating the response to Alexa, you could dump the response somewhere else at the same time that would be available outside of Alexa.
Is it possible to just have our Android app answer the question to only our Alexa custom skill and not have the entire default behavior of Echo? For example: I created a custom skill, say, calculate Can I make an android app which uses Alexa Voice service API to answer the questions only related to calculate and no other questions? (i.e. no default behavior like weather, music)
Yes, it's possible to override the commands. First of all, create your custom skills using Amazon Skill Kit, then use android or iOS Alexa Application for
Android.
In "Settings", go to your product if echo/dot or your android/iOS application and enable your skill.
Why does the example on developer documentation app say "companion app"? Do I need Echo to use it? Can I not make an app which will answer questions but does not require echo?
The documentation context of companion app is only to use your hardware as an Alexa device. So using the Amazon login with Amazon SDK library the developer has the to authorize the user and get token from Amazon server for your hardware to communicate with the Alexa server.
Yes, you can make an android or iOS app for talking to the Alexa server. The link below is to a well-developed library for the same.
https://github.com/willblaschko/AlexaAndroid
Is it possible to get text and audio both as output using Alexa API?
No you will never get the text intepretation you will only get the response from Alexa in the form of JSON.
I am looking into a way to integrate limited-vocabulary speech recognition into my Android app (since Google Voice Search relies on Internet access which isn't always available to users).
I found a Nuance page enticing programmers to download NDEV or the "Dragon Mobile SDK" but if it requires Internet access just like Google Voice Search, why should I be even interested in it?
Unfortunately, Nuance's page is not very informative and requires you to register (and sign a dragonian agreement...) before you get a chance to spend time to download & play with the kit only to discover that it's not what you're looking for...
Do you happen to know whether the speech-recognition part of NDEV or he "Dragon Mobile SDK" require Internet access?
Do you know of any off-the-shelf SDK for Android that doesn't require Internet access (i.e. some limited-vocabulary speech recognition can be done locally on the Android device)?
Yes, NDEV does require Internet access. Their SDK is just an API wrapper + the UI with the microphone button (and the code necessary to capture the audio properly).
If you need local speech recognition, have a look at Sphinx, it's a great project.
On most phones (with hardware powerful enough) Google's own speech recognition can be set to work offline.
Simply go to Settings->Language and Input-> Voice Search -> Offline Speech Recognition
There you'll be able to download data for any language you need available offline.
Google has recently made great progress with their speech recognition software, which is used in several open source products, e.g. Chromium Web Speech and Android Handsfree texting. I would like to use their speech recognition as part of my server stack, however I can't find much about it.
Is the text recognition software available as a library or package? Or alternatively, can I call chromium from another program to transcribe some audio file to text?
The Web Speech API's are designed only to be used in the context of either Chrome or Android. There is a lot of work that goes on in the client so there is no public server to server API that would just take an audio file and process it.
If you search github you find tools such as https://gist.github.com/alotaiba/1730160 but I am pretty certain that this method of access is 100% not supported, endorsed or confirmed to keep working.
The method previously stated at https://gist.github.com/alotaiba/1730160 does work for me. I use it on a daily basis in my home automation programs. I use a python script to capture audio and determine what is useful audio or just noise, then it sends the little audio snippet to google and returns the text all under a second!! I have successfully integrated it into my programs and if you google around you will find even more people that have as well!
I want see the source code for the voice enabled-keyboard feature for android.
Can someone tell me where to find the code?
I assume you're referring to the speech recognition feature demonstrated on the Nexus One with Android 2.1.
If this application is open sourced as part of Android, it will be posted on the Android Open Source Project website at https://android.googlesource.com.
However, Android 2.1 has not yet been posted; it should hopefully be available soon.
In the meantime, you could take a look at the source to the voice dialler application.
As far as I know this code is not currently planned to be open sourced -- it is owned by Google as part of their voice recognition server technology. The IME is a fork that Google made of the standard platform input method, adding voice search to it, much like other manufacturers make their own proprietary customizations.