i am trying to build a game with unity which have Arabic speech recognition to be used on android devices. I am stuck at the speech recognition tool. which one is more suitable with unity ? can i use google API directly ? if yes can i control the data ?
or sphinx is more suitable for game on android device ? I have read about sphinx but the acoustic model of Arabic language was not available but for it.
I have tried to make it using MATLAB but it will cost me a lot to build it on mobile using MATLAB coder and also it does not have all the required libraries.
any help about this ?
Official unity plugin for pocketsphinx is here:
https://github.com/cmusphinx/pocketsphinx-unity-demo
Arabic model has to be trained, yes. In the end you could have a good system, but you have to invest enough into it.
Related
Is it possible to integrate external TTS engine with Pepper Robot?
I want to integrate Third party Speech engine with pepper robot. Please guide me on the same.
You can integrate an external TTS engine with Pepper. Either offboard (like the services offered by IBM, MS Azure or Google) or onboard (ideal would be something in Java or Kotlin for Android Pepper, but anything is possible). If you have a specific technology in mind, please provide more details and we can give you a more precise answer.
Bear in mind that this may introduce latency in speech synthesising compared to the default text to speech engine.
Edit - sorry, I missed your Android tag. The below mentioned APIs only work on Pepper 2.5 (Choregraphe Pepper)
Alternatively, there are a number of different voices available on Pepper, perhaps one will suit your needs. Use the naoqi API function ALTextToSpeech.getAvailableVoices to list the different voice options, then ALTextToSpeech.setVoice to set the voice to one of those options.
I am currently working with a Pepper robot (academic version and the QiSDK and NaoQi 2.9). Since I am using the academic version I can't use the cloud based automatic speech recognition service from Softbank which is not included and therefore e.g. I can't use wildcards or other chatbot engines besides QiChat.
Does anybody of you know how I can implement my own speech recognition service for Pepper? I can't find where I can get access to the audiostream of Pepper's microphones.
I've read the documentation from Softbank:
https://developer.softbankrobotics.com/pepper-qisdk
and
https://qisdk.softbankrobotics.com/sdk/doc/pepper-sdk/ch4_api/conversation/reference/basechatbot.html
And I've tried to create a SpeechRecognizer based on Android, which works, but uses the Tablets microphone and not Peppers.
Remote Speech Recognition is a service that you will need to by on top if it was not included with your original Pepper offer!
Regards,
Jonas
I was also curious and contacted the softbank support.
Summary:
With version 2.9. you have no access to the head microphones and can only access the tablet mic.
I am exploring the APIs provided by Google. Firstly, I was experimenting with Google Cloud Vision API with Python in PyCharm in order to try to perform Optical Character Recognition with various texts.
So I wrote a basic program in Python in PyCharm which was calling this API, I gave to it as an input an image which included text e.g. the image/photo of an ice-cream bucket and then takes the text written on this bucket as an output.
Now I want to test the barcode scanner of Google Mobile Vision API. So ideally I would like to call the Google Mobile Vision API in a python program in PyCharm which calls this API, give as an input an image/photo of a barcode and take as an output the details saved in this barcode.
My question is if this can be (easily) done with PyCharm or if I should download Android Studio to do this simple task?
In other words, can I call easily a mobile API in an IDE which is not for mobile app development like Android Studio but in an IDE for desktop applications like Pycharm?
It may be a very basic question but I do not know if I missing something important.
The mobile vision API is designed only for Android and iOS. As far as I know, Pycharm does not work well with Java, so I would say that you would have to create an Android/iOS project in order to test it (It would be a lot harder trying to make it work with python than simply installing Android studio and cloning a mock project).
I wanted to know how can I make a Engine to support Hindi for Text-To -Speech for Android.
I don't want to use any paid Engine version nor I want to have a web-service kind of interface providing the transliteration. I wanted to build an actual Hindi Support TTS Engine. How can I start? What are the steps.
Thanks,
I am trying to build an application (which would use the Hindi language and other regional languages) to get speech voice commands.
I also need text to speech functionality in my application.
I was wondering if there was any way I could get a speech recognition library on Android? I did a quick Google search and found a couple of libraries for Hindi on the Internet but I am not sure if I can include them in my Android project. Can I?
You can use pocketsphinx for your app.It has JAVA and Python API's for capturing and recognizing speech. By default it recognizes only English. But if you provide your own Language Model (LM) and Dictionary File(.DIC), you should be able to get it working.
Not directly straightforward as using an API, but can be certainly doable
http://cmusphinx.sourceforge.net/wiki/tutorialandroid
Android does support a few languages but not very many, I think the list currently is US, Japanese and Chinese. Not Hindi yet though I would think they are working to add new language support soon.
Over on the Android Developers page they have a good example.