Ok so I have been researching this for a while to no avail.
I am creating a google glass app that will look at a building and tell you the distance from that building. I have the coordinates of the building hardcoded and I now need to display the distance (text variable) to the camera preview.
I have looked at other examples of "overlaying text on the camera preview" that pertain to Android, but nothing that pertains to Google Glass. So can someone explain how to overlay a text box on the camera preview?
Related
I downloaded the ML Kit sample code for Android and i'm having a hard time to go about setting a limited detection area for the live camera detection (I'm only interested in text recognition, i got rid of everything else).
I need to limit the text recognition to only a part of the screen (say, a tiny rectangle or square in the center of the screen). Has anyone done such a workaround with ML kit?
Please take a look at the ML Kit Translate Showcase App which shows how to limit Text recognition to a specific section of the screen.
I am wondering how can I use the Google text recognition (OCR) with ARCore?
When I use the OCR sample to put some text above the detected text? It is jumping around as images are coming in. However I would like to anchor it to the text so when the camera moves, it is attached to it like ARcore
I couldn't find a way to do that? Google Lens does that
Any help or pointer is appreciated
Thank you
I have live camera preview, now project requires eye pupil movement tracking (which is different from Eye is open/close). Means which part of the screen is being seen by the user this time and what is the most viewed part of screen at the end of Video stream. Although, I tried with OpenCV & Google Vision api, but they are not able to perform the task perfectly.
Please suggest if you have any solution.
https://android.jlelse.eu/a-beginners-guide-to-setting-up-opencv-android-library-on-android-studio-19794e220f3c
https://developers.google.com/vision/
Are any of the current text capture APIs (e.g. Google's Text API) fast enough to capture text from a phone's video feed, and draw a box that stays on the text even as the camera moves?
I don't need fast enough to do full OCR per-frame (though that would be amazing!). I'm just looking for fast enough to recognize blocks of text and keep the bounding box displayed in sync with the live image.
There are two major options for good results. They are both C++ but there are wrappers. I've personally played with OpenCV for face recognition and the results were promising. Below links with small tutorials and demos.
OpenCV
Tessaract by Google
Firebase
onDeviceTextRecognizer is simple and working for me.
I am trying to show four videos at once using Google Cardboard. These videos are normal 2D videos that were shot on a normal 16:9 camera. What I want and need is to have one video in front of you then you turn your head 90 degrees and you see another video, turn again and see another until you hit the front video again. Please see my Pablo Picasso Microsoft Paint skills to visualize what I am talking about...
So basically what I need is like four VR movie theater screens that a person can look around in. Is there a program I could use or do I have to do some programming to make this happen? Searching this is not easy with all the articles of VR that pop up. Any help that can point me in the right direction would be greatly appreciated!
I actually found an app that did this all for me. The app is called 360 Virtual Reality Player(Google Play Store) and it takes any 2D video and makes it into a head-tracked VR video. Once I found this app, all I needed to do is stitch the videos together with a black bar in between them using OpenCV to get the desired effect.