Since i began programmation this forum provides me all that i always need! i want to thank you for... Thanks for all!!!!
Now i'm here to asking a problem that i not found it yet here.
I'm working on an android application. In my app, i have to read an android panorama which must be on a distant server. I have two problems to do this and hope u to save:
1. I take a panorama from my phone and When i connect it to PC from copying my panorama, this one become a simple jpeg image. I don't know how and why!!!
2. I have no idea on how to view panorama on android. I search on google, on android forums and still at my beginning point, i have to present my application next week!!!
So i give myself to you for bringing me out from this depth.
Thanks.
there's this library that do that with spherical cubic and cylindrical panoramic imagesPanoramaGL
and there's the utility with the google Play Services libs but only for spheric images refer to this :
Android Support for Photo Sphere
Related
My application's output would be a panorama photo. I looked around for solutions but I couldn't find one which I liked. Almost all of what I saw were an exact copy of a many years old project back on code.google, using GL ES 10 only.
I'd like possible good performance and I also want to avoid reinventing the wheel. Google's Open Spherical API (https://developers.google.com/streetview/open-spherical-camera/) is great, but I need an implementation for a viewer.
I'm wondering if it's possible to reuse StreetViewPanorama's capabilities to implement a 360 viewer leaving out the Google Maps portion. Looks like that getStreetViewPanorama requires a MapView. If not StreetViewPanorama, is there another API I could use?
Or is there some intent I can fire? In that case I should tell though that I periodically would like to update the displayed 360 image.
I've been wanting to create an android's cardboard app, in which the user enters a spherical panoramic view of some panoramic picture I've taken, and that I could gather data about where the user is looking in the process.
I've seen the "Cardboard Demo" provided by google, they have a feature called "photo sphere" in which the user can view photos exactly the way I want, but I want to implement it differently.
Can anyone give me some direction on how such a panormaic viewer with cardboard?
I was also trying to build the same kind of app. You can do a Skybox implementation in Opengl.(Cube mapping)
As you said, if you already have panorama image, then you can map the same to a Sphere.
You can Refer this Blog
I have ended up using Google's Cardboard SDK for Unity.
Basic idea :
Get Unity.
https://unity3d.com/get-unity/download
Get Google CardBoard
https://www.google.com/get/cardboard/get-cardboard/
Download the following SDK for unity
https://github.com/googlesamples/cardboard-unity
Try first to get the CardBoard/DemoScene (It's in the previous link) running on your app. Use this link for guidance :
https://developers.google.com/cardboard/unity/get-started
The DemoScene is basically a 3D game for Google cardboard, you can start by loading that project into unity and playing around with the different elements. From here it's pure unity, you can use the cardboard elements (They are the only important thing here, and define two offset cameras that are synced with the movement of the phone for 3d viewing of the cardboard) from the DemoScene project and create your own scene.
As due to gathering data on where the user is looking at , I extracted the
angles of the cardboard camera objects on unity.
Good Luck!
Refer this link- Open Photoshpere from SD card in android to view in Google Cardboard
You can use the Rajawali framework-
https://github.com/ejeinc/RajawaliCardboardExample
I want to do Face recoginition (Not Face Detection) in my Android and iOS app. I have studied a lot on Web and found following possible solutions:
1.) openCV: I don't want to go into writing my own API using this. Also, I don't have prior experience in JNI for Android.
2.) Betaface API So far this is good.
3.) Sky Biometrics is also Good.
Now, I am searching for the solution from 3-5 days and came to know that I can use above API (so far I have decided to purchase license for Sky Biometrics). And this API will provide me a list of Features for the faces it recognised.
But, Now I am confused That how to use these features and save in my local data base to recognise faces from the pictures.So My queries are following
1.) How to convert Face features to Actual working Face recognition API means What is the actual algo or solution I can use to merge diffrent face features of a same person to identify him correctly.
2.) Uploading images and then creating database for Face-features set is a very time taking process. Do any one knows any Android/iOS Face Recoginition SDK to do this so that I can do this accurately and timely with no or less time taking process.
3.) Both solution-2 and 3 can be used with Images. Is there any other solution is available which can do the same with less efforts but with more accuracy.
OpenBR may be also interesting for you: http://openbiometrics.org/
Finally I am using Rekognition API. And this is good enough to serve my purpose.
I am using the Vuforia SDK to build an Android application and am curious as to how the marker tracking works. Does the app convert the video frame into byte codes and then compare these against the .dat file generated by creating the marker? Also, where is this code found in the Vuforia sample app, is it in the C++ ? Thanks.
Well, you don't see the code for recognition and tracking because they are Intellectual property of Qualcomm and usually should not be revealed. Vuforia is not an open-source library.
Vuforia first detects "feature points" in your target image [Web-based target management] and then uses the data to compare the features in target image and the receiving frame from camera.
Google "Natural feature detection and tracking" which falls under Computer Vision area and you will find interesting stuff.
No, detection and tracking code placed in libQCAR.so
But question "how its work" is to complex to answer here. If you want to be familiar with object detection and tracking - start to learn method as mser, surf, sift, ferns and other.
Vuforia uses edge-detection technique. If there are more vertices or lines in the high-contrast image then its a high rated image for Vuforia. So, we can say that their algorithm is somewhat similar to SIFT.
Is there an android Framework that can be used in an app to recognize a 3D image and send the user to a video. This should fall under augmented reality, but so far everything I have viewed uses 2D image and stuff to produce a 3D image on the screen... My situation is backwards from that. I tried using vuforia but I couldn't get the sdk to work, and unity needs an android license. DroidAr doesn't seem to fit the bill either. Or are there any tutorials on this matter? Thanks.
I have not used the feature, but Metaio has a 3D object, "markerless" tracking feature as well as the ability to do video playback within the SDK. I am sure if you would rather simply redirect to a video (YouTube) or something this would not be exceptionally difficult.
http://www.metaio.com/software/mobile-sdk/features/
Metaio's mobile SDK is similar to Vuforia, so if you had trouble with that you might have difficulty getting it up and running. If your programming skills aren't up to that, you might consider looking into Junaio, an AR browser made by Metaio. With Junaio you simply create a content channel rather than having to build the app from scratch. Again, I have not actually tried this feature yet but the documentation seems to indicate that 3D tracking is available in Juniao:
http://www.junaio.com/develop/quickstart/3d-tracking-and-junaio/
Good luck!