I'm about to make an app, I need to know if the person is running using the the front camera, so I need to know where are their eyes and the mouth.
The problem is that I've been searching but I don't find anything like that on Titanium just for pictures and I have to do it in real time.
is this possible on Titanium?, or I have to do it native?
sorry for my bad english
Update:
The app is for IOS and Android
For iOS you can use the following Titanium module which wraps the CIDetector framework:
http://gitt.io/component/by.farwayer.facerecognizer
Related
Question regarding using React Native. Is it possible to develop an app where it will record a video only without a camera view? Basically, an app that is just a video recorder without a monitor to see what the camera is seeing. If yes, I would appreciate if you can share a little bit of an idea on how to go about it.
Hiding it from the phone's viewing area whether by position, or giving it a width or height of 0, or making it transparent are not the kind of solutions that I'm looking for.
If you are expecting a module from RN in JavaScript that would allow you to do that, then I believe that module is not around. At least in the official framework.
What you could do, is to create a NativeModule. RN allows you to actually execute native code from Android and iOS and then communicate through a bridge.
You can see how to create this modules here: https://facebook.github.io/react-native/docs/native-modules-ios.html
I want to make an android application that has "augmented reality" functionality in android application like start the camera and fill the colors virtually in real time. Please give me suggestion on it.
Thanks in advance.
http://bitstars.github.io/droidar/
look in here...
Should be good base for your future work
There are a lot of SDK's available today to integrate augmented reality in android application.
You will find the below mentioned links useful to choose SDK based on your app requirement.
https://creator.zoho.com/reitmayr/augmented-reality-sdks/view-embed/AR_SDKs
http://socialcompare.com/en/comparison/augmented-reality-sdks
If you want to go with a example, try the below
https://www.sitepoint.com/how-to-build-an-ar-android-app-with-vuforia-and-unity/
We are working on a Mobile Application which has to be developed on IOS and Android Platforms. Most of the application is straight forward which is display content from a back end CMS system for which we are considering using PhoneGap. The application has one module in which the user can browse through a virtual house and be able to modify colors / replace a particular item like a chair etc. This can be easily done using Unity3d which we have in place. Using Unity3d for the whole application is probably going to be an overkill and there will be a lot of effort required on the Application Front or if we go the PhoneGap route managing the 3d thing is going to be a challenge.
The question I have is if there is any way we can export the Unity Scene and use it in a PhoneGap build since we will get best of both packages? If not that what are the options to get Unity3d working (UI and Services for Application) for doing an Application or doing some stuff with PhoneGap which helps to get the 3d part sorted.
Thanks in advance.
Figured out he solution.
Using Phonegap completely. For the 3d Module using Panorama image and displaying it using https://github.com/nicekei/jQuery-html5-canvas-panorama-plugin for interactivity. There are many other plugins available for jquery and mobile friendly. I choose to use the above one. You can find more on http://www.jquery4u.com/plugins/10-jquery-panorama-image-display-plugins/. For walk through using annotations to link it to another page which will give a 3D Panorama for that room.
You can render stills from a 3d software and stitching them using any Panoramic stitcher Photoshop also does the job. Hope it helps.
I want to develop a Camera 2 kind of application. So i just wanted to know how to apply live effects to android Camera .Is there any android library for that? plz help me to solve the issue.
thanks in advance....
What have you tried? I assume by live effects you mean shown on the preview.
You can try what the answer here suggested, but any sort of complicated filter implemented in java is going to slow things down. For good performance you'll probably have to write some native code to process the previews. The good news is, if you're writing NDK code you can make use of libraries like OpenCV to do the filtering for you.
Is there an android Framework that can be used in an app to recognize a 3D image and send the user to a video. This should fall under augmented reality, but so far everything I have viewed uses 2D image and stuff to produce a 3D image on the screen... My situation is backwards from that. I tried using vuforia but I couldn't get the sdk to work, and unity needs an android license. DroidAr doesn't seem to fit the bill either. Or are there any tutorials on this matter? Thanks.
I have not used the feature, but Metaio has a 3D object, "markerless" tracking feature as well as the ability to do video playback within the SDK. I am sure if you would rather simply redirect to a video (YouTube) or something this would not be exceptionally difficult.
http://www.metaio.com/software/mobile-sdk/features/
Metaio's mobile SDK is similar to Vuforia, so if you had trouble with that you might have difficulty getting it up and running. If your programming skills aren't up to that, you might consider looking into Junaio, an AR browser made by Metaio. With Junaio you simply create a content channel rather than having to build the app from scratch. Again, I have not actually tried this feature yet but the documentation seems to indicate that 3D tracking is available in Juniao:
http://www.junaio.com/develop/quickstart/3d-tracking-and-junaio/
Good luck!