Hey Guys,
I am looking for a source code\framework which will allow me to
record the device's springboard, in other words to "take video
captures of your screen".
Something like 'Display Recorder' for the iPhone, Link.
I already found this project "capture video", but unfortunately it will give low quality video (5/6 FPS)...
Is there any way to improve it?
Kind Regards,
M
Romain Guy posted recently about how he makes recordings of his Honeycomb device, and the solution was to feed the HDMI output into a video recording device attached to a desktop machine.
My suspicion is that if that's what he proposes, there's no way to do it that well on-device. Even if you can access the screen in all scenarios, many devices don't have the power to encode a video stream at good enough speed, which explains why Capture Video is giving such poor results.
Related
I would like to get started on this, but I can put everything on my mind right. So maybe someone could help me to know steps that are required to achieve this: I have laptop with integrated video camera and I would like to make it as IP camera and then embed it in my android app, that will let user to View that camera(my Laptop integrated camera) is streaming.
So far I came up with this:
I need to use software (webcamXP, webcam 7) to make my integrated laptop cam to an IP cam.
(Maybe there is a way that I would not need to do that, because software developers put their logo on streaming video)
I will probably need some kind decoder? WebCam 7 let's me stream video as : Javascript, Motion JPEG, Flash JPEG stream, Flash FLV stream.
And that is, I don't know what I will need more, any help will be appreciated. :)
It is possible do a little bit of research
I have searched a lot on google but unfortunately have not found any solution of screen video recording. If you have used the Screencast Video Recorder from google play, you can better understand my question. Can anyone help me out that how to achieve screen video recording programmatically.
AFAIK Screencast records by reading /dev/graphics/fb0, which is slow.
Programmatically the highest performance method is to create a virtual display. Currently only works on Android 4.4. See the screenrecorder.cpp for how to do this:
https://android.googlesource.com/platform/frameworks/av/+/android-4.4_r1/cmds/screenrecord/screenrecord.cpp
For Android 4.1 and above you can use the SurfaceComposerClient (both from native or Java code), although it's fairly broken on 4.3.
Unless you want to write a screen recorder yourself I wouldn't bother and just use one of the off-the-shelf recorders.
for some more detail see this answer.
I have a requirement where I need to detect human prescence from a live video feed which is coming from a phones camera
I was wonering if this was possible from the phone itself on an android phone (latest models such as the HTC Desire HD and higher perhaps)? And if possible can anyone guide me to a place (with links or such) to get an idea on how to proceed
However if this was not possible from within the phone itself is it possible to take a live video stream from the phone and transmit it to a server, which for example process the feed using open cv and sends an output back to the phone, can anyone tell me if the transmission of the live video feed to a server is possible and any guidance is appreciated as well.
Any suggestions?
Check this out
https://github.com/billmccord/OpenCV-Android
It may not be the most convenient thing to set up. But seems like it would work.
This might be helpful too.
http://www.slideshare.net/pickerweng/opencv-220-for-android
I've been asking around in developer circles about this, but so far no one has been able to confirm if it is possible. The default behavior for recording video in Android is to not use the external mic. I'd like to create an app that uses an external mic if it is available, but it seems like this might be tricky for some reason. Anyone have insight into this?
It seems like it would just be a matter of selecting it at this point in the recording setup:
recorder.setAudioSource(MediaRecorder.AudioSource.DEFAULT);
but it seems like there may be some oddness in doing that.
Thanks,
Jon
Default behavior in Android 12 (and I understand starting at 8 or 9) is that external microphones become the primary and default source for input.
It probably depends on the app, but at least video in the default camera app takes my external mic as default.
Take a look at the documentation here (look in the section titled "Performing Audio Capture") it covers setting up the audio capture.
I'm researching the development of an Android (2.2) app/service that will enable users to record short (I do emphasize short, < 30seconds) video on their phones and then upload that video (HTTP) to a server that will then transcode the video to other formats. That same user can download videos from other Android users and play them.
Now, I get a bit lost with everyones recommended approaches to all the issues in doing something like this because I haven't seen any ask this in a cohesive context. Ideally I would like a non commercial solution to this (as in no vendor/service being needed for the the video hosting/transcoding), but, feel free to include those as a recommendation (I've marked this as a wiki) as I know many like to use youtube and vimeo for the middle layer in all this.
The questions are
What server technologies do you
recommend for hosting and
transcoding?
What technology do you
recommend for streaming the video (it
would be nice to offer a high and
low quality encoding depending on
the users network connection)
What video format and software do you recommend for converting the uploaded video on the server to be viewable later by other Android owners.
Im assuming it's bad to do any transcoding on the phone prior to upload (battery/proc issues), but, if I'm wrong with that assumption what do you recommend?
Some things that may help you...
The video will only need to render on an Android device, and in the future in a webkit html5 browser.
Bandwidth isnt cheap (even with numerous 30 second videos), so a good mix of video quality and video file size is important (streaming if needed to ensure quality vs. download).
This is for android 2.2 devices with a video camera of course and medium to high density screen of 800x400 min.
Open source solutions (server to receive the uploads, code to do the transcoding, server to do the streaming) are preferred, but not required.
CDN's are an option, but I don't think that really figures in to the picture right now.
Check out this page to see all the video formats that Android supports for encoding and decoding.
http://developer.android.com/guide/appendix/media-formats.html
For encoding use FFmpeg or a service like encoding.com