Live video feeds in my own android application - android

I would like to develope a application for viewing a IP camera , DVR video feeds from my own application. Can anyone please tell me the best possible way to achieve it so that the delay is as minimum as possible. What all servers are required to stream the video the formats etc..

androidfan , I believe you need to setup a media server to send & receive video streams.
Red5 will be a good option in your case.
Red5 is an Open Source Flash Server written in Java that supports: Streaming Video , Audio & RTMP protocols.

Related

How to Live stream RTMP/RTSP to a WOWZA server on Android

I am developing an Android application in which I want to publish as well as stream a video...
What I want is:
My app records a video and that video is sent to the server
The recorded video will be streamed live to another Android device at the same time..
P.S.:
I am using wowza server and RTMP stream. I would like to stream RTMP video (.flv)... If no solution is available, I would like to switch to RTSP and for that also, need a working link to follow.
Are there any suite of classes that make this easy?
Using libstreaming you can stream live video to the Wowza server form the android device running Android 4.0(API Level 14),or more recent.
For More details Check this link: https://github.com/fyhertz/libstreaming/wiki/Using-libstreaming-with-Wowza-Media-Server
Download the examples from here: https://github.com/fyhertz/libstreaming-examples

Live Video Broadcasting in Android using Wowza Server

I have tried broadcasting videos in Android app through Wowza server and it works pretty well with this demo of javaCV.
But the only problem with this is that it uses flv file format to broadcast on Wowza server. It uses FFmpegFrameRecorder to broadcast live video to wowza server (not VOD). To set format of the video broadcasted at server, it uses following method:
recorder.setFormat("flv");
So the main problem with this method is that when I use other formats, it doesn't work. For E.g.:
recorder.setFormat("mpeg"); //or something like mov, etc
Please suggest a way to broadcast video from Android device to wowza server with any format other than flv. Please provide any links or tutorials with which I can start.
Thanks in advance.
Edit: I am able to play only audio while streaming vlc video format in Android and iOS devices.
Your issue is not container, but protocol. For live streaming, your best options are RTMP, RTP/RTSP. RTMP is essentially FLV with VCR style commands (play pause stop). There is no streaming protocol for MP4, except HTTP. and this would require you to produce a new mp4 ever few seconds and reassemble server side.
Im not sure you FLV/RTMP is off limits to you, because it probably is the best. But next I would suggest RTSP, and maybe WebRTC.

Playing video from custom live stream (Android)

My android app needs to play a live video from a remote RTMP server (Adobe Flash Media Server).
As android's android.media.MediaPlayer doesn't support rtmp protocol, I found a library which provides me with streaming functionality. I.e. I can connect and receive portions of video stream as byte array.
The question is how I can use this incoming video stream data to display it in a view?
It seems that current standard API doesn't allow me to do that. MediaPlayer accepts either file or url. For audio data there is an AudioTrack which allows to achieve similar goal to mine, but for audio. For video I don't see an option.
Any suggestions are appreciated, maybe there is a third party media library which may provide the functionality for Android.

Possible to capture video and stream it to flash server on realtime in android?

I working on a project where the client side need to capture video and audio from the camera, use some library (proberly ffmpeg) to convert from mp4 to flv and send it to flash server on realtime,in the other side the client need to get flv and convert it to video type of android on real time. Is it possible to do??
Tnx
If you use Adobe Air (with or without Flex), your client can access an Android device's camera directly. You can then use a NetConnection and NetStream to publish the camera feed directly to a Flash Media Server, without having to transcode the video.
It really can be as straight forward as this example shows :)

Mandatory to use Darwin or wowza or VLC to stream live video in android?

I want to know is it mandatory to use any of the streaming servers like Darwin,Wowza or VLC to stream an RTSP live stream video? I am receiving an RTSP link from my client and it tends to change everytime. I can successfully play it in the VLC player but on phone I cant see anything. I tried playing a sample link having .3gp extension and it worked fine. But my links dont have an extension. They look like this rtsp://122.166.229.151:1950/1346a0cf0ef7c2. Please help me.If its compulsory to use an extension or a server, I will continue working in that direction.
A streaming server (as you describe) isn't strictly necessary - as long as you can pull RTSP from whatever your source is, you should be able to see it. Most IP cameras have onboard RTSP servers (although I wouldn't put too many connections on it). If you can see it in VLC, the phone should be able to consume it as well, given that the codec used to encode is one supported by the android device (in most cases, if you're doing H.264 Baseline 3.0 with AAC, you should be good to go).
A streaming server like Wowza can make that stream available to a wider audience than pulling directly from the source device, but if you're not intending to broadcast to a wide audience, it's not required for streaming to Android devices.
Newer versions of Android (Gingerbread and later) are also able to consume Apple HTTP Live Streaming.

Categories

Resources