Android 2.3.4 RTP Server and Client - android

I'm trying to implement RTP Server to transport Voice from Micro to another Phone. And RTP Client to play Audio from another Phone to my phone.
So do you have a library, demo or any mention please talk to me.
Additional, I also need a RTP Server application to test.
UPDATE
In case, I can get an URL like this: RTP://192.168.43.123:5678. So How to streaming it in Android 2.3.4 ?
Thank alot
Note: I must using Android 2.3.4 :(

I think you should have a look at this library: https://github.com/fyhertz/libstreaming
It uses a RTP Server and several Codecs, but unfortunately it is only for android 4.0 and above.
There is another library which works with API Level 8, but doesn't provide RTP: https://github.com/Teaonly/android-eye

Related

Transmit Live Android Audio without SIP

I am looking to transmit audio in realtime from an Android application I am working on to a server in a way similar to how a baby monitor functions (one way listening).
I created a test app that uses SIP to initiate a VOIP call between our client and server applications. The problem is that now I need a way to do this on non-SIP enabled devices. I have tried recording the audio from the device microphone into a buffer, then sending the buffer in chunks to the server through HTTP objects and re-assembling the audio for playback with poor results.
Does anyone have any suggestions for streaming realtime audio from an Android device to a server application for processing? SIP works so well, but I don't have time to implement a SIP stack on all of our non supported devices.
XMPP/jingle (aka gtalk) is the usual alternative. There are C libraries as well as some support in java using the smack libraries. (The smack jingle support is old and doesn't work well, but IIRC someone is working on a new version)

Internet Streaming Radio on an android application

I have an Internet Radio and to be honest I went through almost everything to get how can I actually make a Streaming android application and what i should use for that .. no use .. not a single useful information ..
can anyone help me ?
This is most commonly done with Shoutcast/Icecast HTTP streaming to a MediaPlayer component contained in a Service.
NPR has open-sourced their app, and it's a great reference application for radio.
http://code.google.com/p/npr-android-app/
Note: Shoutcast streaming is currently broken in Google TV, so your app will not work on those devices unless you choose another protocol.
If you wish to only support android 3+ (including Google TV) you can also try streaming over HLS HTTP. This protocol is much less commonly used for radio, though.

Advice about streaming live video to android/ios/pc

i would like some advice about the best way to stream a only-video live stream from a server to:
Android (>4.0 is ok)
PC with web-browser
iOS
I would like to keep latency as low as 1/2 second.
I can use:
flash: works on PC but no iOS and no Android(works only on some tablets)
HLS: not good because of latency
proprietary library: it should work but i have to implement it everywhere
RTSP: works only on Android
Any other way? Is a proprietary library the way to go?
I'm working on Linux but i'm mainly interested in "use this technology" and not "use this code".
Not sure, but you can try HTTP streaming of MP4/3gp formats using a web server. Both Android and iOS supports HTTP streaming. But you need to implement Progressive Download.
Please specify on which OS you want to implement your server.
For Windows - you can use following binary to relocate your moov atoms to the beginning of media file to enable them for progressive download
http://notboring.org/devblog/2009/07/qt-faststartexe-binary-for-windows/
Let us know your progress.
You can implement FFmpeg Server for Live broadcast. It gives you various options. Enable/Disable options from its configuration file located at /etc/ffserver.conf
You can get detail documentation at
http://ffmpeg.org/ffserver.html
Rtsp might be the way to go , but that 1/2 second latency might be hard to get.
I guess for video only and if you don't buffer at all , this may work for ios anyway
https://github.com/mooncatventures-group/FFPlayer-tests
Android supports rtsp , but its not very good.
You can compile ffmpeg for android and write a simple player using OpenGL. I can't share the code because we did it for a client but its not to difficult.

Gstreamer to android

I'm trying to take a stream from a webcam and stream it to an android device. I use Gstreamer to grab the video and stream it out through a TCP server. That part of it works fine. The trouble I'm running into is that I need to make a custom app to receive the stream on the android and I can't get gst-android to compile (For reasons unknown to me, the adb is not a runnable, thus I can't set up the flingersinks.) Any suggestions? Is there something other and gst-android that I can use for this?
Which android version are you targeting? As far as I know, the ndk version of gstreamer will still have problems to render video as no one contributed a working video sink. The surfaceflinger api is not available to ndk apps :/

Mandatory to use Darwin or wowza or VLC to stream live video in android?

I want to know is it mandatory to use any of the streaming servers like Darwin,Wowza or VLC to stream an RTSP live stream video? I am receiving an RTSP link from my client and it tends to change everytime. I can successfully play it in the VLC player but on phone I cant see anything. I tried playing a sample link having .3gp extension and it worked fine. But my links dont have an extension. They look like this rtsp://122.166.229.151:1950/1346a0cf0ef7c2. Please help me.If its compulsory to use an extension or a server, I will continue working in that direction.
A streaming server (as you describe) isn't strictly necessary - as long as you can pull RTSP from whatever your source is, you should be able to see it. Most IP cameras have onboard RTSP servers (although I wouldn't put too many connections on it). If you can see it in VLC, the phone should be able to consume it as well, given that the codec used to encode is one supported by the android device (in most cases, if you're doing H.264 Baseline 3.0 with AAC, you should be good to go).
A streaming server like Wowza can make that stream available to a wider audience than pulling directly from the source device, but if you're not intending to broadcast to a wide audience, it's not required for streaming to Android devices.
Newer versions of Android (Gingerbread and later) are also able to consume Apple HTTP Live Streaming.

Categories

Resources