I need to stream audio to my app as it is being recorded live. How do I do this for Android? I have searched a few sources and didn't get a straight answer on how to do this.
The official documentation says Android supports live streaming over http/https in Android 3.0+. I need to be able to stream to devices with 2.3+. Any suggestion would help me. Thanks!
I would recommend to implement an icecast/shoutcast server. You can find some tips about how to do the Android side on this question: Icecast/shoutcast streaming on Android
Related
I am going through webRTC for Android. I need to develop video and audio chat application. I read many things in webRTC. I am confused where to start with. I did not get proper link.Many said that refer below site.
https://webrtc.org/reference/getting-started
But I could not find this page itself. Please help me to build webRTC for Android development.
Note: I want opensource code. I don't want any licenced libraries.
Thank you for the help.
Check this, and you can start by studying this sample android app.
How to do live streaming in google glass through android code, i am using eclipse IDE to develop android apps. I have downloaded GDK and created the project using GDK. I have browsed for live video streaming, i didnt get any blogs related to that(i dont know how to use Mirror API to do live streaming )..Can anyone helped me to go up?
If you're using the GDK, then you can just use the standard android VideoView or MediaPlayer. It's explained pretty well in the Android API Guide. http://developer.android.com/guide/topics/media/mediaplayer.html
You probably want to build and test this in Android first (a phone or emulator). That should be faster since it's easier to interact with the touch screen. Once you have that working you can load it onto your Glass and proceed from there.
As far as I know there isn't a way to do this with the Mirror API. Only with the GDK.
I am working on an application for google tv with Visio device , i want to know which streaming protocols are avalable for google tv out of these. Can any body give me the reference along with the reply.
HTTP, RTMP, PNM, RTSP, MMS, RTSPU, RTSPT, MMSU, MMST.
Please help any kind of help will be appreciated.
Many thanks
HTTP is going to be your best bet on Vizio wrt ease of setup and development. You can use VLC to experiment with different stream proto's. The thing you should pay close attention to is what codec you are transmitting your media in - https://developers.google.com/tv/android/docs/gtv_media_formats
I need to know if android version 2.1 , 2.2 or 2.3 support online radio streaming, if not could you please tell me which one supports?
As far as I know, streaming audio is possible with all mentioned versions. One of the simpler examples I've seen can be found here http://blog.pocketjourney.com/2009/12/27/android-streaming-mediaplayer-tutorial-updated-to-v1-5-cupcake/
This example shows the usage of android.media.MediaPlayer with a HTTP protocol URL.
After carrying out a lot of research I have come to the conclusion that Java and the Java Media Framework (JMF) is not suitable for developing a streaming server that supports the RTSP protocol on the server side for streaming video and audio. I have read very good things about Live555 media server and the testOnDemandRTSPServer source code for a basis of design. My only worry is that this is written in C++ and I am a predominantly Java programmer. This server is a large portion of my final year project at university so my degree kind of hangs on its successful implementation and I am running out of time. If any one has any experience with implementing a RTSP server that can stream to an android handset or belive they can point me in the right direction to learn how to do it, please let me know. Thanks in advance.
My project also has the RTSP server module to be run on Android phone. I think we can build rtsp library as name.so file and can interface with java by using JNI.
This also works for Android!
http://net7mma.codeplex.com/
You can see the article on CodeProject # http://www.codeproject.com/Articles/507218/Managed-Media-Aggregation-using-Rtsp-and-Rtp
The live555 RTSP server is a fully fledged RTSP server that implements most payloads (H.263, H.264, MPEG2, PCM, AMR, AAC, etc. You can read up on the website whether it already supports the media types you want to stream. It also features an RTSP client. With respect to streaming to an android handset: that is the whole point of RTSP: it doesn't matter what type of client you're streaming to, and as for the server side development, there isn't really much dev to do, unless you need to implement an unsupported media type. The code can be quite complex if you're not well versed in c++, but it sounds like your goal is more related to setting up streaming to android as opposed to implementing the RTSP server and client? So check if live555 supports your media types and if it does, I wouldn't bother writing one in JAVA, that can be quite involved. If you do choose to go that route, your best friend is of course the RFC (http://tools.ietf.org/html/rfc2326).
As for the client, I'm not sure if android already has an RTSP library/client. The one other thing you have to consider is which media types are supported by android.