I am writing Android app to stream audio/video to Wowza server in RTSP interleaved mode. Using AAC and H.264 encoders. I created packetizers for both audio and video. The problem that I am facing is that when I send both streams simultaneously I am losing video stream. I only get audio on Wowza and VLC. When I do not stream audio video works just fine. This proves that my packetizers and RTP streaming code perform as expected. It looks as if I cannot send video fast enough to sustain the stream.
Similarly architected code on iOS provides stable video and audio feed.
Thank you
Related
I have used javacv library to stream the live video from mobile to wowza media server and it's working fine. And I tried to record the live video in media server for broadcasting the video to multiple devices but I can hear only audio from the recorded video file.
Please let me know where I have done wrong and let me know the steps for re-stream the video to multiple device. I'm new to the Media server platform.
Thanks.
Has anyone had any success playing back an audio RTMP stream on Android using http://code.google.com/p/android-rtmp-client or know of any other non-flash solutions. The example that comes with the android rtmp source records the audio to a file but I'm looking for example code that plays back over speakers (or bluetooth).
The easiest way to play an audio RTMP stream in Android has a partial discussion of what's needed.
I would like to stream microphone.
And use http server on android so that
user just go http://xxxx.xxx.xx.xxx/xxx.wav can listen to what I say?
How can I do so?
I would try to develop a small HTTP server which serves an FLV stream.
You can take ipcamera-for-android as an example. This app serves a FLV video stream however you could reuse the server and the FLV encoder part.
As FLV supports PCM streams you can simply copy the buffer of the microphone to your stream.
Another variant
You can encode the microphone stream using the built-in MediaRecorder.AudioEncoder.AAC. Afterwards you can simply serve the AAC as a stream to your client.
I'm playing on my Android Nexus One some videos of a few cisco cameras using a VideoView. While this works fine, I'm unsure if it's possible to save the movie to a file.
I'm opening an URL like rtsp://192.168.1.22:554/live.sdp
How can I save it to the SDcard ? Handle it like a file maybe ... Is that possible ?
You can implement/use library, your own RTSP client which will pipe the incoming RTP packets into a file.
if you want to also play the video stream you can then give the media player a local RTSP server address to your RTSP server and then pipe the same RTP packets to the media player as well.
If you need you can find an open source RTSP server/client here
I'm creating mobile version of YouTube-like website. And i'm going to use Darwin Streaming Server for streaming low-res clips compressed with x264 and NeroAAC and hinted with MP4Box.
When i'm playing it with VLC player streaming works fine, as expected. When i'm trying to open URL with my HTC Hero, it switches to player mode, then starts "loading video" animation, then after some time it shows "unable to connect to server". Sample movies provided with DSS streamed fine regardless bitrate.
I tried few encoding options, but always the same result. I suspect nocabac and level=11 but it didn't changed nothing.
Is there some more specific encoding options for such type?
Thank You!
264 and AAC are supported, but is better to use:
a)RTSP streaming URL for mobile video playback. H.263 video and AMR audio.
or
b)RTSP streaming URL for mobile video playback. MPEG-4 SP video and AAC audio.
Android Streaming is only supported with 3gp format files
rtsp://myStreamingserver/myvideo.3gp