I have an Icecast AAC+ (or AACP, AAC-HE) Live Audio Stream that I want to stream with HTML5 audio on a website. Playback on iOS Devices already works perfectly but I'm struggling with Android. I'm testing on a Samsung S3 with Android 4.4.3 and Chrome 39. Now to the strange part: When I press play and wait for about 1 minute playback starts. Can anyone explain why it takes so long for the stream to start?
Thanks in advance!
Related
I have a web application which plays an mp4 video at the beginning. This video is not playing in the latest version (v57) as well as in the previous version (v56) of Chrome in Android devices. The video was playing fine until Chrome v54.
It's an mp4 file. Has Google brought in any restrictions on the mp4 video files? The video plays fine in PC or Mac. The issue is only with Android devices.
Can someone throw some light?
Thanks.
I found if the Chrome Data Saver feature is turned on, videos would not appear or play.
I am building an application using electron.js that runs a socket server on a local network.
The idea is that the socket server sends messages to connected clients to dynamically load multiple html5 videos in a browser on a smart TV.
The videos are small demonstration videos and need to autoplay and loop.
There can be up to 12 videos playing at once on one device.
The problem that I am experiencing is that I cannot get smooth playback of the videos or not all videos will play.
I have tried the following with 12 videos:
Native Smart TV browser. Result: Only plays 3 videos and playback is ok
Chormium Browser on Raspberry Pi 3. Result: Can play all 12 videos but playback is choppy. Forcing hardware acceleration crashes the browser
Chrome Browser on Quad core Android box. Result: Can only play 6 videos and playback is good
ionic cordova app on Android Box. Result: Can only play 6 videos and playback is good
react-native app on Android Box. Result: Can only play 6 videos and playback is good
My question is, what are my limitations or possible solutions?
Am I limited by hardware or is there something I can do in my application builds that may resolve the issue?
It is worth pointing out that on my desktop chrome browser, all 12 videos play without any problem but this is not an option for me because they need to run on the TV.
Edit: It is worth mentioning I am using mp4 video
I am going to answer this myself to help others if they come across the same issue.
I ended up installing ffmpeg and encoded webm video instead of mp4 and could get all 12 videos playing.
Then by reducing the bitrate, I was able to get all 12 videos playing at the same time on the mini PCs at an acceptable quality.
Wowza Media Server running Live Video Streaming.
When I view the live video using RTMP and HLS streaming using wowza examples which has live video players for Flash and IOS. I am able to view video for both but whenever the camera is moved rtmp url shows live video without any delay but HLS stream shows the delay of 10 seconds.
Then I tried running a mobile application using cordova(phonegap) for ios devices. I am using HTML video tag in cordova application and I am able to view live video on IPad simulator using HLS streaming but whenever the camera moves there is a delay of 25 seconds while viewing the live video on IPad.
Can someone please let me know what configuration needs to be done on Wowza Server Side to reduce this delay in Live video streaming for IOS devices?
And also can someone please advice any other player other then HTML Video tag for cordova application?
Three chunks are required by iOS devices for streaming to begin. Each chunk is set to 10 seconds by default.If you use a keyframe interval of 1 frame per second you can lower the cupertinoChunkDurationTarget to 1 second(1000) and get the latency down to closer to 3 seconds.
Please check here for more : http://www.wowza.com/forums/content.php?88-Cupertino-Streaming-segmenter-parameters-%28iOS
I want to use HTML5 getusermedia to record audio, play back and save to server. I am trying with Google Nexus (Android 4+) and Google Chrome 29. Is it possible? When I try the codes I got from net, it asks permission to record using microphone and using the phone, I save a wav file. But no sound on playing this. The same site works from chromium browser in my PC. Am I doing anything wrong? or is it that audio recording is not actually working as it claims?
Thanks
We now have Web Audio input working in Chrome for Android Beta (31.0.1650.11).
I'm working on certain project that allows some video editing, uploading edited video to server (Amazon S3) and playing this video on device.
The iOS version is already completed. I'm using there GPUImage to edit and encode videos.
I'm using quicktime container and h264 codec.
Now I'm working on android version of this project and have encountered several problems.
While I'm able to download and play these videos (encoded on iPhone) on the android devices, I'm not able to play the stream of these videos on most of android devices.
I have done some research and one of advices was to add some code to iOS postencoding that converts to mp4 and fixes MooV http://fabiensanglard.net/mobile_progressive_playback/index.php, but it doesn't help.
The only thing that I have achieved after that "fix" is to hear audio track and not getting error message on Galaxy SII device, but on other devices there is the same story.
After that I have downloaded VLC player from play store and it plays these videos.
So my question is, what do I need to do to play h264 stream?
How can I implement my own video player or how can I use VLC source code to play these videos.