I have a live event where we'd like to explore the use of tablets (IOS/Android/Windows) attached to seats for the audience to access webpages and applications, but would also like the functionally to push a video stream to each tablet simultaneously and synchronized to an audio feed from the house PA.
So far, I haven't found any hardware/software combination that can pull this off yet. Obviously syncing video over IP isn't an easy task, but I'm hoping someone has a clue if it's possible or not. I'd also explore the use of tablets that have a hdmi/video input, but so far they don't seem to exist either.
Flash Media Server and use p2p.
Just realized your using IOS. IOS doesn't do Flash.
Related
I would like to stream a video between two android devices (android-android). There wouldn't be any server, so the streaming has to be direct between devices. Devices would be in the same network so they could communicate via WiFi.
I've tried using MediaRecorder - MediaPlayer via sockets, but I've received many exceptions.
I also looked for library, but I just want to stream a video between two devices directly.
Any solutions?
If your video if for real time communication, e.g. a web chat or sharing some CCTV in real time with minimal delay then a real time video communication approach like WebRTC would be one additional possibility - this type of approach prioritises low latency over quality to ensure minimum delay. See here for Android WebRTC documentation:
https://webrtc.org/native-code/android/
If the requirement is just to allow one device act as a server for non-real time videos then the easiest approach may be to use one of the available HTTP server libraries or apps to allow one device act as a server that the other one can simply connect to via a browser or player. An example Android HTTP server that seems to get good reviews is:
https://play.google.com/store/apps/details?id=jp.ubi.common.http.server&hl=en
I have a plan to develop an instrument app, when we shake the android phone, it will produce "angklung" (Google it) sound.
THE PROBLEM:
How to make one android phone can share its produced sound (by shake
gesture) to the other android phones having my application?
The connection that I want to use is mobile data connection and wi-fi.
I think this person has the same problem, but I don't know how to communicate with him. Stream android to android
But there is no help..
I need solution/example/suggestion for this problem. So far I succeed to produce the "angklung" sound when it is shaken.
I have no idea how to start this application. I've searched in the internet but there is no help :(
Thanks for your help.
I would give you the suggestion of streaming the audio data to a server and beaming that to other android devices (that are registered to your app). As the question/issue you have asked are way bigger than couple of lines code, hence am pointing you to some good resources, dig those deep & good luck.
Live-stream video from one android phone to another over WiFi
Stream Live Android Audio to Server
I need to develop an application for Set top box, where in any Android mobile youtube application can connect to my youtube application and REMOTEly control STB youtube application. Such as play, pause, navigate..etc
As far as I have searched, youtube uses DIAL for Airplay kind of feature.
But, If I see youtube leanback pairing with mobile; It works on some pairing code; which does not look like DIAL specification.
Are there youtube Android applications, that works on DIAL?
Well, if you have been able to see the pairing code, you've seen the DIAL protocol in action. From there, a proprietary protocol takes over to control the device. (unfortunately, I haven't been able to figure out how it works exactly yet. It somehow proxies all commands through Google via an AJAX-based protocol).
You probably aren't even supposed to do what you want to do. If you are still trying to get YouTube on the set top box, you might want to look into https://www.youtube.com/tv and leapcast.
Oh and to clarify, unlike what almost anybody seems to think, DIAL is not like AirPlay. AirPlay streams data to another device, while DIAL just controls another device (the net effect might be the same though). DIAL is more like UPnP, and even uses a subset of UPnP.
UPDATE
I have reverse-engineered the protocol. The result is a lightweight audio-only media player acting like a YouTube TV. It can be found on GitHub.
Under the hood it uses Google's Channel API for near-real-time communication over HTTP without polling.
I have been working on an android app that streams videos live on a server using android built-in camera and anyone can watch that live stream from my website which is deployed on the server.
So can any one help me on how should i start working on my project because at present i have no direction to start with.
More specific example is:-
Like a person goes to a picnic and he wants his friends and family to see whats going on with the tour and his family can see live what he's doing live.....
There is an open-source project that does a very similar thing:
http://code.google.com/p/ipcamera-for-android/
It basically uses the LocalSocket of the camera to read the video and stream it from a webserver. You should be able to find lots of information in the source code.
If you want to stream over the internet, for everyone to see i can recommend you the service justin.tv which lets you broadcast you stream to the whole internet. If tried it, and it works very good!
However, if there's no wifi you will probably have a very laggy connection, unless you convert the video in a smaller size...
I have a requirement where I need to detect human prescence from a live video feed which is coming from a phones camera
I was wonering if this was possible from the phone itself on an android phone (latest models such as the HTC Desire HD and higher perhaps)? And if possible can anyone guide me to a place (with links or such) to get an idea on how to proceed
However if this was not possible from within the phone itself is it possible to take a live video stream from the phone and transmit it to a server, which for example process the feed using open cv and sends an output back to the phone, can anyone tell me if the transmission of the live video feed to a server is possible and any guidance is appreciated as well.
Any suggestions?
Check this out
https://github.com/billmccord/OpenCV-Android
It may not be the most convenient thing to set up. But seems like it would work.
This might be helpful too.
http://www.slideshare.net/pickerweng/opencv-220-for-android