I have a plan to develop an instrument app, when we shake the android phone, it will produce "angklung" (Google it) sound.
THE PROBLEM:
How to make one android phone can share its produced sound (by shake
gesture) to the other android phones having my application?
The connection that I want to use is mobile data connection and wi-fi.
I think this person has the same problem, but I don't know how to communicate with him. Stream android to android
But there is no help..
I need solution/example/suggestion for this problem. So far I succeed to produce the "angklung" sound when it is shaken.
I have no idea how to start this application. I've searched in the internet but there is no help :(
Thanks for your help.
I would give you the suggestion of streaming the audio data to a server and beaming that to other android devices (that are registered to your app). As the question/issue you have asked are way bigger than couple of lines code, hence am pointing you to some good resources, dig those deep & good luck.
Live-stream video from one android phone to another over WiFi
Stream Live Android Audio to Server
Related
I'm currently working on an app that will stream video from one phone, to another using wifi-direct.
I've already installed and tested the sample wifi-direct app that comes with the SDK. It works great sending images from one Nexus 4 to another.
Some problems I've experienced:
I tried to have the app send a video by changing the "image/*" string to "video/*" wherever it appeared. And ".jpg" to ".mp4".
After running the app with these changes, I am still able to connect devices, but one N4 is stuck at "Opening a server socket". And the video I took never gets sent.
Perhaps I went about it wrong.
Regardless, my real goal is to stream video from one phone to the other using wifi-direct only (no data connection / wifi).
Could someone help me figure out what steps I should pursue in order to accomplish this? Should I start with the wifi-direct sample app as a base? Or should I try to write it from scratch myself? How should I go about streaming the video using wifi-direct?
Thanks!
Updated
I may have found the solution, I can probably make a mobile app with phonegap that talks to my rails app and make a lighter version of a mobile version of the rails app
I may be getting an educational rails application to work on soon. The client is an educational consultant, she wants to build a rails application and one of the feature is that a teacher can record a feedback that belongs to a particular student for an assignment. The app needs to be able to record an audio and play an audio.
I have been researching this feature both on stackoverflow and google but I don't find the answers very complete.
I got a couple of options,
1) I have found this blog, a developer built an audio recorder/player with flash http://cykod.com/blog/archive/December2010
and basically its teaching me to grab the audio from a browser with a microphone with flash (needs user permission) and then send it to the server therefore you can escape using red5 (media server).
She wants it to be mobile friendly. I don't think the recording works on the android phone because of flash and I am sure the android phone lacks a microphone. I don't know much about ios either because I don't specialize in mobile. I got my android phone to download flash so it can play the sound.
2) using red5server but I think the player/recorder will still be flash and it doesn't work on the phone
3) http://www.sajithmr.me/jrecorder-jquery and jrecorder
I got some solutions for rails but is there a mobile friendly solution ?
I am sure the android phone lacks a microphone
A phone without a microphone would be pretty useless! :)
With Android you can use MediaRecorder to record and save an audio file. I'm sure it works, I've done it before.
Although I have never tried, I think you can do it in PhoneGap with the Capture options and on iOS with the Audio Video Foundation framework.
That's for the mobile part. You don't need flash there, there are much better options.
On your web application, you can surely use Flash to record audio from the computer's microphone.
However, if you're lucky enough your users will use recent browser that support HTML5 audio recording and playback. Check out capturing audio tutorial and libraries such as audio.js.
HTH
I have been working on an android app that streams videos live on a server using android built-in camera and anyone can watch that live stream from my website which is deployed on the server.
So can any one help me on how should i start working on my project because at present i have no direction to start with.
More specific example is:-
Like a person goes to a picnic and he wants his friends and family to see whats going on with the tour and his family can see live what he's doing live.....
There is an open-source project that does a very similar thing:
http://code.google.com/p/ipcamera-for-android/
It basically uses the LocalSocket of the camera to read the video and stream it from a webserver. You should be able to find lots of information in the source code.
If you want to stream over the internet, for everyone to see i can recommend you the service justin.tv which lets you broadcast you stream to the whole internet. If tried it, and it works very good!
However, if there's no wifi you will probably have a very laggy connection, unless you convert the video in a smaller size...
I'd like to add NAS support to one of my application, and it is critical that the application is capable of streaming the content and that it not has to download it to the device. The application will be streaming video content, so once the video is over, there shouldn't be any large video files on the device.
What I've tried so far:
jCIFS - Works beautifully, but isn't capable of streaming (to my knowledge). I've successfully created video files on the device using jCIFS, but they're still there, when the video playback stops.
Temporary files - I know that Android is supposed to be able to support temporary files, but I'm not sure how it works or if it's any good in this situation. Just a thought, basically.
My application must be able to launch a video intent with a video on the NAS device, and it should be playable in any video player. I know that some applications on Market support NAS devices (and SMB / CIFS connections), but I don't know how it works.
Any suggestions or ideas would be much appreciated.
I think you have to create an in between http server as discussed in the following question: Android ServerSocket programming with jCIFS streaming files
I am trying to do the same thing, already tried VPNC, no luck.
Now I am trying to use neo router my android can connect to the server but isnt able to browse any files. It might work for you. search for VPN setup (your router name here) android. It will give you some stuff to try.
I have a requirement where I need to detect human prescence from a live video feed which is coming from a phones camera
I was wonering if this was possible from the phone itself on an android phone (latest models such as the HTC Desire HD and higher perhaps)? And if possible can anyone guide me to a place (with links or such) to get an idea on how to proceed
However if this was not possible from within the phone itself is it possible to take a live video stream from the phone and transmit it to a server, which for example process the feed using open cv and sends an output back to the phone, can anyone tell me if the transmission of the live video feed to a server is possible and any guidance is appreciated as well.
Any suggestions?
Check this out
https://github.com/billmccord/OpenCV-Android
It may not be the most convenient thing to set up. But seems like it would work.
This might be helpful too.
http://www.slideshare.net/pickerweng/opencv-220-for-android