I'm actually doing my thesis right now, and I started doing it in unity 3d platform. What i'm doing is something to do with tuning music instrument. A drum tuner to be exact. So I need to use the microphone as a real-time listener just like the guitar tuners are (guitar tuna is my benchmark). But the forums and its manual says that the class microphone is only for recording and unity is not made for such app.
Is there anybody here tried something like my thesis app (or any tuning device)? what platform should i use? Please help me. Thank you!
Related
I want to broadcast video captured from extrenal camera and want to through live video on android and ios mobile application with least latency.
I am looking for help, either webrtc could help me out in this problem. If answer is yes. Kindly guide me in this Regard.
Requirements:
That is one way communication like Television, peers will just watch
video.
I want to scale my architect up-to thousands to millions of users(android/ios) in future.
If you are looking to build a highly scalable broadcast system then webrtc may not be the right choice. You should look for streaming servers which are more suited for broadcast.
For a complete list of streaming engines open the link below:
https://en.wikipedia.org/wiki/Comparison_of_streaming_media_systems
Some popular ones are wowza, red5, nimble etc.
I'm new to programming and I want to create an app. I have to use the phone's microphone in real time to implement it.
Since I have only used enviroments that i have been told in class I don't know what to use now. I have tried using corona but i heard that it doesn't implement audio (i'm not 100% sure about this).Is this true? (I am programming in Windows 8)
If i can't,what enviroment do you recommend that i should use? I only need the intensity of the voice i don't need full voice recognition.
Thanks in advance :)
Corona audio API only support audio playing, and does not give access to the microphone.
To support the microphone in your app, you can create a native plugin and integrate it using Corona Native
I have a plan to develop an instrument app, when we shake the android phone, it will produce "angklung" (Google it) sound.
THE PROBLEM:
How to make one android phone can share its produced sound (by shake
gesture) to the other android phones having my application?
The connection that I want to use is mobile data connection and wi-fi.
I think this person has the same problem, but I don't know how to communicate with him. Stream android to android
But there is no help..
I need solution/example/suggestion for this problem. So far I succeed to produce the "angklung" sound when it is shaken.
I have no idea how to start this application. I've searched in the internet but there is no help :(
Thanks for your help.
I would give you the suggestion of streaming the audio data to a server and beaming that to other android devices (that are registered to your app). As the question/issue you have asked are way bigger than couple of lines code, hence am pointing you to some good resources, dig those deep & good luck.
Live-stream video from one android phone to another over WiFi
Stream Live Android Audio to Server
Updated
I may have found the solution, I can probably make a mobile app with phonegap that talks to my rails app and make a lighter version of a mobile version of the rails app
I may be getting an educational rails application to work on soon. The client is an educational consultant, she wants to build a rails application and one of the feature is that a teacher can record a feedback that belongs to a particular student for an assignment. The app needs to be able to record an audio and play an audio.
I have been researching this feature both on stackoverflow and google but I don't find the answers very complete.
I got a couple of options,
1) I have found this blog, a developer built an audio recorder/player with flash http://cykod.com/blog/archive/December2010
and basically its teaching me to grab the audio from a browser with a microphone with flash (needs user permission) and then send it to the server therefore you can escape using red5 (media server).
She wants it to be mobile friendly. I don't think the recording works on the android phone because of flash and I am sure the android phone lacks a microphone. I don't know much about ios either because I don't specialize in mobile. I got my android phone to download flash so it can play the sound.
2) using red5server but I think the player/recorder will still be flash and it doesn't work on the phone
3) http://www.sajithmr.me/jrecorder-jquery and jrecorder
I got some solutions for rails but is there a mobile friendly solution ?
I am sure the android phone lacks a microphone
A phone without a microphone would be pretty useless! :)
With Android you can use MediaRecorder to record and save an audio file. I'm sure it works, I've done it before.
Although I have never tried, I think you can do it in PhoneGap with the Capture options and on iOS with the Audio Video Foundation framework.
That's for the mobile part. You don't need flash there, there are much better options.
On your web application, you can surely use Flash to record audio from the computer's microphone.
However, if you're lucky enough your users will use recent browser that support HTML5 audio recording and playback. Check out capturing audio tutorial and libraries such as audio.js.
HTH
I've been experimenting with making an android app for the past week or two and have reached a point where I need to decide on a general plan of action concerning the code I'm writing.
started with SoundPool, easy to use, not very flexible. went on to AudioTrack, seems good but slow.
So now I'm looking at the ndk..
Does the ndk have direct access to AudioTrack? or something else?
What is the general concensus on this kind of thing?
A guess is to make the UI in java and the 'sound engine' in C++
I'd like to get started on the right track before I write too much.
edit:
It is a music studio that plays and manipulates wav files from the sdcard as well as realtime sound synthesis.
The Android SDK doesn’t offer a fully featured audio processing solution like Core Audio on iOS. The available features are exposed through OpenSL ES interface. Any audio processing feature's availability is dependent on the device manufacturer’s Android flavor and configuration, so you can not rely on them.
To wit, the infamous Android fragmentation problem is even bigger in audio.
Even worse, if a device reports an audio feature available, it may still not work properly. For example: audio playback rate on a Samsung Galaxy Note III.
Realtime sound synthesis? None of the Android SDK’s features are designed for that.
The best way is doing the UI in Java and dealing with sound in C++.
There are some 'audio engine' offers on the web, but many of them are just “wrappers” around the same thing.
As cofounder of Superpowered, allow me to recommend the use the Superpowered Audio SDK, which is a total audio processing solution designed for real-time and highest performance, without any dependency on Android’s audio offerings, and runs on all.
http://superpowered.com/
There are a ton of Audio Latency issues in Android. There's really not anything that can be done about it. It seems like ICS (4.0) may have done some improvements on it, from what I've read.
You could subscribe to Andraudio and you'd actually be better off directing Android Audio questions through their emailing list than through Stackoverflow:
http://music.columbia.edu/mailman/listinfo/andraudio