I have developed a streaming app for Android. My app is streaming AAC audio.The app works fine on 1.5, 1.6, and 2.1. I tested it on Froyo, the android 2.2 and it does not work there.I have try everthing - Streaming directly from server url - Download incrementaly the content, store in a temp file, start reading. - Download incrementaly the content, store in a ContentProvider and start reading.No luck, I always got errors : 06-23 02:47:02.257: ERROR/MediaPlayer(266): Unable to to create media player 06-23 02:47:02.257: INFO/System.out(266): setDataSourceFD failed.: status=0x80000000 Does Froyo support AAC? Anyone happens to stream AAC?
I am streaming an MP3 audio and instead making this MP3 audio to play, I want it to be converted to PCM format and pass this PCM audio to AudioTrack(as it takes PCM data as input.) instead of passing my MP3 stream to MediaPlayer. Can anyone please help me to convert the MP3 to PCM stream?
I want to record audio in MP3 format in my android application. I know I can read PCM audio using the audioRecord class. Is there a java library compatible with android that will encode PCM to MP3? Is it possible to use the Lame mp3 encoder library?
According to the MS website, Windows Media Player queries an MTP device to see what formats it supports. My Razr Maxx HD supports ogg and m4a. When attempting to sync WMP does not sync those files stating they might not be supported on the device. Being that I've had so much trouble in the past even getting any of the Android devices to be properly detected as an MTP device I'm worried this is a driver issue.I should be able to get WMP to at least convert them.
this is the device mis-reporting the formats, Windows not querying them properly, or an actual WMP issue? Would like to be able to get sync working without any conversions. Maybe there is a way to hardcode formats within the mtp inf file?
I'd like to be able to keep tabs on my dogs while I'm away. Ideally, I picture an app running on my PC that can stream audio to my cell, when I run an app on my cell. Any ideas around this? And there can be monkeys in the middle (PC to webpage to phone).
For my app, I need to record audio from MIC on an Android phone, and send it over TCP to the other android phone, where it needs to be played. I am using AudioRecord and AudioTrack class. This works great with a file - write audio to the file using DataOutputStream, and read from it using DataInputStream. However, if I obtain the same stream from a socket instead of a File, and try writing to it, I get an exception. I am at a loss to understand what could possibly be going wrong.
I am working on a music streaming application and cannot find APIs I can use for it (MediaPlayer supports files only local or HTTP). I have a few questions and would appreciate if someone can help me answering them.
Does Android SDK 1.5 support audio streaming (MP3 stream for example)? If does what API calls I can use? If doesn't what are the plans for future SDK releases that would support streaming? Are there any suggested work arounds until the streaming is supported (like temp files etc)?
I have installed the 2.1 update on my HTC Hero yesterday, and happy as a puppy I instaleld my streaming app... Just to discover that it won't stream decently... It has to stream an mp3 stream and it works on my ADP 1.6 device, on a Samsung Galaxy S test-device (with 2.1) and on the 2.1 emulator...
Logcat is gettign the follwoing errors :
06-30 10:14:18.360: INFO/AudioPolicyService(56): stopOutput(): before lock 06-30 10:14:21.259: DEBUG/dalvikvm(1906): GC freed 28 objects / 1944 bytes in 186ms 06-30 10:14:21.439: INFO/AudioHardwareMSM72XX(56): AudioHardware pcm playback is going to standby. 06-30 10:14:21.439: DEBUG/StreamProcess(56): OutputStreamProcess::ResetAOLC() 06-30 10:14:22.919: INFO/AudioPolicyService(56): startOutput(): before lock 06-30 10:14:22.987: INFO/AudioHardwareMSM72XX(56): AUDIO_START: start kernel pcm_out driver. 06-30 10:14:26.329: DEBUG/dalvikvm(287): GC freed 11 objects / 488 bytes in 268ms 06-30 10:14:47.789: INFO/AudioPolicyService(56): stopOutput(): before lock 06-30 10:14:50.879: INFO/AudioHardwareMSM72XX(56): AudioHardware pcm playback is going to standby. 06-30 10:14:50.879: DEBUG/StreamProcess(56): OutputStreamProcess::ResetAOLC() 06-30 10:14:53.911: INFO/AudioPolicyService(56): startOutput(): before lock 06-30 10:14:54.022: INFO/AudioHardwareMSM72XX(56): AUDIO_START: start kernel pcm_out driver.
The result is a playing player, which stops (pauzes) every 10 seconds, the restarts This will work for about 8-10 restarts and then will simply stop without ever thrwoing an exception or something like that Which makes it somewhat difficult (if not impossible) to catch for us developers So has anyone been experiencing the same problem ? Does anyone have a clue if in this update another OpenCore lib has been used ? Does anyone have any solution ?
When I pair my sprint htc hero to my car stereo, it always opens the default music player and starts playing a song. Can I prevent this from happening? I generally want to listen to pandora - and not my mp3's. Most annoying part is that the mp3's play simultaneously with pandora until I manually open that music app and pause the music.
There must be a way to either turn off the default action or to plug another program name into that defautl.
I'm planning on building an application where audio media is going to be streamed to the mobile phone for the user to listen.The targets are smartphones: iPhone/Blackberry/Android/(J2ME ?).I see that streaming on iPhone has to be done with HTTP Live streaming, but I don't see it supported by other platforms.Should I broadcast the streams via rstp ? http ? Is there any way to use a unified solution for all the different mobile platform ? If anyone already had to go through this, help would be greatly appreciated.
I am new in Android. I am using android os 2.2. I am trying to play streaming audio and video from url. I don't know what to do for that. after 3 days og googling i come to know that I have to use mediaPlayer and MediaController classes.I had use different differend combination of codes found on internet but not succeed.Can anybody guide me what to do for streaming audio/video playing. What about the UI if we use MediaPlayer or Controller class. is there any thing to related these in xml file of layout.if not then even please tell me what may be the code for that.please help me.
I'm working on an app for a client that requires streaming of an AAC audio stream. Unfortunately, there's nothing I can do about the stream format on the server. I'm working on Android and have discovered that Android's media player does not support raw AAC streams (which is what I'm getting). I found a project on Google Code that supports it (I tested it with the stream) but it's GPL'ed and that doesn't work for my client. I don't have much experience with this sort of thing so forgive me if my ideas aren't great. I know Android can play AAC encoded content if it is in an MP4 wrapper so I had thought about creating an MP4 wrapper on the fly on the client-side or perhaps even just doing some conversion to another format on the fly. Are these reasonable options? Does anybody have better suggestions?
Edit To rephrase, is it possible to put a raw AAC stream from a web server in an MP4 container in real time? If so, does anybody know of resources to help me with the process?
I need to record audio and video files in 3gp/mp4 format at the same file and same time.
When i run my application file is created with videofile.3gp but video does not record in sd card on the emulator.please let me clear one thing does video records on the emulator? if i run this code on android supported device do these error clear. Please help me. I am new to this topic.please find the code and errors below.
* @copy Rights * audio.java * sample code for Eminosoft Developerworks Article * Android development Team * www.eminosoft.cm
I am trying to playback voice data which is received in form of RTP packets in my application. The class Receiver.java extends thread is playing back the audio. The question is 1. For real time streaming operation how to configure the write(byte, int offsetInBytes, int sizeInBytes) methods? 2. What is the function and purpose of getPlaybackHeadPosition() What is playback head position kindly explain! Note: The RTP packets I am receiving are not from a traditional Datagram Socket but rather then from Ad Hoc Network so there is unknow delay for receiving the RTP packets. I need to design the Receiver class to play back the voice as soon as packets are recieved. Kindly help me in this regard You received this message because you are subscribed to the Google Groups "Android Developers" group. To post to this group, send email to email@example.com To unsubscribe from this group, send email to android-developers+ unsubscribe@ google groups.com For more options, visit this group at http://groups.google.com/group/android-developers?hl=en
I want to stream audio directly from my mac via bluetooth to my nexus one. This might sound crazy, but I actually want to be able to control the audio from my mac. I just want to be able to connect an audio jack to my nexus one and let it stream the audio via bluetooth. Is there an app for that?
In my music app, i would like to play live streaming like radio broadcasting with mediaplayer as Mediaplayer mp = new Mediaplayer(); mp.setDataSource(LiveStreamingURL); mp.prepare(); mp.start(); but after playing for few seconds it is stopped. I dont know how to overcome it. Give me your suggestions for this.
Where to get streaming (live) video and audio from camera example for Android? Suppose I want to create some live video streaming service app so I'll have some cool server at the back end. And I know how to do that part. Suppose I have some stand alone app for PCs now I want to go on to mobile devices. So I want to see some sample app grabing audio and video streams from Phone, Synchronizing them, encoding somehow, and sending LIVE stream to server. I need any Open-Source sample that will do this or something like this. Where can I get such one?
The dev guide suggests this MediaPlayer mp = new MediaPlayer(); mp.setDataSource(this, Uri.parse("http://wamu-1.streamguys.com:80")); mp.prepare(); mp.start(); I can't seem to get it working though. I'm getting "java.io.IOException: Prepare failed.: status=0x1"
I am having the following problem: I am using Android's MediaPlayer to play an mp3 through an HTTP connection. The audio is played correctly, but the MediaPlayer returns an incorrect duration (calling mediaPlayer.getDuration() ). Depending on the mp3 the duration is either shorter or longer. The code that plays the audio is the following:
My application cannot handle the JNIE overhead, as I need to make 2 JNIE calls to the native C/C++ code for every 20msec. So I am trying to use the native media libraries and the realted C/C++ files provided by Android in the folder "//external/srec/audio/test" of the donut build. Ex. AudioHardwareRecord, AudioInRecord etc.. I am implementing all the calls in the native layer.
I am using the sampling rate as 8000 and want to read 160 samples (320 bytes) at a time.
I used the following values to create the AudioRecord..
It says "success" after creating this interface but I am getting the error "Bad Value" if I am doing the InitCheck.
And it is crashing once I start reading the samples or if I try to retrieve the configured sampling rate.