I have installed the 2.1 update on my HTC Hero yesterday, and happy as a puppy I instaleld my streaming app... Just to discover that it won't stream decently... It has to stream an mp3 stream and it works on my ADP 1.6 device, on a Samsung Galaxy S test-device (with 2.1) and on the 2.1 emulator...
Logcat is gettign the follwoing errors :
06-30 10:14:18.360: INFO/AudioPolicyService(56): stopOutput(): before lock
06-30 10:14:21.259: DEBUG/dalvikvm(1906): GC freed 28 objects / 1944 bytes in 186ms
06-30 10:14:21.439: INFO/AudioHardwareMSM72XX(56): AudioHardware pcm playback is going to standby.
06-30 10:14:21.439: DEBUG/StreamProcess(56): OutputStreamProcess::ResetAOLC()
06-30 10:14:22.919: INFO/AudioPolicyService(56): startOutput(): before lock
06-30 10:14:22.987: INFO/AudioHardwareMSM72XX(56): AUDIO_START: start kernel pcm_out driver.
06-30 10:14:26.329: DEBUG/dalvikvm(287): GC freed 11 objects / 488 bytes in 268ms
06-30 10:14:47.789: INFO/AudioPolicyService(56): stopOutput(): before lock
06-30 10:14:50.879: INFO/AudioHardwareMSM72XX(56): AudioHardware pcm playback is going to standby.
06-30 10:14:50.879: DEBUG/StreamProcess(56): OutputStreamProcess::ResetAOLC()
06-30 10:14:53.911: INFO/AudioPolicyService(56): startOutput(): before lock
06-30 10:14:54.022: INFO/AudioHardwareMSM72XX(56): AUDIO_START: start kernel pcm_out driver.
The result is a playing player, which stops (pauzes) every 10 seconds, the restarts This will work for about 8-10 restarts and then will simply stop without ever thrwoing an exception or something like that Which makes it somewhat difficult (if not impossible) to catch for us developers So has anyone been experiencing the same problem ? Does anyone have a clue if in this update another OpenCore lib has been used ? Does anyone have any solution ?
Right is updated with this new OTA update, however, i forgot to put my goldcard in so now ive lost wifi and god knows what else, i remember there beign a thread (maybe to do with orange handsets) but i cant find it. And is it fixable?
Is anyone else having this issue? Ever since the OTA update yesterday, if I am on the phone and another call comes in, at that very moment, the current call I am on gets disconnected and I have to click either "answer" or "reject" button. Its like the WAITING part of call waiting doesn't work anymore. like the phone can only handle 1 call at a time. I've tried rebooting but dont know what else to try.
I did the over the air 2.1 update and now my People app sync'd all 2000 of my Gmail contacts and there appears to be *no* option to sync just the 'My Contacts' group of my Gmail account.There was an option for that in Eris 1.5 (although I remember it being a little hard to find). But I've searched everywhere, and there is no "Sync Groups" option on any of the tabs of the People app. I also tried wiping the Contacts Storage app and resyncing, but that was no help. There are no options in Settings->Accounts & Sync either.This is a crippling regression bug, how could they have shipped the 2.1 update like this? Or am I just missing some hidden sync option?
I received notification this morning that my phone was ready for an update! Unfortunately after the update rebooted the phone it stopped at the Moto symbol and would not go any further. Called verizon and walked throught he normal battery pull and factory reset to no avail. The only answer they had was to send me a replacement phone which will be here to tomorrow. Bummer it broke but good customer service softened the issue.
I'd like to be able to keep tabs on my dogs while I'm away. Ideally, I picture an app running on my PC that can stream audio to my cell, when I run an app on my cell. Any ideas around this? And there can be monkeys in the middle (PC to webpage to phone).
For my app, I need to record audio from MIC on an Android phone, and send it over TCP to the other android phone, where it needs to be played. I am using AudioRecord and AudioTrack class. This works great with a file - write audio to the file using DataOutputStream, and read from it using DataInputStream. However, if I obtain the same stream from a socket instead of a File, and try writing to it, I get an exception. I am at a loss to understand what could possibly be going wrong.
I have developed a streaming app for Android. My app is streaming AAC audio.The app works fine on 1.5, 1.6, and 2.1. I tested it on Froyo, the android 2.2 and it does not work there.I have try everthing - Streaming directly from server url - Download incrementaly the content, store in a temp file, start reading. - Download incrementaly the content, store in a ContentProvider and start reading.No luck, I always got errors : 06-23 02:47:02.257: ERROR/MediaPlayer(266): Unable to to create media player 06-23 02:47:02.257: INFO/System.out(266): setDataSourceFD failed.: status=0x80000000 Does Froyo support AAC? Anyone happens to stream AAC?
I am working on a music streaming application and cannot find APIs I can use for it (MediaPlayer supports files only local or HTTP). I have a few questions and would appreciate if someone can help me answering them.
Does Android SDK 1.5 support audio streaming (MP3 stream for example)? If does what API calls I can use? If doesn't what are the plans for future SDK releases that would support streaming? Are there any suggested work arounds until the streaming is supported (like temp files etc)?
When I pair my sprint htc hero to my car stereo, it always opens the default music player and starts playing a song. Can I prevent this from happening? I generally want to listen to pandora - and not my mp3's. Most annoying part is that the mp3's play simultaneously with pandora until I manually open that music app and pause the music.
There must be a way to either turn off the default action or to plug another program name into that defautl.
I'm planning on building an application where audio media is going to be streamed to the mobile phone for the user to listen.The targets are smartphones: iPhone/Blackberry/Android/(J2ME ?).I see that streaming on iPhone has to be done with HTTP Live streaming, but I don't see it supported by other platforms.Should I broadcast the streams via rstp ? http ? Is there any way to use a unified solution for all the different mobile platform ? If anyone already had to go through this, help would be greatly appreciated.
I am new in Android. I am using android os 2.2. I am trying to play streaming audio and video from url. I don't know what to do for that. after 3 days og googling i come to know that I have to use mediaPlayer and MediaController classes.I had use different differend combination of codes found on internet but not succeed.Can anybody guide me what to do for streaming audio/video playing. What about the UI if we use MediaPlayer or Controller class. is there any thing to related these in xml file of layout.if not then even please tell me what may be the code for that.please help me.
I'm working on an app for a client that requires streaming of an AAC audio stream. Unfortunately, there's nothing I can do about the stream format on the server. I'm working on Android and have discovered that Android's media player does not support raw AAC streams (which is what I'm getting). I found a project on Google Code that supports it (I tested it with the stream) but it's GPL'ed and that doesn't work for my client. I don't have much experience with this sort of thing so forgive me if my ideas aren't great. I know Android can play AAC encoded content if it is in an MP4 wrapper so I had thought about creating an MP4 wrapper on the fly on the client-side or perhaps even just doing some conversion to another format on the fly. Are these reasonable options? Does anybody have better suggestions?
Edit To rephrase, is it possible to put a raw AAC stream from a web server in an MP4 container in real time? If so, does anybody know of resources to help me with the process?
I am trying to playback voice data which is received in form of RTP packets in my application. The class Receiver.java extends thread is playing back the audio. The question is 1. For real time streaming operation how to configure the write(byte, int offsetInBytes, int sizeInBytes) methods? 2. What is the function and purpose of getPlaybackHeadPosition() What is playback head position kindly explain! Note: The RTP packets I am receiving are not from a traditional Datagram Socket but rather then from Ad Hoc Network so there is unknow delay for receiving the RTP packets. I need to design the Receiver class to play back the voice as soon as packets are recieved. Kindly help me in this regard You received this message because you are subscribed to the Google Groups "Android Developers" group. To post to this group, send email to email@example.com To unsubscribe from this group, send email to android-developers+ unsubscribe@ google groups.com For more options, visit this group at http://groups.google.com/group/android-developers?hl=en
I want to stream audio directly from my mac via bluetooth to my nexus one. This might sound crazy, but I actually want to be able to control the audio from my mac. I just want to be able to connect an audio jack to my nexus one and let it stream the audio via bluetooth. Is there an app for that?
In my music app, i would like to play live streaming like radio broadcasting with mediaplayer as Mediaplayer mp = new Mediaplayer(); mp.setDataSource(LiveStreamingURL); mp.prepare(); mp.start(); but after playing for few seconds it is stopped. I dont know how to overcome it. Give me your suggestions for this.
Where to get streaming (live) video and audio from camera example for Android? Suppose I want to create some live video streaming service app so I'll have some cool server at the back end. And I know how to do that part. Suppose I have some stand alone app for PCs now I want to go on to mobile devices. So I want to see some sample app grabing audio and video streams from Phone, Synchronizing them, encoding somehow, and sending LIVE stream to server. I need any Open-Source sample that will do this or something like this. Where can I get such one?
The dev guide suggests this MediaPlayer mp = new MediaPlayer(); mp.setDataSource(this, Uri.parse("http://wamu-1.streamguys.com:80")); mp.prepare(); mp.start(); I can't seem to get it working though. I'm getting "java.io.IOException: Prepare failed.: status=0x1"
I am having the following problem: I am using Android's MediaPlayer to play an mp3 through an HTTP connection. The audio is played correctly, but the MediaPlayer returns an incorrect duration (calling mediaPlayer.getDuration() ). Depending on the mp3 the duration is either shorter or longer. The code that plays the audio is the following:
I am currently new to android development I been working on the examples in the Dev Guide in the android website. I want to get a stream from a server I have to play in the emulator when I insert the url it doesn't seam to want to play. My question is there a way to get the emulator to play audio or is it all enabled also does MediaPlayer require a special kind of format like mp3 or ogg? his is the code i am running on my 'onCreate()' method. MediaPlayer mp - new MediaPlayer(); mp.setDataSource(MY_URL_); mp.prepare(); mp.start()
AudioRecord can record audio data in pcm format in realtime, but there is no audio-encoder who can encode pcm data to amr format. may be third-party java code can do this, but the performance may be very low.MediaRecorder just record audio data to disk. not realtime.It seems that this is a real miss-impossible?
I want to stream a media file on web, and android developer website has said that MediaPlayer.setDataSource() can set the data source (file-path or http/rtsp URL) to use. But I got an error in both G1 device and emulator when streaming a rtsp url file: Command PLAYER_PREPARE completed with an error or info PVMFailure error(1, -1). Does anyone know what is this error, or anyone knows where can I get the error description (1, -1) means? the code can work successfully when playing a audio/video file or streaming a http protocal sudio/video file, but can't stream rtsp protocal file. Is android not supported rtsp streaming?
My favorite TV and FM stations all have links on there websites for live streaming vidoe/audio. They all have Quicktime or Windows Media links but my Evo will not plahy them. I've tried several Radio apps on the Market but no luck with them either. Is this a planned fix or do I have to go back to to a I-phone or a windows based phone?
I can't seem to get my Desire to stream audio on my Parrot MK600 car hands free kit. It pairs and connects, I can make calls over the hands free and it downloads the contact but when I play music from the stock Music app it still comes out over the handset.
I subscribed to JetFlix, in the market. (Jetflicks! TV - Android app on AppBrain) It is a $3 app with a $10/month subscription. I watch it on my phone in my car at the lake during my lunch break. I know that probably sounds sad, but it is AWESOME. I'm a manager in the dang corporate world, and I forget that I am at work by the time I'm done with a movie and come on back to the office. It's my dirty little secret. ANYWAY, there are some great shows available on this application, and I was thinking of replacing my Netflix subscription with it if I could just stream it to my TV at home. I haven't bought an HDMI cable yet, because I don't have a use for one just yet. But is anything "streamable" to the tv? I read that only pictures and videos that were shot with the camera stream though the HDMI cable. If that is the case, what is the point? I keep all that stuff on a hard drive anyway. So my question: can I hook up an HDMI cable and stream Jetflicks to my TV? The Jetflicks movies play on the native Internet browser. So I guess a more accurate question would be: Does the native internet browser display on the TV through the HDMI cable? If so, is there audio?