Android :: How To Play Streaming Audio / Video From URL
Nov 1, 2010
I am new in Android. I am using android os 2.2. I am trying to play streaming audio and video from url. I don't know what to do for that. after 3 days og googling i come to know that I have to use mediaPlayer and MediaController classes.I had use different differend combination of codes found on internet but not succeed.Can anybody guide me what to do for streaming audio/video playing. What about the UI if we use MediaPlayer or Controller class. is there any thing to related these in xml file of layout.if not then even please tell me what may be the code for that.please help me.
My favorite TV and FM stations all have links on there websites for live streaming vidoe/audio. They all have Quicktime or Windows Media links but my Evo will not plahy them. I've tried several Radio apps on the Market but no luck with them either. Is this a planned fix or do I have to go back to to a I-phone or a windows based phone?
In my music app, i would like to play live streaming like radio broadcasting with mediaplayer as Mediaplayer mp = new Mediaplayer(); mp.setDataSource(LiveStreamingURL); mp.prepare(); mp.start(); but after playing for few seconds it is stopped. I dont know how to overcome it. Give me your suggestions for this.
The dev guide suggests this MediaPlayer mp = new MediaPlayer(); mp.setDataSource(this, Uri.parse("http://wamu-1.streamguys.com:80")); mp.prepare(); mp.start(); I can't seem to get it working though. I'm getting "java.io.IOException: Prepare failed.: status=0x1"
I am currently new to android development I been working on the examples in the Dev Guide in the android website. I want to get a stream from a server I have to play in the emulator when I insert the url it doesn't seam to want to play. My question is there a way to get the emulator to play audio or is it all enabled also does MediaPlayer require a special kind of format like mp3 or ogg? his is the code i am running on my 'onCreate()' method. MediaPlayer mp - new MediaPlayer(); mp.setDataSource(MY_URL_); mp.prepare(); mp.start()
Where to get streaming (live) video and audio from camera example for Android? Suppose I want to create some live video streaming service app so I'll have some cool server at the back end. And I know how to do that part. Suppose I have some stand alone app for PCs now I want to go on to mobile devices. So I want to see some sample app grabing audio and video streams from Phone, Synchronizing them, encoding somehow, and sending LIVE stream to server. I need any Open-Source sample that will do this or something like this. Where can I get such one?
I subscribed to JetFlix, in the market. (Jetflicks! TV - Android app on AppBrain) It is a $3 app with a $10/month subscription. I watch it on my phone in my car at the lake during my lunch break. I know that probably sounds sad, but it is AWESOME. I'm a manager in the dang corporate world, and I forget that I am at work by the time I'm done with a movie and come on back to the office. It's my dirty little secret. ANYWAY, there are some great shows available on this application, and I was thinking of replacing my Netflix subscription with it if I could just stream it to my TV at home. I haven't bought an HDMI cable yet, because I don't have a use for one just yet. But is anything "streamable" to the tv? I read that only pictures and videos that were shot with the camera stream though the HDMI cable. If that is the case, what is the point? I keep all that stuff on a hard drive anyway. So my question: can I hook up an HDMI cable and stream Jetflicks to my TV? The Jetflicks movies play on the native Internet browser. So I guess a more accurate question would be: Does the native internet browser display on the TV through the HDMI cable? If so, is there audio?
I want to stream a media file on web, and android developer website has said that MediaPlayer.setDataSource() can set the data source (file-path or http/rtsp URL) to use. But I got an error in both G1 device and emulator when streaming a rtsp url file: Command PLAYER_PREPARE completed with an error or info PVMFailure error(1, -1). Does anyone know what is this error, or anyone knows where can I get the error description (1, -1) means? the code can work successfully when playing a audio/video file or streaming a http protocal sudio/video file, but can't stream rtsp protocal file. Is android not supported rtsp streaming?
I am on the brink of buying my first Android phone (HTC Desire) and have just stumbled across some postings about Vodafone's restrictions on streaming audio and video. I have read a few articles on this now, but I really don't understand exactly what the blocks apply to. It seems to apply to "RTSP" streaming, but I'm not clear about what this is and whether this just applies to live streaming, or also to recorded video such as clips on the BBC News site.Does this apply only to Vodafone, or also to other networks? I was thinking of getting a new contract with Vodafone because the coverage with my current provider Orange in my area is very poor, and because Vodafone have the best price for the Desire. However if this streaming problem applies to all video on Vodafone, I may well pay a little more and go for O2. I don't know whether this would make any difference, but I will be going for an unbranded and unlocked phone through CPW.
I've read many topics from many android forums about this issue, but I have yet to see any posted solutions. If anyone know what program to convert videos, what settings to use, what resolution, what frequencies, what bitrates that will result in sync'd playback all the way through a movie, let me know.
Now i tried to play a video file .3gp from the internet youtube link via my android device ! not emulator. And it worked! But when i tried to play a video file .3gp or .mp4( i converted the mp4 from 3gp ) its not working. Im getting error "Cannot play video Sorry, this video cannot be played"
Tried to play with the link to my device wich is: I:DCIM100MEDIAVIDEO0030.3gp or .mp4 or tried this: I:/DCIM/100MEDIA/VIDEO0030.3gp with "" like: "I:/DCIM/100MEDIA/VIDEO0030.3gp ";
I tried to remove the I: but it didn't change anything.This is the main code im using to play the video files on my device. And there some lines i marked with // wich i used to play from the internet and im not using now.
Could you please let me know how to do the audio record in android emulator and play the same recorded audio. Could you please help me in proceeding in development.I tried with MediaPlayer API's also.Its not working.
I am trying to develop an android application which will be able to record a video and send it in real time to a remote computer. The computer then will play it live. I have managed to establish a connection between the android and the computer, and to record a video on the android phone. My problem is, how will I be able to send this video to the computer? Does anyone have an idea?
I connected my phone via USB to my RR sport to upload audio files to the RR hard drive. However, I keep getting connection error. Apparently my device is not being recognized as a storage. I did the unmount thinking. Also, I am wondering how I can play audio from my phone through the RR audio system. RR is 2010.
I'd like to be able to keep tabs on my dogs while I'm away. Ideally, I picture an app running on my PC that can stream audio to my cell, when I run an app on my cell. Any ideas around this? And there can be monkeys in the middle (PC to webpage to phone).
For my app, I need to record audio from MIC on an Android phone, and send it over TCP to the other android phone, where it needs to be played. I am using AudioRecord and AudioTrack class. This works great with a file - write audio to the file using DataOutputStream, and read from it using DataInputStream. However, if I obtain the same stream from a socket instead of a File, and try writing to it, I get an exception. I am at a loss to understand what could possibly be going wrong.
I have developed a streaming app for Android. My app is streaming AAC audio.The app works fine on 1.5, 1.6, and 2.1. I tested it on Froyo, the android 2.2 and it does not work there.I have try everthing - Streaming directly from server url - Download incrementaly the content, store in a temp file, start reading. - Download incrementaly the content, store in a ContentProvider and start reading.No luck, I always got errors : 06-23 02:47:02.257: ERROR/MediaPlayer(266): Unable to to create media player 06-23 02:47:02.257: INFO/System.out(266): setDataSourceFD failed.: status=0x80000000 Does Froyo support AAC? Anyone happens to stream AAC?
I am working on a music streaming application and cannot find APIs I can use for it (MediaPlayer supports files only local or HTTP). I have a few questions and would appreciate if someone can help me answering them.
Does Android SDK 1.5 support audio streaming (MP3 stream for example)? If does what API calls I can use? If doesn't what are the plans for future SDK releases that would support streaming? Are there any suggested work arounds until the streaming is supported (like temp files etc)?
I have installed the 2.1 update on my HTC Hero yesterday, and happy as a puppy I instaleld my streaming app... Just to discover that it won't stream decently... It has to stream an mp3 stream and it works on my ADP 1.6 device, on a Samsung Galaxy S test-device (with 2.1) and on the 2.1 emulator...
Logcat is gettign the follwoing errors :
06-30 10:14:18.360: INFO/AudioPolicyService(56): stopOutput(): before lock 06-30 10:14:21.259: DEBUG/dalvikvm(1906): GC freed 28 objects / 1944 bytes in 186ms 06-30 10:14:21.439: INFO/AudioHardwareMSM72XX(56): AudioHardware pcm playback is going to standby. 06-30 10:14:21.439: DEBUG/StreamProcess(56): OutputStreamProcess::ResetAOLC() 06-30 10:14:22.919: INFO/AudioPolicyService(56): startOutput(): before lock 06-30 10:14:22.987: INFO/AudioHardwareMSM72XX(56): AUDIO_START: start kernel pcm_out driver. 06-30 10:14:26.329: DEBUG/dalvikvm(287): GC freed 11 objects / 488 bytes in 268ms 06-30 10:14:47.789: INFO/AudioPolicyService(56): stopOutput(): before lock 06-30 10:14:50.879: INFO/AudioHardwareMSM72XX(56): AudioHardware pcm playback is going to standby. 06-30 10:14:50.879: DEBUG/StreamProcess(56): OutputStreamProcess::ResetAOLC() 06-30 10:14:53.911: INFO/AudioPolicyService(56): startOutput(): before lock 06-30 10:14:54.022: INFO/AudioHardwareMSM72XX(56): AUDIO_START: start kernel pcm_out driver.
The result is a playing player, which stops (pauzes) every 10 seconds, the restarts This will work for about 8-10 restarts and then will simply stop without ever thrwoing an exception or something like that Which makes it somewhat difficult (if not impossible) to catch for us developers So has anyone been experiencing the same problem ? Does anyone have a clue if in this update another OpenCore lib has been used ? Does anyone have any solution ?
When I pair my sprint htc hero to my car stereo, it always opens the default music player and starts playing a song. Can I prevent this from happening? I generally want to listen to pandora - and not my mp3's. Most annoying part is that the mp3's play simultaneously with pandora until I manually open that music app and pause the music.
There must be a way to either turn off the default action or to plug another program name into that defautl.
I'm planning on building an application where audio media is going to be streamed to the mobile phone for the user to listen.The targets are smartphones: iPhone/Blackberry/Android/(J2ME ?).I see that streaming on iPhone has to be done with HTTP Live streaming, but I don't see it supported by other platforms.Should I broadcast the streams via rstp ? http ? Is there any way to use a unified solution for all the different mobile platform ? If anyone already had to go through this, help would be greatly appreciated.
I'm working on an app for a client that requires streaming of an AAC audio stream. Unfortunately, there's nothing I can do about the stream format on the server. I'm working on Android and have discovered that Android's media player does not support raw AAC streams (which is what I'm getting). I found a project on Google Code that supports it (I tested it with the stream) but it's GPL'ed and that doesn't work for my client. I don't have much experience with this sort of thing so forgive me if my ideas aren't great. I know Android can play AAC encoded content if it is in an MP4 wrapper so I had thought about creating an MP4 wrapper on the fly on the client-side or perhaps even just doing some conversion to another format on the fly. Are these reasonable options? Does anybody have better suggestions?
Edit To rephrase, is it possible to put a raw AAC stream from a web server in an MP4 container in real time? If so, does anybody know of resources to help me with the process?
I am trying to playback voice data which is received in form of RTP packets in my application. The class Receiver.java extends thread is playing back the audio. The question is 1. For real time streaming operation how to configure the write(byte, int offsetInBytes, int sizeInBytes) methods? 2. What is the function and purpose of getPlaybackHeadPosition() What is playback head position kindly explain! Note: The RTP packets I am receiving are not from a traditional Datagram Socket but rather then from Ad Hoc Network so there is unknow delay for receiving the RTP packets. I need to design the Receiver class to play back the voice as soon as packets are recieved. Kindly help me in this regard You received this message because you are subscribed to the Google Groups "Android Developers" group. To post to this group, send email to firstname.lastname@example.org To unsubscribe from this group, send email to android-developers+ unsubscribe@ google groups.com For more options, visit this group at http://groups.google.com/group/android-developers?hl=en
I want to stream audio directly from my mac via bluetooth to my nexus one. This might sound crazy, but I actually want to be able to control the audio from my mac. I just want to be able to connect an audio jack to my nexus one and let it stream the audio via bluetooth. Is there an app for that?