Android :: Live Video Streaming - How Will I Be Able To Send This Video To Computer?
Sep 21, 2010
I am trying to develop an android application which will be able to record a video and send it in real time to a remote computer. The computer then will play it live. I have managed to establish a connection between the android and the computer, and to record a video on the android phone. My problem is, how will I be able to send this video to the computer? Does anyone have an idea?
I need help on live video streaming. I want to do the following things in live video streaming. 1. I want to broadcast the live video which is recording by camera from one android mobile directly to a particular ip address. 2. now i want to play that video to another android mobile from that ip address without storing that video.
The problem I have is in the implementation of a video streaming application. What I need to do is to stream live what I record with the androids' camera, to a remote computer. I managed to establish a connection between the mobile phone and the computer, but I do not know how to send the video.I searched through forums, and I found out that in order to send a streaming video you have you use the ParcelFileDescriptor. So I created an instance of MediaRecorder with encoding H263 and .mp4 file format, I created a Socket and then I created a new ParcelFileDescriptor connecting it with the socket. If that is correct then the android application that I created is functional.What I don't know now is what to do at the side of the computer in order to receive and view the streaming video. I tried using JMF, but it doesn't support .mp4 video files. What do you suggest me to do? Or where to look in order to find a solution?
Where to get streaming (live) video and audio from camera example for Android? Suppose I want to create some live video streaming service app so I'll have some cool server at the back end. And I know how to do that part. Suppose I have some stand alone app for PCs now I want to go on to mobile devices. So I want to see some sample app grabing audio and video streams from Phone, Synchronizing them, encoding somehow, and sending LIVE stream to server. I need any Open-Source sample that will do this or something like this. Where can I get such one?
Has anyone tried streaming live video from there Hero to ustream.tv? Android supports it now, so it should be able to be done. I used to be a lifecaster but since I had no way of streaming video on the go, this would be a dream come true if I could do so on the Hero since it has Android.
Kindly provide me the code/steps for Android application which will provide the Video Streaming & Video playing functionalities by accesssing the Videos from various video websites available on Internet.
For a research project we need to capture video on the HTC magic and transmit this video over a network to other devices, both HTC magic's and desktop pc's (which have their own implementations of the codecs).For the transmission over the network we are using some of our own native libraries, which have all compiled fine and work. These libraries work via the RTP (Real time transfer) protocol. This poses a significant problem because RTP expects detailed information about the contents of each packet (exact timestamp, sampling rate, which part of which frame is in each packet etc.) So we tried the method described at http://www.mattakis.com /blog/kisg/20090708/broadcasting-video-with-an This appears the only way to get the MediaRecorder to send it's data to our native application. So we run a server-socket on the native side, send MediaRecorder output to a FileDescriptor of a javasocket. The native side does receive data. 2 problems however :
1. Like said in the blogpost : header information etc. is not filled in, so if we write the data to file from the native-side, it's not playable.
2. The socket we write to is TCP, so a stream implementation. So there is no clear way to know where each frame begins and ends (because the data comes in continuously). And that's the main problem : we can't package the data in good RTP packets for transmission! Nor can we get good information about the timing etc. So we looked into OpenCore a little bit, hoping to be able to do it through JNI-interfaces ourselves (or maybe even directly with the native implementations from our native code). This has proven a daunting task, because so little documentation can be found on these things. So the question is : Is it possible to get the encoded video (and audio) data with clear frame-separation and good timing-information for RTP transmission? (It is not an option to use anything but RTP, as the research is partly about RTP possibilities)
I tried for video streaming in sdk 1.5. The video format .mp4.But it gives error like "This video is not valid for streaming to this video". But in http://developer.android.com/sdk/android-1.5-highlights.html .mp4 is supported for streaming.
Is their either a trick or an application that would let a video that is being streamed to loads in the background and allow you to move to another application or back to the browser while the video loads? As soon as I get off screen from the video player, all is lost.
I have a simple question "*Is Video Streaming possible in Android Emulator*",I am using the example in Android SDK but it give me message "*Sorry Can't play this Video*".I am using Youtube Video to stream in android Emulator.
I 'm developing a video streaming android application on HTC Tattoo. I 've Lighttpd server at my server side which i use to stream videos to a web site. On Android , Do I need to enable any module in Lighttpd server?
How to get videos from android to the wii (hacked of course). I have tried to just plug in a usb cord in the back like I do for my psp but every time I go into an app for watching movies it wont show my phone. I know I can just take the sd card out but its a real pain to take the back off my phone and the case I have on it too.
I've just recently released a first Beta version of AndroidVideo.mobi app, wich allows you to play tv shows and episodes on your phone. I'd be glad if anyone had time to test it and let me know of any quirks and bugs in it.
I am new in Android. I am using android os 2.2. I am trying to play streaming audio and video from url. I don't know what to do for that. after 3 days og googling i come to know that I have to use mediaPlayer and MediaController classes.I had use different differend combination of codes found on internet but not succeed.Can anybody guide me what to do for streaming audio/video playing. What about the UI if we use MediaPlayer or Controller class. is there any thing to related these in xml file of layout.if not then even please tell me what may be the code for that.please help me.
I have created an application that displays XML data information on the top section of my Linear layout. Below that I have a video that is streaming and takes a while to start playing. How would I set a dialogue to display until the video appears?
I've seen plenty of info about how to stream video from the server to an android device, but not much about the other way, ala Qik. Could someone point me in the right direction here, or give me some advice on how to approach this?
My company makes DVRs that specialize in streaming live and recorded video. The video is encoded using standard MPEG-4 codecs so the codecs in Android should have no trouble with them. However, the video is recorded using non-standard file formats and is streamed using our proprietary protocol (among other things we embed additional data such as watermarks with each frame of video). Is there any way I can take my stream pull out the frames and have it play on an Android device?
I want to live stream video from android device. I found out that ff server (ffmpeg) can be compiled and run on android. Now i need to get video from camera in android device and get it to the fserver in real-time.
I'm currently looking for a new cellphone and i'm in favor with the Xperia X10, but first i want to be sure that it's possible what i want to do. I want to watch movies from my laptop without connecting the laptop to a tv. I imagine it like this: Variation a) The phone accesses my files via WLan and plays the movie via tv out ... without converting the movies first Variation b) My laptop plays the movie and streams it via WLan to the phone which relays it to the tv. Is something of this possible ? and another (not important) question: Some phones support the HID Bluetooth profile. Does that mean i can connect my BT mouse to the phone ?
I have been having some issues with the Qik app. It has worked then stopped. I uninstalled it, then re-installed it. Sometimes it just loads the camcorder app, not the qik app, which is really odd. I figure it is built using the camcorder api but still very odd.
The customer service from qik has been poor as well.