I need to send audio chunks like a stream to server while recording audio in iPhone. What is the best method to implement this? I looked at HLS. But it supports server to client streaming. I need client to server streaming. Please suggest a best method to implement audio streaming from client (iOS device) to server.
HLS is not supported from client to server. Only from server to Client.
"Streaming audio or video to iPhone, iPod touch, iPad, or Apple TV"
https://developer.apple.com/library/content/documentation/NetworkingInternet/Conceptual/StreamingMediaGuide/Introduction/Introduction.html
Probably not the answer you were hoping for.
Related
I have been exploring on how to live stream from iPhone. I will have to publish a stream at URL to Wowza server that I came to know. Other thing is that I will require a library for iOS to encode and compress the camera output and will have to send that stream over RTMP protocol to the Wowza server. At the receiving and, there should be a player which can decode, decompress that stream comes from the Wowza to the device like iPhone (a user who wants to see live stream).
My question is, if encoding is done through particular iOS SDK, RTMP has a role as a Protocol, a player at receiving end has a role of decoding, then what is the role of Wowza ? What is its function that makes it very important in the live streaming process ?
I have been searching on the function of a Media Streaming server since 3 days, but I could not understand the exact function of Media Streaming server like a Wowza.
I am desperate to the answer..
Any explanation will be appriciated, thanks in Advance !!!
I actually did a year-long project involving media streaming on iOS and I used Wowza. The role of Wowza is to function as a media server that can receive the video that is broadcast from an iOS device over the RTMP protocol. With Wowza, you have options to send http parameters that instruct the server to begin or stop recording the live video that is being streamed. You also have the option of embedding video players in websites for live view.
I have read about HTTP Live Streaming from Apple. So far I understand that it was created for streaming video to iOS devices. But is it possible to use this approach to stream from iOS device while recording on camera? If so can you give me a clue or tell how to do it?
It is possible, yes. But I would not recommend it. HLS is a pull based protocol. Good for delivering to clients, but bad for ingesting. Would need to run a web server on the device, and package the media into a transport stream. Its a lot of effort that is easier to handle on the server.
I'm developing a website who needs an external RTMP stream.
I'm using jwplayer to run the stream using Flash (examples and information about here).
My problem is the stream do not works at iOS.
Somebody suggests a solution?
IOS does not support rtmp protocol for that you has to use http protocols, i.e. ,http live streaming
I have to record the high volume sound. It can be upto 60 seconds. The application is already made on flash and the current flash application uses rtmp (red5 server) and streams the recording to the server in real time as FLV. I have uploaded the audio file on server as mp3 from iOS application but this approach is not useful as the volume is very high as mentioned above. So I want to use rtmp (red5 server) for iOS application. I want to ask is it possible that I can record the audio and upload on this server and also can steam from there?
Edit-1: I want to ask that whether Apple can reject the application if we use the RTMP iOS library in our application?
Edit-2: I did some research and found some 3rd party libraries like:
http://www.aftek.com/afteklab/aftek-iphone-RTMP-library.shtml
I want to know, whether it is possible to directly connect from iOS to RTMP server?
Or these libraries use the middle layer approach with HTTP to connect to RTMP?
Example: iOS -> HTTP Server -> RTMP Server (red5)
I will appreciate the help.
I'm trying to find out how i can use wowza media server(which i just installed) to receive a live audio stream from an iOS device(Xcode simulator) and play it on a web browser(safari) using http live streaming. i just need direction, guidance or a tutorial to start with this, or just the basic concept of how it works.
sorry for the newbie question, but i really really did try digging up the documentation, their is nothing about iOS http live streaming specifically, they concentrate more on Flash streaming(flv).
Thanks.
Use Wowza's MPEG-TS capabilities
There are a couple of Google hits and topics on the Wowza forums. This one looks fine for getting you started: Using an MPEG Transport Stream (MPEG-TS) encoder with Wowza Pro (MPEG-TS).
You can then just click a simple link in Safari on iOS devices and it should work fine. The like will look something like:
http://myserver.com/appName/myStream/playlist.m3u8
Wowza recently released an addon called "Gocoder". By using it you can encode your live streaming from wowza supported ios devices and broadcast it to any screen.
You can download the encoder from here
https://itunes.apple.com/us/app/wowza-gocoder/id640338185?ls=1&mt=8
Here is the step to configure the Gocoder addon with live application
http://www.wowza.com/forums/content.php?500