How do you record RTSP Stream using MobileVLCKit IOS? - ios

So I currently have a RTSP Stream that is ran with the MobileVLCKit library's. I am able to receive the video stream with no problem, what I need to do now is record that RTSP stream onto the iPhone its being streamed to. Essentially I need to save the incoming RTSP stream to the phone while also letting the user view it live.

Related

How to broadcast wowza video from one iPad to multiple iPad devices - swift

I want to broadcast existing videos to multiple users through wowza...
Suppose I want to broadcast any 1 uploaded video (in wowza server) to my multiple users? so how can I do that.. can wowza call any API to start streaming in other users devices? Means when I started streaming video from my application then it should start in other devices also through wowza API.
Are you talking about broadcasting (streaming) an MP4 file as a simulated Live stream (playout) or as Video On Demand (VOD)?
Obviously you cannot force devices to start playing a stream. That'd only work if you develop an App that can listen for commands and trigger playback accordingly. Wowza doesn't have such an App, nor any built-in features that can do this.
If you want devices to access a stream on-demand you can simply upload the file to Wowza's content folder. If you want to have a programmed playout, like a TV channel, then you can check out this article: https://www.wowza.com/docs/how-to-schedule-streaming-with-wowza-streaming-engine-streampublisher
(the source code of the plug-in that is used in that article is available from https://github.com/WowzaMediaSystems/wse-plugin-streampublisher)
As from your Question, You can broadcast stream successfully and you might have used Wowza Go Coder SDK for doing it.
On Broadcasting live stream, broadcasted videos stored in CONTENT directory which is inside of Stream Engine installation directory structure.
You can find all the streamed videos in it.
Now, You want to broadcast particular stored video then you can do it by loading specific URL for that video. Broadcasting for that Video is not possible but, you can play that particular video as below which will be accessible to all your application users :
In IOS, Link or URL for Videos stored in CONTENT directory is as below :
http://[Host Address]:[PORT]/vod/mp4:sample.mp4/playlist.m3u8
rtsp://[Host Address]:[PORT]/vod/sample.mp4 (In Android)
Here, sample is the name of your Stream. You have to broadcast live stream video with different stream names so that all videos are accessible.
In this way, you can play stored live stream videos.

Detect when switched to an audio only HLS Stream

In our app, we have an HLS index file which contains multiple HLS streams for different Bandwidths. Our Lowest Badwidth stream is only audio.
We want to show a message to our users when the app has degraded to the Audio Only stream.
Is there a way to detect when we have switched to an audio only stream?
Thanks!

What is the role of Streaming server like Wowza?

I have been exploring on how to live stream from iPhone. I will have to publish a stream at URL to Wowza server that I came to know. Other thing is that I will require a library for iOS to encode and compress the camera output and will have to send that stream over RTMP protocol to the Wowza server. At the receiving and, there should be a player which can decode, decompress that stream comes from the Wowza to the device like iPhone (a user who wants to see live stream).
My question is, if encoding is done through particular iOS SDK, RTMP has a role as a Protocol, a player at receiving end has a role of decoding, then what is the role of Wowza ? What is its function that makes it very important in the live streaming process ?
I have been searching on the function of a Media Streaming server since 3 days, but I could not understand the exact function of Media Streaming server like a Wowza.
I am desperate to the answer..
Any explanation will be appriciated, thanks in Advance !!!
I actually did a year-long project involving media streaming on iOS and I used Wowza. The role of Wowza is to function as a media server that can receive the video that is broadcast from an iOS device over the RTMP protocol. With Wowza, you have options to send http parameters that instruct the server to begin or stop recording the live video that is being streamed. You also have the option of embedding video players in websites for live view.

CMSampleBufferRef to Audio Video for RTMP streaming

I am using iOS libRTMP to connect to an RTMP server. Now my problem is, how can I send the audio and video to an RTMP server? I've seen that CMSampleBufferRef is used for processing and sent to the server. The data is being processed with some flash video like metadata.
I tried VideoCore but it fails with the server I am using(Evostream) but works with other servers. Medialibs works but I need an open source option.
How do we convert CMSampleBufferRef to a format that RTMP understands?

How to play a non-rtmp stream

I have an issue to stream a video conference for many mobile device(Andriod, iPhone etc) and I have troubles with streaming.
RTMP stream uses a flash and didn't work with this devices.
What stream can I use? and how can I convert rtmp to this supported stream?
For mobile devices you would need to convert your video to HTTP Live Streaming (HLS). If it's truly a live stream, you could use something Zencoder live streaming to convert it. If the RTMP stream is just reading from an MP4, you could use Handbrake or Zencoder to transmux it to HLS.

Resources