It is documented all over the web that the iOS does not support multiple HTML5 audio. I ran a test and here are the observation:
iOS 5 (iPad1) - can play only one audio at a time.
iOS 6 (iPad2) - can play multiple audio.
iOS 7 (iPad3) - can play multiple audio.
It looks like on iOS 6 & 7, we can play multiple HTML5 audio, but to my surprise i could not find any discussion regarding this over the web, also apple documentation still says "Currently, all devices running iOS are limited to playback of a single audio or video stream at any time"
I am not sure if multiple HTML5 audio will work across all devices which support iOS6 and iOS7, also will apple continue to support this with next version of iOS, does anyone has any idea about this?
Related
I am having issues with downloading and playing an HLS URL at the same time.
I have followed WWDC'16 video and also below link.
But either I am able to play the audio or I am only able to download it. I am not able to achieve doing both the tasks at the same time.
And the download is also working only on iOS 11 beta releases, not on iOS 10 devices.
Am I missing something or there is some other way of achieving it?
There are a lot of changes related to HLS and FPS for iOS 11
And there are two examples from Apple for developers:
FairPlay Streaming Server SDK beta - only for XCode 9
FairPlay Streaming Server SDK - for XCode 8
As for me: on iOS 10 I can play and download HLS+FPS without any issues, but the same code for iOS 11 is not working (as I wrote - there a lot of new instances and workflows in it).
Audio in HMTL5 video works fine in Safari on iOS 10.3.1 on iPhone, though it doesn't in standalone web apps (same html code & video file). Video play is fine. Just no sound.
There are a few other related discussions in the past, eg, Why HTML5 video doesn't play in IOS 8 WebApp(webview)?. I tested on iOS 10 using the html provided by that post.
Not sure whether it is a new bug introduced in iOS 10, or a bug that has never been fixed by Apple since earlier versions. Does anyone experience this issue? Are there any workarounds? Thanks in advance.
I had the same question. I googled all over the place. Then I realized what was causing the problem: I had the physical "soft mute" switch (next to the volume buttons) on my iPad turned on. Infuriatingly, this muted the volume on web apps, but not Safari web pages.
I am developing an video player app on iOS and I am now thinking about how to support DLNA so that my app can mirror its online video to the TV with some DLNA-supported device.
Notice that the online video is playing on my app via WIFI or cellular network and I could switch it to be played on TV and my app becomes a remote control and a server to the TV.
Which framework should I use?
I already know Cyberlink and PlatinumKit.
I have worked on DLNA with iOS and Android device. I did not used the Cyberlink or PlatinumKit. I learned how it works and write with swift and java.
Here is my blog about the subject, if you only want the part of DLNA to find DLNA device and stream the video to the device, then control events like play, pause and seek. You could find all the material you need here.
https://eliyar.biz/DLNA_with_iOS_Android/
I have a scenario where I have to play videos using flash player in iOS device's browser. Offically Apple does not support flash, so is there any way of playing flash videos in iOS devices?
You could try a flash to html5 converter like Google swiffy:
https://www.google.com/doubleclick/studio/swiffy/
Swiffy has been regularly upgraded since 2011 and does a very good job, although expect much bigger file sizes.
Also, looped event sounds in your FLA files (if any) will play only once (i.e. they don't loop). What is more, event sounds play in Swiffy in PC browsers, but currently do not play on iOS devices. Stream sound isn't supported in Swiffy yet, on any device.
I have tried to live stream audio (AAC-LC) from iOS for 3 months without much success...
I tried Audio Queues, which work well but there is a strange delay (~4s) and I don't know why (high level API ?)
I tried Audio Units, it sometimes works on the simulator but never with the phone using a modified code from this source
I am really lost, can anyone help me ?
EDIT
I have to do a live streaming application (iPhone-> Wowza Server via RTSP). The video part works well with little delay (1s). Now I'm trying to add audio in addition to video but I'm stuck with the SDK.
tldr : I need to capture microphone input then send AAC frames over the network without getting huge delay
This app, which I just now completed, broadcasts audio between any two iOS devices on the same network:
https://drive.google.com/open?id=1tKgVl0X92SYvgpvbljRzilXNQ6iBcjqM
Compile it with the latest beta release of Xcode 9, and run it on two iOS 11 (beta) devices.
The app is simple; you launch it, and then start talking. Everything is automatic, from network connectivity to audio streaming.
Events generated by the app are displayed in an event log in the app:
Even though the code is simple and concise, the event log was provided to make understanding the app's architecture quicker and more easily.