Embed config is not supported in RemoteEmbeddedPlayer - android-youtube-api

When I try to play a youtube video on android using the youtube official api I get this error in logcat
E/YouTubeAndroidPlayerAPI: Embed config is not supported in RemoteEmbeddedPlayer.
and then the onError callback is triggered with NOT_PLAYABLE reason.
Is it something configured incorrectly for the video?
This is the video:
https://www.youtube.com/watch?v=sw6Mg81YMg0

had same problem when I created my project for Android TV as the target device, then I create a new project for phone and tablet. Now it works
hope this helps.

Related

Flutter HLS (m3u8) format video from asset not playing on the iOS device

I'm facing a problem with the video player. HLS (m3u8) video from the asset is not playing on the iOS device. I've downloaded all the ts chunk files with playlist.m3u8 and copied them to the asset folder. Android there is no issue it's playing fine from the asset but on the iOS device it's not playing and the error log also not throwing.
Any solution? How to play HLS (m3u8) video from the asset/local on the iOS device.
Using: video_player: ^2.3.0
Thanks in advance!
If you need to access videos using HTTP (rather than HTTPS) URLs, you will need to add the appropriate NSAppTransportSecurity permissions to your app's Info.plist file, located in /ios/Runner/Info.plist.
See Apple's documentation to determine the right combination of entries for your use case and supported iOS versions.

How to open an app with the iOS11 integrated qrcode reader

I want to use the new in-build qrcode reader in the iOS camera to open an app.
I cannot find the right documentation in the API.
Thanks for your help
Best Regards
Christian
I'v just tested with various URLs. It seems if the URL is http the Camera app opens Mobile Safari. If you know the target app URL (which is declared in the Info.plist) you can open the app with QR codes.
(ex) youtube://user/Apple --> Opens Youtube app

Google Cast plugin for Unity can't stream Video clip

I'm trying to play a Video inside Unity, and stream it to Google Cast.
Google provides a plugin that enables the connection to a Cast device, and it works fine once a correct Cast App ID is given.
Recently Unity provided a component 'VideoPlayer' that enables video playback inside a mobile device. And I tried to use both of them to stream video content on the Cast device. But when I play the video, the app stops responding with a signal 'SIGABRT' at
reinterpret_cast<PInvokeFunc>(_native_GCKUnityRenderRemoteDisplay)();
I also tried to play the video using AVPro plugin but the same issue appeared.
The plugin works just fine without a video, and the last update of the plugin is Apr 2016 so I think the plugin has some issue with the Unity's latest VideoPlayer component.
Is there something I can do about it?
There are currently no plans to update the Google Cast Unity plugin.

Live stream implementation using RTSP protocol

I'm trying to gain access to a live stream through the RTSP protocol on iOS. I'm trying to run the example from this website: http://www.gdcl.co.uk/2013/02/20/iOS-Video-Encoding.html and it's advertised that you can just take the url (rtsp://) and paste it into quicktime player, VLC or some other means, but whenever I try it fails. When I try in quicktime player it gives me this error: The document “Macintosh HD” could not be opened. The file may be damaged or may not be a movie file that is compatible with QuickTime Player.
What am I doing wrong? Is the example broken or do I need to update some specs in the code. I'm running iOS 9.3 and it's told to work > 7.0.
I was able to play this back on VLC when compiling and running on my iOS device. You need to ensure that you are on WiFi (vs LTE or 3G). I'm on iOS 9.2.1 and played back with VLC version 2.2.2.
You can then take it a step further as I was successful in ingesting it into Wowza via Stream file with the following configuration:
{
uri : "rtsp://[rtsp-address-as-published-on-the-app]",
streamTimeout:12000,
reconnectWaitTime:12000,
rtpTransportMode:"udp",
rtspValidationFrequency:15000,
rtspFilterUnknownTracks:true,
rtspStreamAudioTrack:false,
rtspStreamVideoTrack:true,
rtpDebugSession:true,
rtspSessionTimeout:12000,
rtspConnectionTimeout:12000
}
I would suggest reviewing what the console logs say in your iOS application (xcode) and then also take a look at your VLC error messages/logs as well to see what the exact issue is when you try to playback.

get shared/server url using CocoaHTTPServer

I am creating application that play video using chromecast device on TV (Apple TV or monitor that support HDMI port). Application is playing url like "https://serveraddress/video.mp4" but its not playing the video stored in app bundle(local video).
I have found that to make video play in remote machine (here it is TV) have to create local server on iphone.
I found this SDK CocoaHTTPServer I have run the sample application but not getting how to take video URL that is stored in app bundle.
Can someone please help me in this.
Thanks in advance.
This work great for me.
https://github.com/swisspol/GCDWebServer
Initialize as given in sample code (give document directory as root).
Than to access local file copy that file in document folder and than access using http://localhost:8080/download?path=new.jpeg

Resources