I studied about the C++ API provided.Can somebody suggest me some methods or links to proceed
how to import those c files into Xcode and work with it?
Does Octave provide any online API?
Thanks in advance.
Related
I just recently got started with he Google Cardboard SDK for iOS and I'm looking to create an simple app in Swift that displays a 3D (Stereoscopic) VR video.
First, I adapted the VideoWidgetDemo Sample in the SDK (https://github.com/googlevr/gvr-ios-sdk/tree/master/Samples/VideoWidgetDemo) from its original Objective C to Swift 4 and it performs well. It uses GVRKit to create a GVRSceneRenderer with a GVRVideoRenderer.
But then I came across a blog post on the Ray Wenderlich site (https://www.raywenderlich.com/136692/introduction-google-cardboard-ios) that uses GVRSDK's GVRVideoView instead, which feels simpler and easier to use. However, there is a very noticeable performance difference. The video displayed by this app stutters/jitters much more than the GVRKit version.
I'm puzzled by the fact that the official Google VR reference documentation site (https://developers.google.com/vr/ios/reference/) doesn't even mention GVRKit, even though all the official samples from the GitHub repo use it instead of GVRSDK. So the samples follow one approach and the reference docs cover a different one.
I haven't been able to find any guidance for when to use one or the other (or even both together if it makes sense), so I'm hoping that someone on StackOverflow can shed some light on this choice.
I'm also curios about the performance difference I'm experiencing with the two different approaches. It would be great if there is a way to achieve the same level of performance with the GVRVideoView than with the GVRVideoRenderer.
Thanks in advance for your insights and suggestions.
Seem that the SDK is deprecated.
I've posted an issue about a GVRSDK on GitHub, and they say that the SDK is deprecated and developer have to switch to GVRKIT.
Here the GitHub issue:
https://github.com/googlevr/gvr-ios-sdk/issues/298
If your goal is to display 360 video in a simple app, check this Cordova plugin:
https://codecanyon.net/item/cordova-ionic-vr-plugin-photo-360-video-360-player-with-cardboard/20392357
It seems that Google came up with a new SDK because Daydream is now deprecated.
The Google Cardboard SDK offers a streamlined API, improved device
compatibility, and built-in viewer profile QR code scanning.
Quickstart : https://developers.google.com/cardboard/develop/ios/quickstart
Github : https://github.com/googlevr/cardboard
I'm trying to write an IOS app using Xamarin that sends data to PC using TCP. I want it to do this with Cable or WiFi. I have been researching around the net for sometime now. But couldn't find any good tutorials. I'am aware that I should be using NSNet and Bonjour Windows SDK but I couldn't find anything more than that. Any suggestions or directions on how to achieve this would be very helpful. Thanks in advance.
If you are using Xamarin then the more preferred choice would be to write things using System.Net.Sockets namespace. This way you can reuse the code with other C# clients, including Android (if needed).
Plenty of examples on TcpClient/TcpServer.
https://msdn.microsoft.com/en-us/library/vstudio/system.net.sockets.tcpclient(v=vs.110).aspx
i need some advice about developing app using QR code scanner's libraries on iOS.
I am already bit familiar with iOS7 AVFoundation framework, Zbar and Zxing.
My goal is to use method that can scan QR code in the shortest time possible.
I am thankful for any experience shared or tips how to improve the scanner performace.
Have you checked out CDZQRScanningViewController which uses iOS7 only API - is this what you are looking for?
Have you noticed a performance issue with the other libraries? maybe you could elaborate on the issue you are having.
I want to get stream of an ip-cam on my iPhone/iPad and want to display it on screen. By R&D i found that ffmpeg is the only way to achieve it but i found nothing on ffmpeg. Is there any other way to achieve it or a confirmed way to get compiled ffmpeg on mac please mention that. Material regarding how to use ffmepg or source code example will be highly appreciated.
Is there nothing built-in framework to achieve it if not then please mention if there is any free framework/sdk to achieve this functionality.
Thanks
There are actually a few.
here are some links
http://www.streammore.tv/
http://www.live555.com/
I am sure if you google you can find more.
I cannot only address the first one, because that is ours, but I didn't want this to sound purely like self promotion.
I'm a complete newbie at Appcelerator's Titanium SDK, so I apologize in advance if this question has an obvious answer.
Is there any way to change a sound's pitch in real-time using appcelerator? Is there a module that can do that?
When looking for the answer, I found out about some objective-c libraries (dirac-2, for example) that can do that. Is there any way to interface to such libraries through Titanium without a dedicated module?
After some searching, I found a module - Zero Latency Sound - that can control a sample pitch in real time.
https://marketplace.appcelerator.com/apps/891?1475082272
Also, this module's support is commendable.