What type of markers are to be used in microsoft touchless SDK I tried few thigs like cap of bottle. but I am not able to work with any demo give in touchless.. So suggest me what to do..??
You need good light and high contrast between the colors of the marker and your background on the touchless SDK. The SDK was rather rudimentary more proof of concept than actual product. Now that the Kinect SDK is released, you might want to consider using that instead.
Related
Does anyone work on object detection/tracking on Hololens? I'm expecting to automatically detect a physical object using a marker or without a marker.
As per my knowledge we have a library like OpenCV based on raw images. But, HoloLens is a powerful computer vision device. Is there a possibility Microsoft might expose some high-level object detection API?
With the actual version of Hololens it not could be done native. Rumors says that the next version will come with a IA chipset for this porpoise.
In the meanwhile, you can use thinks like Vuforia Object Detection which works smoothly.
Anyway, there is a version of the OpenCV in the Unity Asset store which come with severals samples of what you want to do.
I just recently got started with he Google Cardboard SDK for iOS and I'm looking to create an simple app in Swift that displays a 3D (Stereoscopic) VR video.
First, I adapted the VideoWidgetDemo Sample in the SDK (https://github.com/googlevr/gvr-ios-sdk/tree/master/Samples/VideoWidgetDemo) from its original Objective C to Swift 4 and it performs well. It uses GVRKit to create a GVRSceneRenderer with a GVRVideoRenderer.
But then I came across a blog post on the Ray Wenderlich site (https://www.raywenderlich.com/136692/introduction-google-cardboard-ios) that uses GVRSDK's GVRVideoView instead, which feels simpler and easier to use. However, there is a very noticeable performance difference. The video displayed by this app stutters/jitters much more than the GVRKit version.
I'm puzzled by the fact that the official Google VR reference documentation site (https://developers.google.com/vr/ios/reference/) doesn't even mention GVRKit, even though all the official samples from the GitHub repo use it instead of GVRSDK. So the samples follow one approach and the reference docs cover a different one.
I haven't been able to find any guidance for when to use one or the other (or even both together if it makes sense), so I'm hoping that someone on StackOverflow can shed some light on this choice.
I'm also curios about the performance difference I'm experiencing with the two different approaches. It would be great if there is a way to achieve the same level of performance with the GVRVideoView than with the GVRVideoRenderer.
Thanks in advance for your insights and suggestions.
Seem that the SDK is deprecated.
I've posted an issue about a GVRSDK on GitHub, and they say that the SDK is deprecated and developer have to switch to GVRKIT.
Here the GitHub issue:
https://github.com/googlevr/gvr-ios-sdk/issues/298
If your goal is to display 360 video in a simple app, check this Cordova plugin:
https://codecanyon.net/item/cordova-ionic-vr-plugin-photo-360-video-360-player-with-cardboard/20392357
It seems that Google came up with a new SDK because Daydream is now deprecated.
The Google Cardboard SDK offers a streamlined API, improved device
compatibility, and built-in viewer profile QR code scanning.
Quickstart : https://developers.google.com/cardboard/develop/ios/quickstart
Github : https://github.com/googlevr/cardboard
Is it possible / a good choice to use Google Cardboard SDK to realise AR?I Only have found VR related things with that SDK. What is the best framework for AR? Is Vuforia a good way to go? Im trying to write an AR app (for Android) which detects/scans room numbers at my university and shows the schedule of this room (which class / time / which prof...)
Thanks for help!
Google Cardboard SDK is made for VR, hence not the best option for AR. There are SDKs built specifically for AR. Check out this comparison
There is an alpha version of DroidScript (JavaScript IDE for Android) available which supports augmented reality for Google Cardboard. There is also a sample on the DroidScript forum that demonstrates Aruco marker detection (Augmented reality). So I guess you could hack something together quite easily if you ask the developers for the latest version.
The latest version of vuforia can be integrated with Google cardboard.
Check this link:
https://developer.vuforia.com/library/articles/Solution/Integrating-Cardboard-SDK-050
Now I have an requirement in Augmented Reality, I suppose to detect the live object(pen/marker) then I have play some interactive content.
I need a suggestion/advise from you is to recommend the appropriate SDK to develop this app.
I used Vuforia for normal/simple AR app but this is involved the real time object detection.
Friends, kindly suggest me the SDK to meet the requirement.
-Murali Krishnan
You can do this with a cylinder object. you would need to create a different object, for each pen style you would want as a trackable.
https://developer.vuforia.com/resources/dev-guide/creating-cylinder-target
As mentioned by ashatte, you should have a look at metaios sdk.
You are looking for 3d Markerless tracking. The sdk is very good for this.
I am using an XBox Kinect with the Kinect for Windows SDK. I want to make an application that will augment a 3D mask (a 3D model of a mask made in 3DS Max) onto the face of anyone using the application. The application will be used in an exhibit locally. I have not tried much because I don't know where to start. So what I want to know is, is it currently possible to augment a 3DS Max model onto a live video stream using the facial recognition and skeletal tracking features in the newest Kinect for Windows SDK, and if so, how/where should I start trying to do/implement this? Any point in the right direction would be great. Thank you! PS And yes, I have read the UI guidelines and the facial documentation. My problem is one of not knowing where to start programming, not one of not understanding the fundamental concepts. Thanks!
If you are serious about getting into developing for the Kinect I would recommend getting this book:
http://www.amazon.com/Programming-Kinect-Windows-Software-Development/dp/0735666814
This goes through developing with the Kinect for Windows SDK from the ground up. There is a face tracking and an augmented reality example so I'm pretty sure you will be able to achieve your goal quite easily.
All the code from the book is here:
http://kinecttoolbox.codeplex.com/
Alternatively, there is an example here which pretty much is what you want to achieve:
http://www.codeproject.com/Articles/213034/Kinect-Getting-Started-Become-The-Incredible-Hulk
It is developed using the Beta version of the SDK, but the same priciples apply.
You can also check out the quick start videos here:
http://channel9.msdn.com/Series/KinectQuickstart
In summary, based on my own experience, I would spend some time going through the beginner examples either in the vides or the book (I found the book very good) just to get familiar with how to setup a simple Kinect project and how the different parts of the SDK work.
When you have developed some throwaway apps with the Kinect, I would then try tackling your project (although, the Incredible Hulk project above should get you most the way there!)
Best of luck with your project