After the release of Apple ARKit, there will be a lot of augmented reality apps will be released and I'm very new to this concept and wanna ask some questions about it.
1-) Can you implement drag and drop for ARScene? What I mean is lets say I have camera feed(ARScene) on the background and have another view on top of that having some images. Can I drag any image from that overlay and place it in my ARScene anywhere I want?
2-) If ARKit doesn't have this capability are there any other frameworks that can support this?
Thanks all.
Related
Weird crazy question I know. My current setup is a SCNScene with a camera controlled by the device's gyroscope. I'm able to add and light normal nodes, however I would like to add 2D UIView objects into the scene like UITextViews or maybe some buttons. The views would need to be inside the scene and thus become no longer visible if the camera moves away from them.
Firstly, is this even possible? Or would this be way more difficult to implement than rebuilding an editable textview as a node? Could this be achieved by categories or...?
I just talked to the scenekit people here at WWDC. He said that as of now it is completly impossible to do it in a nice, functional way. The only options he offered as a possible solution is to create the UIView element somewhere off screen or behind the scene and continuously take screenshots of the object and apply those images as a texture to a SCNNode. He also pointed out the performance will be very poor with this because taking screenshots is heavy and you can't get going very quick with it. So I guess this is a no-go until UIView adds support because, according to the engineer, it's impossible to implement this because of a UIView limitation and not a SceneKit one.
In my app, i want users to click a photo, then i want to allow them to edit it for crop/rotate/straighten. Can i use built-in photos app for that? or is there any open-source/api to have similar view controller.
I know it's a possible duplicate of Is there a view controller for image crop and rotate works like iOS 8 photo.app? but i found no answer here as well.Please let me know if it's possible?
Unfortunately, no.
UIImagePickerViewController allowEditing property, when set to YES, provides you a very simple cropping UI, but it doesn't support advanced rotation as iOS Photo app.
So you'll have to make it yourself.
The github projects in the question you linked might be a good starting point for that. (here's another one)
Good luck !
I wrote one in Swift, you can continue to develop based on this: QCropper
I would like to animate using Adobe Edge (html5). I would then like to proceed to import my animation in Xcode for an iPad app (IOS 8). I will be using Objective-C for this project. Is there any way to do so?
If not what are my alternatives? Keep in mind I want to animate using some sort of GUI application.
I think it depends on what sort of animations you want to create and how they will be used in your app. One possible way is to integrate the SpriteKit library into your UI, which can be useful for making animations or interactive artwork in your UI.
If you go this route, you can create an animation as a movie clip symbol in Adobe Animate. (Adobe Edge Animate has been discontinued and effectively replaced by what is currently called Adobe Animate). Then right-click your symbol in Animate and select Export PNG Sequence. This will produce the frames of your animation as PNG images, which you can then animate in your app using SpriteKit. Animate can also export art assets as a sprite sheet or texture atlas.
You would need to get familiar with the SpriteKit library to go this route. As other users pointed out, there are alternate approaches you could take.
The only way I know this would work would be to use a webview in your ObjectiveC code. Then you would point your web view to the local HTML/JS files generated by Edge Animate. Basically this is how Cordova does things.
You could user alternatives like Phonegap, or create your own UIWebView and charge the www folder.
Basically I want to know if it's possible to add an effect similar to when you open Newsstand with no magazines and you encounter an effect where you can kind of see the user wallpaper with a blur effect added onto it.
I sort of managed to get the wallpaper onto the app with a few .plist settings changes, but want to know if theres a better way that also adds the blur effect.
regards,
Patricio
You can use the UIImage category that Apple provides in its sample code, that allows you to blur a UIImage with a variety of styles and colours.
https://developer.apple.com/downloads/index.action?name=WWDC%202013
Search the page for UIImageEffects, download it. Open the project, and you'll find a UIImage category you can move to your project.
More on this topic can be found in the Engaging UIs in iOS7 WWDC 2013 talk.
i'm looking for an AR API (possibly free) or SDK that may be used on IPAD.
I've tried Wikitude so far, but it's not adapted for IPAD for the moment, and most of the well known API seems to be only made for iPhone..
Any clue on this one?
Any chance that this API offer the possibility to add the AR view as a subview (for example as an element of a tab bar..:))
Thanks a lot
Why are you going for 3rd party library? You can apply AR using local library. On your ViewController make 2 UIView. The one in the back implement Camera View on it. And the one on the front implement whatever you want to show on the live preview of camera. Then if you want to capture the screen just simply take a screenshot and it'll be saved.