How to duplicating perspective zoom home screen feature in iOS? - ios

As we all know, iOS 8 having a perspective zoom feature in the home screen which zooms depend on the angle of a device and how it moves.
I want to duplicate that feature inside my app. I found a question to find out the angle of a device here.
How can I determine the angle of the device?
Measuring tilt angle with CMMotionManager
I need to discuss the ideas to zoom the image depend on the angle of a device. Any help would be appreciated.
Edit: It is a duplicate of this question: iOS 7 parallax effect in my view controller
Thanks for you help.

That feature is called a 'parallax effect' and there is built-in support in Cocoa Touch for it. You don't need to re-invent the wheel to add that effect to your apps. Search the web for information on UIInterpolatingMotionEffect. Here, in particular, is Apple's brief documentation on it.

Related

Adding Zoom-in functionality in ARKit Session(camera view not the AR object)

I am working on an app with ARKit, and want to detect objects at some distance by using ARKit camera session. Could anyone help if we can add any zoom in functionality to the AR Camera view?
Although did not find any concrete answer to this question but, some people suggested to transform the ARView itself which I think cannot be the right solution. Also, few talked about adding gesture on ARView and do some transforming but not accepted. Please help if anyone has any idea.
As per my understanding and experience in ARKit, camera zooming is not supported in ARKit. But you can zoom on the ArView.
Check this link please:
Reference to camera zoom issue ARKit
Another reference regarding scenekit

Creating a 360 photo experience on iOS mobile device

I am interested in VR and trying to get a bit more information. I want to create a similar experience on iOS where I can take a 360 image and be able to view it on a iOS device by tilting the phone around and using the devices gyroscope, as I tilt the phone around it will pan around the 360 image (like on google street view where you can use the tilt gesture).
And something similar to this app: http://bubb.li/
Can anybody give a brief overview how this would be do-able, any sources that could help me achieve this, API's etc...?
Much appreciated.
Two options here: You can use a dedicated device to capture the image for you, or you can write some code to stitch together multiple images taken from the iOS device as you move it around a standing point.
I've used the Ricoh Theta for this (no affiliation). They have a 360 viewer in the SDK for mapping 360 images to a sphere that works exactly as you've asked.
Assuming you've figured out how to create 360 photospheres, you can use Unity and Unreal, and probably development platforms to create navigation between the locations you captured.
Here is a tutorial that looks pretty detailed for doing this in Unity:
https://tutorialsforvr.com/creating-virtual-tour-app-in-vr-using-unity/
One pro of doing this in something like Unity or Unreal is once you have navigation between multiple photo spheres working it's fairly easy to add animation or other interactive elements. I've seen interactive stories done with 360 video using this method.
(I see that the question is from a fairly long time ago, but it was one of the top results when I looked for this same topic)

Detect vertical or horizontal position iOS

I have an iPhone 4 and I would like to detect whether the phone is vertically straight (see first image) or horizontal (see second image). I am not talking about portrait vs landscape view detection.
I am new to iOS development and I think what I am looking for is the accelerometer (specifically the pitch property perhaps) or gyroscope. I have looked online and I am not quite understanding the code or whether I am looking for the right properties. Can someone please tell me how this can be done?
Thank you!
Try the accelerometer api and check the position of the phone.
here is a tutorial: http://tech.pro/tutorial/968/iphone-tutorial-reading-the-accelerometer

Motion detection of iOS device in 3d Space

Ive been working with the iOS sensors a bit off late and i wanted to write an app that would accurately track the motion of the phone in space. I wanted to know if its possible to track the motion of the device and detect gestures, such as drawing a circle with your phone or even moving in a straight line.
I've been searching online about this, and i wanted to know two things:-
1.Is it possible to do this with the CoreMotion framework.
2.If Yes, what is the alternative for older devices that do not support CoreMotion. Without the double integral method using the accelerometer!
This would really help!
Any other alternative ideas are most welcome!
Thanks in advance!
As your write, you cannot do the double integral.
For gesture recognition, I would try dynamic time warping. See my earlier answer here.

iOS augmented reality with compass and location

I'm trying to develop a mini "Around Me" like using camera, compass and location. I would like to display place's images on my screen.
For the moment I have my location and my orientation with compass. I would like to know how can I determine the position of the place I want to display.
Thanks for your help ;)
Once you have relative distance and bearing, which you can determine from two points in the same coordinate space using algorithms found on this page, figuring out where a known coordinate is with respect to a known viewpoint is basically a perspective projection, the math is outlined on this Wikipedia article. The rotation of the camera is given by the compass, and the tilt by the accelerometer (the position is of course, GPS).
I'm trying to find a better document - there are a couple of extra things to consider - like the camera parameters etc, but this is a good starting point.
If it's too involved (like if you're not comfortable with rotation matrices) we can break it right down to the simple trig.
The code in the iPhone ARKit project does this, and quite a bit more. While you may not be able to use their complete library, it is a great reference on the subject of augmented reality.
Check out 3DAR, it lets you add an AR view to a MKMapView app very easily. There's a video tutorial on this process, as well as some sample code, on the 3DAR site, www.3dar.us
You can create a location based AR app in Junaio. It's an AR browser. Free to use and deploy in (as long as it's not a custom app and in Junaio).

Resources