Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
Questions asking us to recommend or find a tool, library or favorite off-site resource are off-topic for Stack Overflow as they tend to attract opinionated answers and spam. Instead, describe the problem and what has been done so far to solve it.
Closed 8 years ago.
Improve this question
I'm developing an app that makes use of an augmented reality feature whereby the user can point the device's camera at a point of interest such as a building and get information on that building.
Are there any frameworks out there for the iOS SDK that allow for this functionality?
I've looked at the Vuforia framework, however this doesn't seem to support this feature.
If anyone could point me in the right direction I'd appreciate it.
Thanks in advance!
If you can put some markers on the buildings, you can try ARToolKit.
If you cannot, it will be a much more difficult problem.
If the buildings look similar, it is impossible to distinguish them by their appearance.
If they look very different, you may try feature detection techniques such as SIFT or SURF.
GPS information will make your job easier, if you know the exact location of these buildings.
Related
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
Questions asking us to recommend or find a tool, library or favorite off-site resource are off-topic for Stack Overflow as they tend to attract opinionated answers and spam. Instead, describe the problem and what has been done so far to solve it.
Closed 8 years ago.
Improve this question
Currently trying to include maps within my ios 7.1 app. Essentially what I am trying to do is read in markers and display them on my map. I have a considerable number of marker to display some hundreds. Is it possible using mapkit to read in from a sqllite table these markers and display them.
thanks in advance
Mapbox is built on top of leaflet, you can use that.
Also, you can always use a uiwebview and put something in there.
But I do suggest either google, or Apple maps, they are pretty stable.
I tried using mapbox in an iPhone project before, but ended up sticking to Apple maps (mapKit), if you know what you want to do you can create some awesome stuff.
Hope that helps?
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
Questions asking us to recommend or find a tool, library or favorite off-site resource are off-topic for Stack Overflow as they tend to attract opinionated answers and spam. Instead, describe the problem and what has been done so far to solve it.
Closed 9 years ago.
Improve this question
I want to make an offline iPhone application that can grab text from a picture. Can anyone suggest the best library which I can use. I heard ZBAR and ZXING can be used only for barcode reading. Is there any other OCR Libraries for iOS to read text from images. I expect your valuable suggestions as soon as possible.
Thanks in Advance..!
Currenlty offline OCR is possible only with Tesseract
You can get source code here
Here is the good tutorial about how to use Tesseract
Also you can perform OCR on multiple language. You can dowlnoad other language trained data here
You might want to look here. They offer a large variation of language they can detect and i heard only the best from their framework.
You can also test the framework live: live demo
Unfortunately i think it is very expensive to use this framework. But if you want an A1(all my Breaking Bad fans ;-)) framework this might be your best shot.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
Questions asking us to recommend or find a tool, library or favorite off-site resource are off-topic for Stack Overflow as they tend to attract opinionated answers and spam. Instead, describe the problem and what has been done so far to solve it.
Closed 9 years ago.
Improve this question
I'm working on a project in which I want to add a draw/sketch feature that draws by the user's touches on screen. That way they can make simple drawings on a view. This is a concept used in many picture apps (i.e. SnapChat). I've done many Google searches but haven't found anything out there that I could use. Does Apple have a guide or class for this or is there are any algorithms or code out there that's available to the public? Please provide me with references or information regarding this topic. Thanks in advance for your time.
I really like the ray wenderlich tutorials.
They also have a tutorial on how to make a simple drawing app.
http://www.raywenderlich.com/18840/
I think it should give you a hint on how drawing works.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
Questions asking us to recommend or find a tool, library or favorite off-site resource are off-topic for Stack Overflow as they tend to attract opinionated answers and spam. Instead, describe the problem and what has been done so far to solve it.
Closed 9 years ago.
Improve this question
I want to do something like this:
With iPhone's camera and AR tech, I could stick one virtual paper to the desk in front of me in the real world and wherever I put my iPhone around that place I stick the paper and it still seems that the paper is on the original position where I've put it at first.
If I want to make it come true, what do I need? And what SDK is more suitable for me to use?
This question is pretty old now. Is it still open? You can try out Qualcomm Vuforia
Take a look at http://socialcompare.com/en/comparison/augmented-reality-sdks to see which AR SDK support iOS. You can start by looking into ARToolkit, Metaio, String, Vuforia, and Wikitude.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
Questions asking us to recommend or find a tool, library or favorite off-site resource are off-topic for Stack Overflow as they tend to attract opinionated answers and spam. Instead, describe the problem and what has been done so far to solve it.
Closed 9 years ago.
Improve this question
I am looking for good tutorials that go through every step of creating an AR application. It would be beneficial if it also covers some of the theory behind optics and such.
A good setup is to use ARToolKit plus osgART.
The tutorial of ARToolKit also explains also some of the principles of Computer Vision:
http://www.hitl.washington.edu/artoolkit/documentation/
osgART is the ARToolKit for OpenSceneGraph and abstracts some of the principles of using a computer vision based tracking in a scenegraph application
Both toolkits have a large and vibrant development community.
MXRToolkit have good documentation and is suitable for beginner. There are some other toolkits too.
http://mxrtoolkit.sourceforge.net/
Not a trivial learning curve but OpenScenegraph is probably the main FOSS app for large applications like flight simulators.
I haven't used it but it might be worth looking at: http://developer.qualcomm.com/ar