Run OpenCV Viz on iOS - ios

Is it possible to port my existing Viz3d-based visualisation to iOS? It is using VTK under the hood so in theory, it should be doable since VTK can run on iOS.
If yes, is there a working example of this, or can you provide one?

Related

When using opencl with torch, lua uses integrated graphics not dedicated gpu

I am using Andrej Karpathy's code to train a rnn. When I give the flag "-opencl 1" to tell it to use an opencl gpu, it uses integrated graphics and nothing else.
I tried reinstalling cltorch and using different flags, but nothing has seemed to work. To add to this I can't see if my gpu is under load because I'm on macos. I looked through the code and I could't find any errors, but I have little experience with lua.
Code can be found here: https://github.com/karpathy/char-rnn.
I expect with the flag "opencl 1" or something to the like, my radeon pro 560x will be used to train on my dataset and not my cpu or integrated graphics.
When reading instructions I thought just -opencl flag needed to be used but it turns out -gpuid needs to be used in conjuction with it as well. This was also after a reinstall of torch and opencl drivers, so that could have been a problem as well.

Is there any Fourier transform Function in Appcelerator?

I am trying to take the FFT of an image inside an application for iOS and android. I was wondering if there is a built in function in appcelerator.
If not, What is the easiest way to implement it.
Appcelerator is a company. What you're trying to find is a function for Titanium. It actually is even easier than that. Titanium is build in JavaScript, and it supports nodejs packages and uses the commonjs approach.
So any commonjs module, or nodejs package (that doesn't rely on the nodejs core packages) are supported by Titanium.

Using Google Cartographer with Turtlebot in a custom world?

I am a beginner in robotics and I am trying to use Google Cartographer to make my simulated Turtlebot build a map autonomously of its environment.
I have done already all the tutorials in ROS and I can make the robot build the map using teleoperation, but I don't know how to make it build the map by itself. I want to use Google Cartographer for this.
I can however run the demo provided by Google and it works (it builds the map of a Museum).
# Launch the 2D depth camera demo.
roslaunch cartographer_turtlebot demo_depth_camera_2d.launch bag_filename:=${HOME}/Downloads/cartographer_turtlebot_demo.bag
The questions:
How can I run it on my own world instead of the bag file of that museum?
Does it need a yaml map like the one I built with teleoperation? what is the command to make it use my yaml map instead of a bag file?
Can I use a .png image with yaml context?
Could it use the gazebo simulated worlds that are .sdf? What is the command to input this then?
These are the specifications I have:
Ubuntu Xenial
ROS Kinetic
Gazebo 7
Turtlebot2
Google Cartographer for turtlebot
Thanks a lot! It's like I understand the concepts but I don't know how to link things together to make them work.

OpenCV codes in Code Composer (CCStudio)

I'm using CCStudio v5 to implement a vision system and want to use OpenCV functions in my code, but I don't know it is possible to use OpenCV code in CCStudio or not!
How I can import OpenCV library into my CCStudio project? Is this depends on my hardware?
There is no official release of OpenCV for system without OS. OpenCV library is available for Windows, linux, mac, Android and Ios operating system.
Here
you can find a link which explain the challenges of having OpenCV running on microcontrollers

openCV from window to ios

I want to train openCV from a server and send the xml generated by openCV to an ios device where an app will recognize the face using the xml trainned by the server. I will use openCV in both app but the server has window (trainning) and the device has ios (recognition).
So my main question is very simple:
The xml generated in openCV window version can be used an openCV IOS version without any trouble? Somebody made something similar who can give me some tips?
In window I will use .Net.
I think they won't have trouble because they are same libraries (openCV), so I suppose they have same internal algorithms but I want to be sure before start the project.
Best Regards and thanks for your time
There is no problem, but you must train with images taken from your devices. It is normal to have multiple xml sets depending on your different cameras. Normally you release these with the binary, and not as a download but still...

Resources