I am working on a project which needs to detect left and right face turn in face recognition. I dont even know whether API's are available even to detect it.
I am able to detect face using OpenCV, but when I turn my face it is not even detecting the face.
Any help is greatly appreciated,
It's called profile in OpenCV and is already part of OpenCV. I guess you must use another classifier for this type of detection
Take a look at these links:
http://alereimondo.no-ip.org/OpenCV/34
http://tech.groups.yahoo.com/group/OpenCV/message/78936
Related
May be this sounds like a stupid one, but i really curious to know that, what is the difference between "Face Detection and Face Recognition" in iOS perspective? And in what case or which kind of situation should i use the two of them. I am new in iOS and have never any previous tinkle down experience about iOS face Detection/Recognition related thing. I am going to make an application, where i have to detect user face (by camera, not after taken photo) with database picture collection. Please give you response if any, and please don't misunderstand my question. O:)
-Thanks a lot in advanced.
In General:
Face Detection:
Detect the face in the image. It searches general human face like segment in the whole image. Output may be one or more than one. The output will be a rectangle or rectangles on the faces in the image.[Paul viola method]
Face Recognition:
Recognize input face from the already trained database with highest match score. A single face should be given as input, and the output will be a name, or class name or unknown face.
[PCA, LDA]
iOS has face detection, but no face recognition. It can tell you where the faces are in an image but can't tell you who they are.
If you want to use the face detection, start with AVMetadataFaceObject or a tutorial like this one.
I am working on an app in iOS that will occur an event if camera detects some changes in image or we can say motion in image. Here I am not asking about face recognition or a particular colored image motion, And I got all result for OpenCV when I searched, And I also found that we can achieve this by using gyroscope and accelerometer both , but how??
I am beginner in iOS.So my question is , Is there any framework or any easy way to detect motion or motion sensing by camera.And How to achieve?
For Example if I move my hand before camera then it will show some message or alert.
And plz give me some useful and easy to understand links about this.
Thanx
If all you want is some kind of crude motion detection, my open source GPUImage framework has a GPUImageMotionDetector within it.
This admittedly simple motion detector does frame-to-frame comparisons, based on a low-pass filter, and can identify the number of pixels that have changed between frames and the centroid of the changed area. It operates on live video and I know some people who've used it for motion activation of functions in their iOS applications.
Because it relies on pixel differences and not optical flow or feature matching, it can be prone to false positives and can't track discrete objects as they move in a frame. However, if all you need is basic motion sensing, this is pretty easy to drop into your application. Look at the FilterShowcase example to see how it works in practice.
I don't exactly understand what you mean here:
Here I am not asking about face recognition or a particular colored
image motion, because I got all result for OpenCV when I searched
But I would suggest to go for opencv as you can use opencv in IOS. Here is a good link which helps you to setup opencv in ios.
There are lot of opencv motion detection codes online and here is one among them, which you can make use of.
You need to convert the UIImage ( image type in IOS ) to cv::Mat or IplImage and pass it to the opencv algorithms. You can convert using this link or this.
I am very new to opencv and able to install it so far. I want to compare a face with other different faces available in library and to find out the closest match. I have tried different features but couldn't find the closer answer.
Any suggestion to choose a detector.
dimensions of input image and images in library are same.
thanks in advance
i think, you want face recognition (who is it?), not detection (is it a face?).
look here for what opencv has to offer there
I'm currently working on a project, where I need to detect a face and then take a photo with the camera. (after the camera focused everything correctly).
Is something like this possbile in iOS?
Are there any good tutorials on this?
i would suggest to use opencv for this as it has proven algorithm and fast enough to work on image as well as video
https://github.com/aptogo/FaceTracker
https://github.com/mjp/FaceRecognition
This solution will work for android too using opencv port to android.
Use GPUImage for face detection.
Face detection example is also available in GPUImage.
see last point in FilterShowCase example project of GPUImage for face detection.
iOS 10 and Swift 3
You can check apple example you can detect face
https://developer.apple.com/library/content/samplecode/AVCamBarcode/Introduction/Intro.html
you can select the face metedata to make camera track the face and show yellow box on the face its have good performace than this example
https://github.com/wayn/SquareCam-Swift
I am working on a demo in ios where a user takes a photograph from camera or from photo gallery, and then some how determine whether the person is japenese, chinese, asian from the skin color, and shape of eyes. Is there any ios library that can help me to achieve this goal.
If there is no such library, can anyone guide if it is even physible or what approach should I follow ?
You're probably going to look into OpenCV for any kind of feature detection. From what I understand this framework allows you to detect specific colors shapes etc.. and should help you achieve your goal.
Then as far as any kind of basic face detection goes, you can use Apple's CIDetector.