I am working on 1 project start up. And i want to do the following this. 1)Detect the camera by wifi and camera id. 2)After detecting the camera.Fetch the stream from the camera to the phone 3)play that stream into the player.
Now i had done google and found that by RTSP we can achieve the following this. But i am not sure or dont have the code to detect the camera from the camera id or by wifi and the detect the stream from that. Can anyone guide me in this with the some sample source code.
If your camera supports ONVIF you can use it to detect it.
Related
Does anyone know of a way to stream wirelessly from a cctv camera to an instagram live feed?
I've looked into yellow duck a bit but I'm not sure if that would work.
I have a analog camera connected to EasyCap video capture device. When I run a basic code which opens webcam video using OPENCV, I can access my in-built webcam but not the other analog camera.
How would you connect any other camera (FPV, IR, etc) to the PC such that OPENCV can access it.
Thanks.
i struggled with the same problem and hope it helps!
the original thread + ANSWER
also relevant XKCD
one more observation: from your description it looks like you already have a webcam running on the laptop (in-built webcam maybe?) you might want to disable it in system manager so as to guarantee that your analog camera cam_index is zero for certain. Otherwise if you leave the webcam enabled as a device, then your analog cam will most likely be incremented to cam_index=1 which amusingly enough seems to be confirmed by it crashing on cam_index=1.
Arguable not a great method to find your camera's index but there you have it!
You can set which camera to connect to open by changing the following deviceID to the desired device you want:
CvCapture* capture = cvCaptureFromCAM(deviceID);
or new API:
VideoCapture cap(deviceID);
Check out documenation for more info.
Use the deviceID of the analog camera instead of the in-built one.
I would like to build a straight forward app that can intercept the usb sound input via the ipod/ipad usb camera kit. Is this possible and if so what area within Core Audio should I look at?
Thanks for your help and any help is helpful!!!
iOS automatically reroutes microphone input from suitable generic USB audio input devices using the camera connection kit to all iOS audio APIs.
We are trying to do a Video recording on IOS with a Overlay of GPS Co ordinates shown in the top right corner.
We can capture video fine on Appcelerator. But not having much luck with getting the overlay in the recording.
We'd like to do this realtime rather than in post processing.
Unfortunately, This is not something you can do using Appcelerator's video recorder. Your best bet is to write a module in native code that handles the recording and realtime overlay.
I want to make something like they have at US dmv's where you sit down and it takes your picture, maybe like photobooth.
I want to connect a high end camera via usb, fire the camera and get the picture.
There's the Picture Transfer Protocol http://en.wikipedia.org/wiki/Picture_Transfer_Protocol a nastly little thing. All the cameras I held in my hands so far, claiming they had proper PTP support failed it somewhere. But in theory one can use PTP to remote control a camera, i.e. trigger the shutter, retrieve the picture and so on.
Rater than reimplementing the whole thing I recommend you get some readily usable PTP library. There are some open source ones listed on http://ptp.sourceforge.net
The easiest method is probably to use OpenCV: http://opencv.willowgarage.com/wiki/
If you need a high end camera - most digital SLRs have a tethered mode where you can control the camera, fire the shutter and retrieve the image data. Each camera maker has a proprietary (but normally free) sdk.
For a webcam type camera - these normally run in video mode, you simply grab an image out f the video stream - as PaulR says - use openCV