I am running into hardware issues that perhaps someone here knows a workaround. I am using a PC and windows.
For several years I have been making interactive installations using video tracking: the Jmyron library in Processing, which has functioned marvelously for me. I use this set up: cctv type microcameras to a multiplexer, the I digitize this signal via a firewire cable to a pci card. Then Processing reads these quads (sometimes more) as a single window, and it has always worked (from windows xp all the way to 7). Then comes windows 8: Processing seems to prefer the built-in webcam to the firewire bus. On previous version of windows, the firewire bus would naturally override the webcam, provided I had first opened a video capture in Windows Maker, and then shut it down before running the Processing sketch. In Windows 7, which had no native video capture software, I used this great open source video editor called Capture Flux. The webcam never interfered. With Windows 8, no matter what I try, Processing defaults to the webcam, which for my purposes is useless. I have an exhibition coming up real soon, and there is no way I am going to have the time to rewrite all that code for Open CV or other newer libraries.
I am curious if anyone has had similar problems, found a work around? Is there a way of disabling the webcam in Windows 8 (temporarily of course, because I need it to be operational for other applications), or some other solution?
Thank you!
Try this:
type "windows icon+x" choose device manager (or use run/command line: "mmc devmgmt.msc")
look for imaganing devices, find your integrated webcamera
right click on it and choose disable - now processing should skip the device.
Repeat the steps to reenable the device.
Other solution would be using commands in processing:
println (Capture.list()); (google it on processing.org) this way you will get all avaliable devices and you can choose the particular one based on its name.
Hope this helps.
Related
Host machine: Debian 10 running NoMachine 7.2.3
Settings:
Specified H264
User Hardware Encoding enabled
Use Specific Frame Rate enabled (60FPS)
Use Acceleration enabled
Client: Windows 10 running NoMachine 7.2.3
Both machines have monitors attached.
Using NX protocol for connection.
FullScreen / Scale to Window / Desktop is currently 2560x1440 (reduced from native while testing this issue)
Specific issue:
I do a ton of work in the terminal and when viewing desktop via nomachine, the terminal caret is randomly not visible. The same issue is less noticeable with right click menus and other areas of "visual updates in small screen space." If this were another remote desktop vendor I would try to find the "don't update just regions" setting to force the entire display to update regularly, but I can't find similar settings for nomachine. I have a dedicated gigabit connection between the two machines with no other traffic on that line, so bandwidth is not an issue.
To recreate:
I disabled caret blink (using universal access / accessibility settings) so the caret is a solid block in terminal / vi. If I edit a text file in vi and move up and down, the caret will only update visually every other line or so (verified on the physical screen it is moving correctly). Same if I highlight or insert, etc. You inevitably miss a character or so or lose your place).
I have tried changing speed vs quality slider, resolutions, swapping from h264 to VP8, etc.
I have disabled:
multi-pass display encoding
frame buffering on decoding
client side image post-processing
Nothing seems to change this specific issue. Yes I can make dragging a quarter-screen-sized terminal window smoother, but that doesn't help me follow the caret in vi/vim. Both machines are nicely spec'd (client has 16G / RTX2080, server has 32G / GTX1080)
Is there a way to get nomachine to update all the screen all the time, or at least better refresh small areas like a terminal caret?
(OP): Based on a night of troubleshooting, the issue seemed to be either:
An issue with the Debian install of the nvidia drivers
The server machine is a laptop with a broken main screen (but with an HDMI external monitor plugged in). The Debian X-server may have been confused as to whether it was headless or not and caused issues with nomachine (which tries to detect headless and start a virtual session).
The solution to this exact problem would be to disable the GUI and force a virtual session, per https://www.nomachine.com/AR03P00973 (dummy dongles won't work because the laptop's main display is not a standard plug).
In my specific case, I needed GUI access on the server at times so I couldn't use the above methods, and I could not remedy the problem with Debian, so I wiped the system and installed Ubuntu 20.04, which is more forgiving with graphics drivers and monitors. After setting up the Ubuntu system as similarly as possible to the Debian system and letting the proprietary nvidia drivers auto install, nomachine connected at the same resolution and worked perfectly, without the lag in small screen areas.
I'm in an unusual spot. The built in screen on the laptop I'm using right now is broken so bad I can barely see anything besides random flickering lights on it. Because of that, I got and plugged in a monitor to my laptop. This moniter works fine once Windows 7 has already loaded.
The issue is: I'm trying to open the BIOS/boot menu to live boot off of my USB, but I can't see anything. How do I make it so that I get visual output on my monitor early enough to open the BIOS/boot menu?
If it's a Dell laptop, try FN&F1 or F1
Dell Forum
I know this sounds weird, but I am making a program in Dart in browser that takes input from the Griffin Powermate USB know controller. So far, I have been using the included drivers and mapped a click on the knob to a click on the mouse. Now I have run into a problem with this: The default drivers send both "buttonDown" and "buttonUp" events at the time of release, and nothing at the time when I would expect "buttonDown"-events.
I assume I will have to ignore the drivers and dive into the details. So I have the following questions:
Is there a magical driver replacement(preferably for OSX, but I can install another OS) that fixes the problem immediately?
Is there a way to access generic USB HID devices in Dart? (like 'node-hid' for node.js)
I need to detect the corner/edge of a document in a captured image in Windows Phone 7. I cannot use openCV as WP 7 does not have support for native code lib.
Can anybody suggest me some algorithm or open source library that I can use for this purpose.
I want to do the similar thing on Windows Phone app as posted in another SO question: DETECT the Edge of a Document in iPhoneSDK
you can try using A-Forge for this task.
the A-forge .NET have great functionality and I think it will not be a problem to use it on windows phone
I made a small sketch with processing using opencv to face detect and track. But when I export as an application and start it nothing happens, same as an applet. First I exported the applet and thought this must be a security issue, and struggled with some hack for processing to sign your java applet.
http://processing.org/hacks/hacks:signapplet
With no luck I thought I would atleast be able to run it as an application but no success.
Anyone know what to do?
If you're on OSX 10.6 there is a known bug with the quicktime camera stuff not working in 64-bit mode. Make sure both your install of processing and the exported application are set to run in 32-bit mode.
Right click -> get info -> Open in 32-bit mode