Heyho,
now, I know that similar questions have been asked throughout the internet but the answers were always the same "don´t do it because the resolution is bad". For the project I am working on the quality is not so important. I just need to know if and how it is possible to use the Vive as a simple (second) monitor. I want to simply see my for example browser window just like I am seeing it on a normal monitor. Stuff like Bigscreen VR or Virtual Desktop aren´t what I am looking for.
Thanks in advance for the answers.
Which operating system? If it is on Linux and X11:
The nvidia driver has the option
Option "AllowHMD" "yes/no"
that if set to yes will present the HMD as a regular display device.
Related
I just set up ncmpcpp on the ubuntu side of my macbook pro, and I'm trying to make an iphone app to control it. How would I go about doing this?
Should I use bluetooth or wifi? (which one would be easier)
And then how would I go about implementing it? What packages should I install in linux? And how would I use them?
I know it's kind of a big topic, and I have several broad questions, but if you can answer any of them or provide any information that would help, I would be incredibly grateful!
Thanks!
That sounds like an ambitious project with a significant scope.
Whenever ideas like this crop up, it's good to take a step back and ask: "What am I trying to accomplish?".
Are you simply wanting to run a music "server" on your Ubuntu portion, and access it via your iPhone? Or are you trying to make a "remote control"?
There are likely apps that will do what you're wanting to do... I know, I know mdp/ncmpcpp is super neat looking, but... practicality!
I'm trying to get the position of finger on laptop touchpad in Delphi. Not the position of cursor on screen. So I can use it for drawing purposes. Is this possible? How can I do this? Is there any Windows API or any component for this?
Thanks for your help.
Update
I found a software for Lenovo touchpad that does the exact thing. It only shows the position of fingers on touchpad and PEiD says it's been written with Visual C++. So I guess it's a possible thing but as David Heffernan said it depends on manufacturer of the touchpad and it's hardware specific.
Coincidentally, I've just spent the last 30 minutes researching this very thing.
Windows supports this through the touch and gestures APIs. These were introduced in Windows 7 but touchpad drivers didn't tend to offer the necessary support until Windows 8 arrived and made it a logo requirement.
Synaptics and Alps seem to be the principal touchpad manufacturers and they have both released updated drivers for Windows 8 which also work on Windows 7. "Multitouch" is the keyword to search for. This is touchpad-model dependent though; I can't find an update for older Alps devices.
In short, this should work on a "Designed for Windows 8" laptop. It may work on Windows 7 and if it doesn't you may be able to get an updated driver.
The short answer is generally no, this is not possible. Touchpad drivers present to the operating system such that they appear and behave like a mouse does. Absolute coordinates are not available. For this application you need a proper touchscreen device or tablet, at least if you are looking for a general solution that is supported by the operating system.
Some touchpads may provide this information through a hardware-specific driver, of course, but you would need to support, where this is even an option, each device independently. Synaptics, for example, provides an SDK and drivers that can expose the absolute coordinate information.
For tablets or other full-screen digitizers that are supported as "Pen and Touch" inputs, this information is usally obtained through the WM_TOUCH message. Some advanced touchpads may support this - you can always query to discover what features are supported. For those that are, you have to register your application's window to recieve touch messages as detailed here :
Getting Started with Windows Touch Messages
Upon receiving a WM_TOUCH message you can obtain detailed information by immediately passing the touch handle to GetTouchInputInfo. Which returns an array of TOUCHINPUT structures, each carrying information about each active touch point on the digitizer surface.
My laptop has two video cards, a high powered NVIDIA one and an onboard Intel one. When I call IDirect3D9::GetAdapterCount however, it only finds the onboard Intel one, probably because the high powered one is being hidden.
I'm able to go into my laptop settings and tell it 'force choose' the NVIDIA card, and then it works, but this is not an acceptable solution for my end-users. I've also noticed that when I run Battlefield3, it's able to properly find the NVIDIA card even without 'force choose' enabled. Maybe there's a special white-list that has Battlefield listed? Or some other secret method?
Any ideas how to acquire that elusive card?
Are you sure the intel chip is enumeratable? Quite often its not. By sticking in a discrete GPU the sandybridge (and older) chipset is generally disabled. You probably want to check the Nvidia optimus test tool.
GetAdapterCount will actually returns count of the monitors in system, not videocards. And as far as I know there is no way to force choose it programmatically.
If you talking about nVidia optimus technology, it choose videochip using driver settings.
I'm writing LabVIEW software that grabs images from an IMAQ compatible GigE camera.
The problem: This is a collaborative project, so I only have intermittent access to the actual camera.I'd like to be able to keep developing this software even when the camera isn't present.
Is there a simple/fast way to create a virtual or dummy IMAQ camera in software? Ideally I'd like the dummy camera grab frames from an AVI or a stack of JPEG's. Something like this must exist, I just can't find it on Google.
I'm looking for something that won't take very long (e.g.< 2 hours effort) and that is abstracted away through the standard LabVIEW IMAQ interface, so that my software won't know or care whether its dealing with a dummy camera or an actual camera.
You can try this method using LabVIEW classes:
Hardware Emulation Using LabVIEW Classes
If you have the IMAQdx driver, you might consider just buying a cheap USB webcam for $10.
Use the IMAQdx driver (assuming you have it), and then insert the Vision Acquisition Express VI, and you can choose AVIs or even pics as a source.
Something like this: GigESim is a camera emulation software. Unfortunately it is proprietary and too expensive (>$500) for my own needs, but perhaps others will find this link useful.
Anyone know of a viable Open Source alternative?
There's an IP Camera emulator project that emulates IP camera with python. I haven't used it myself so i don't know if it can be used by IMAQ.
Let us know if it's good for you.
I know this question is really old, but hopefully this answer helps someone out.
IMAQdx also works with Windows DirectShow devices. While normally these are actual physical capture devices (think USB Webcams), there is no necessity that they have to be.
There are a few different pre-made options available on the web. I found using Open Broadcaster Studio and this Virtual Cam plugin to be easy enough. Basically:
Download and install both.
Load your media sources in the sources list.
Enable the VirtualCam stream (Tools > VirtualCam). Press Start.
I have been developing using a single 15.4" laptop for a while (duh!). Actually I am quite comfortable. I use compiz Grid, scale, window and some other nice add ons, to easily navigate.
I am now impressed by other dual monitor related questions on Stackoverflow and am buying more monitors.
Since I am not exactly used to these multiple monitor setups, I have a few (possibly basic) queries.
If I buy just one monitor, will I be able to set up it for a different resolution than on my laptop, on Ubuntu. What packages I need to install on ubuntu to better manage multiple monitors.
If I buy 2 22" monitors and intend to use the laptop screen as well, wont the third one appear like an odd man. Does it happen even if I have just 2 monitors.
I have used Dell and HP monitors in my earlier company and I feel HP is so much better. Obviously I would love to have a monitor that has little corners and nearly all visible screen, like iPhone. Is there a specific recommended model.
For a 3 monitor setup, is it absolutely necessary to buy an extra graphics card. Are there any other better solutions.
You can have two monitors at different resolution on Ubuntu just fine. I used to connect an Ubuntu laptop to a TV using HDMI and it worked fine. Something annoying is what monitor gets to be the main one, because you get all your desktop there. I thing generally the biggest one is the main one, which is not what I wanted for the TV, but probably what you want.
I'm not sure if it's possible to get a third monitor in any way on a laptop. I'd recommend you to go for the biggest monitor you can (budget and laptop capabilities, can you go to 30"?) and use the laptop as secondary screen space (useful for a browser, or docs, or im/tweeter/mails, etc.
A CRT monitor can be set up with most any resolution you like, while a LCD monitor works best at it's max resolution.
You can mix monitors with different resolutions without problems in Windows, I can't imagine that it would be very hard in Linux either.
Be carefully dual head on Linux still sucks. The only time i got a triple screen configured with a GUI frontend and without .XConfig hacks was with Mandriva Spring 2007 never before and never after with any of Suse, Mandriva, CentOS.
It really sucks. It seems that configuration like dual head on one card and single head on another one is just not in the head of the programmers. You can enable all cards (maybe dual head is not always supported) or none.
Windows and MacOSX are the only systems which work fine. But MacOSX (and Hackintoshs) do not support dual head on all supported cards, for example my two 7300GTS cards can't run as dual head on MacOSX.