How to use graphics memory as RAM? [closed] - memory

Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 11 years ago.
Improve this question
Since graphics cards provide large amounts of RAM (0.5GiB to 2GiB) and API access to the GPU is not that difficult with CUDA, Stream and more portable OpenCL I wondered if it is possible to use graphics memory as RAM. Grahics RAM might have a larger latency (from CPU) than real RAM but its definitively faster than HDD so it could be optimal for caching.
Is it possible to access graphics memory directly or at least with a thin memory management layer within own applications (rather than free usable for the OS)? If so, what the the preferred way to do this?

Yes, you can use it as swap memory on Linux. Refer to the link here for more details.
With Linux, it's possible to use it as swap space, or even as RAM disk.
Be warned
It's nice to have fast swap or RAM
disk on your home computer but be
warned, if a binary driver is loaded
for X, it may freeze the whole system
or create graphical glitches. Usually
there is no way to tell the driver how
much memory could be used, so it won't
know the upper limit. However, the
VESA driver can be used because it
provides the possibility to set the
video RAM size.
So, Direct Rendering or fast swap.
Your choice.
Unlike motherboard RAM and hard
drives, there aren't any known video
cards that have ECC memory. This may
not be a big deal for graphics
rendering, but you definitely don't
want to put critical data in it or use
this feature on servers.

Related

Can a VM with more VRAM than RAM do machine learning efficiently? [closed]

Closed. This question is not about programming or software development. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 2 days ago.
Improve this question
I'm working on a virtual machine that has been given 12 GB of RAM and has a Quadro RTX 6000 with 24 GB of VRAM. I'm trying to do machine learning on this virtual machine.
My intuition right now suggests that it is not using the video card as efficiently as it could be with the limited amount of RAM, and would work better with more. Is this correct?
The following suggests it is so, but is not very clear on that.
OpenCL - what happens if GPU memory is larger than system RAM
In short, how much RAM should I expect to need for machine learning and computer vision for this video card typically?
It very much depends on the software you are using. In some cases, GPU software can use significantly more VRAM than RAM, when the model runs only on the GPU and there is no need to have a copy of it in RAM.
As an example, although CFD and not ML, the FluidX3D software uses between 3.2x and 5.4x more VRAM than RAM. Here in your case the 24GB VRAM capacity would still be the limiting factor.
If it's 1:1 RAM:VRAM allocation, then you're limited by the 12GB RAM. In the end, you have to test your software and check the allocation ratio with tools like top/htop and nvidia-smi.

Should someone focus on having as less memory leaks as possible or having the fastest computing time? [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 7 years ago.
Improve this question
I was wondering:
When programming, should one focus on having as less memory leaks as possible or more on the CPU computing time?
What are the pros/cons?
Thanks!
#basile's answers is correct. And it's worth clarifying what you mean by "memory leaks".
The strict definition of a memory leak is when a block of dynamic memory is never deallocated after being used. I would argue that this is never acceptable under any circumstances.
But, fortunately, avoiding memory leaks and using CPU time efficiently are not usually competing ideals.
It sounds like your question is more along the lines of "is it more important to cut down on CPU cycles, or is it more important to use as little memory as possible?" This is a common and completely valid question because there are many instances in programming where you can shave CPU cycles by dumping stuff into memory, or on the other hand, you can save memory by brute-forcing something.
Alas, there's no clear-cut answer. There are times when one is appropriate, and times when it goes the other way. As you grow as a programmer, you learn best practices for being efficient with both. And, in the real world, as long as you program responsibly, you will probably never see an actual situation where you have to sacrifice one or the other. Especially with the speed of modern chips.
If a program runs very quickly (e.g. a small fraction of second) and you want to run it zillion times, memory leaks do not matter at all (because in a very short time it will allocate only a small reasonable amount of memory, and the OS will reclaim the memory used by a process when that process terminates).
If your program does not run quickly (in particular if it runs continuously, e.g. because it is a server or a daemon), memory leaks are of paramount importance.
BTW, memory leaks may mean slightly different things (not the same in C as in Ocaml).
If coding in C or C++, use valgrind to detect memory leaks.
Read also about garbage collection (see also the GC handbook). At the very least, the terminology and the algorithms related to GC should concern you. In C, you might sometimes consider using Boehm's conservative garbage collector.

Memory breakdown based on its speed [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 8 years ago.
Improve this question
In one technical discussion the person asked me which things you look into when you buy a laptop.
Then he asked me to Sort different types of memory e.g RAM etc on the basis of speed.In simple words he wanted memory hierarchy .
Technically speaking a processor's registers are the fastest memory a computer has. The size is very small and people generally don't include those numbers when talking about a CPU.
The quickest memory in a computer that would be advertised is the memory that is directly attached to the CPU. It's called cache, and in modern processors you have 3 levels - L1, L2, and L3 - where the first level is the fastest but also the smallest (it's expensive to produce and power). Cache typically ranges from several kilobytes to a few megabytes and is typically made from SRAM.
After that there is RAM. Today's computers use DDR3 for main memory. It's much larger and cheaper than cache, and you'll find sticks upwards of 1 gigabyte in size. The most common type of RAM today is DRAM.
Lastly storage space, such as a hard drive or flash drive, is a form of memory but in general conversation it's grouped separately from the previous types of memory. E.g. you would ask how much "memory" a computer has - meaning RAM - and how much "storage" it has - meaning hard drive space.

CUDA : why are we using so many kinds of memories? [closed]

Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 8 years ago.
Improve this question
I've learned the CUDA programming and I went into some problem. The major one is in CUDA "why do we use so many kinds of memories(Global, local, shared, constant, texture, caches,registers)?" unlike in CPU where we have only three main memory(Ram, caches, hd etc).
The main reasons for having multiple kinds of memory are explained in this article: Wikipedia: Memory Hierarchy
To summarize it, it a very simplified form:
It is usually the case that the larger the memory is, the slower it is
Memory can be read and written faster when it is "closer" to the processor.
As mentioned in the comment: On the CPU, you also have several layers of memory: The main memory, and several levels of caches. These caches are much smaller than main memory, but much faster. These caches are managed by the hardware, so as a software developer, you do not directly notice that these caches exist at all. All the data seems to be in the main memory.
On the GPU, you have to manage this memory manually (althogh in newer CUDA versions, you can also declare the shared memory as "cache", and let CUDA take care of the data management).
For example, reading some data from the shared memory in CUDA may be done within a few NANOseconds. Reading data from global memory may take a few MICROseconds. One of the keys to high performance in CUDA is thus data locality: You should try to keep the data that you are working on in local or shared memory, and avoid reading/writing data in global memory.
(P.S.: The "Close" votes that mark this question as "Primarily Opinion Based" are somewhat ridiculous. The question may show a lack of own research, but is a reasonable question that can clearly be answered here)

Tools for checking memory fragmentation [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 8 years ago.
Improve this question
I have recently read topics about memory fragmentation:
How to solve Memory Fragmentation and What is memory fragmentation?
I want to see some memory allocation map such as author in these article http://pavlovdotnet.wordpress.com/2007/11/10/memory-fragmentation/
Could you recomend some tools to get memory allocation map like that, so I could see if the memory is fragmented and what is the biggest free space available.
I'm on Windows so I would prefer tools working on this system.
Here is a tool that visualizes GC memory and heap usage, also the source code is provided. Another similar app is linked in the comments there as well.
If you need to be able to profile memory usage for a .NET solution, you could check out ANTS Memory Profiler, it can run alongside a project in Visual Studio and keep tabs on how processes and objects are using memory.
There is indirect solution to the problem. I have developing server application for a few years. Initially we are doing the allocation on demand and as a result after a running for few weeks the performance of the server degraded. As a workaround we followed this approach -
Suppose you have user defined classes X,Y,Z, .. which you need to allocate from heap at runtime. Allocate n number of objects X at startup. Put all these objects in free pool list. On demand , take each object of x and provide it to your app. When in use, put it in busy pool list.
When app wants to release it, put it back to the free pool list. Follow this startegy for Y. Z etc.
Since you are allocating all the needed objects at startup and never releasing back to the OS memory manger until your program exits, you will not face the performance degradation caused by memory fragmentation.

Resources