Dask Scheduler Memory - dask

our dask scheduler process seems to balloon in memory as time goes on and executions continue. Currently we see it using 5GB of mem, which seems high since all the data is supposedly living on the worker nodes:
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
31172 atoz 20 0 5486944 5.071g 7100 S 23.8 65.0 92:38.64 dask-scheduler
when starting up the scheduler we would be below 1GB of memory use. Restarting the network doing a client.restart() doesn't seem to help, only a kill of the scheduler process itself and restart will free up the memory.
What is the expected usage of memory per single task executed?
Is the scheduler really only maintaining pointers to which worker contains the future's result?
----edit----
I think my main concern here is why a client.restart() doesn't seem to release the memory being used by the scheduler process. I'm obviously not expecting it to release all memory, but to get back to a base level. We are using client.map to execute our function across a list of different inputs. After executing, doing a client restart over and over and taking snapshots of our scheduler memory we see the following growth:
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
27955 atoz 20 0 670556 507212 13536 R 43.7 6.2 1:23.61 dask-scheduler
27955 atoz 20 0 827308 663772 13536 S 1.7 8.1 16:25.85 dask-scheduler
27955 atoz 20 0 859652 696408 13536 S 4.0 8.5 19:18.04 dask-scheduler
27955 atoz 20 0 1087160 923912 13536 R 62.3 11.3 20:03.15 dask-scheduler
27955 atoz 20 0 1038904 875788 13536 S 3.7 10.7 23:57.07 dask-scheduler
27955 atoz 20 0 1441060 1.163g 12976 S 4.3 14.9 35:54.45 dask-scheduler
27955 atoz 20 0 1646204 1.358g 12976 S 4.3 17.4 37:05.86 dask-scheduler
27955 atoz 20 0 1597652 1.312g 12976 S 4.7 16.8 37:40.13 dask-scheduler
I guess I was just surprised that after doing a client.restart() we don't see the memory usage go back to some baseline.
----further edits----
Some more info about what we're running, since the suggestion was if we were passing in large data structures, to send them directly to the workers.
we send a dictionary as an input for each task, when json dumping the dict, most are under 1000 characters.
---- even further edits: Reproduced issue ----
We reproduced this issue again today. I killed off the scheduler and restarted it, we had about 5.4 GB of free memory, we then ran the function that I'll paste below across 69614 dictionary objects that really hold some file based information (all of our workers are mapped to the same NFS datastore and we are using Dask as a distributed file analysis system.
Here is the function (note: squarewheels4 is a homegrown lazy file extraction and analysis package, it uses Acora and libarchive as its base for getting files out of a compressed archive and indexing the file.)
def get_mrc_failures(file_dict):
from squarewheels4.platforms.ucs.b_series import ChassisTechSupport
from squarewheels4.files.ucs.managed.chassis import CIMCTechSupportFile
import re
dimm_info_re = re.compile(r"(?P<slot>[^\|]+)\|(?P<size>\d+)\|.*\|(?P<pid>\S+)")
return_dict = file_dict
return_dict["return_code"] = "NOT_FILLED_OUT"
filename = "{file_path}{file_sha1}/{file_name}".format(**file_dict)
try:
sw = ChassisTechSupport(filename)
except Exception as e:
return_dict["return_code"] = "SW_LOAD_ERROR"
return_dict["error_msg"] = str(e)
return return_dict
server_dict = {}
cimcs = sw.getlist("CIMC*.tar.gz")
if not cimcs:
return_dict["return_code"] = "NO_CIMCS"
return_dict["keys_list"] = str(sw.getlist("*"))
return return_dict
for cimc in cimcs:
if not isinstance(cimc, CIMCTechSupportFile): continue
cimc_id = cimc.number
server_dict[cimc_id] = {}
# Get MRC file
try:
mrc = cimc["*MrcOut.txt"]
except KeyError:
server_dict[cimc_id]["response_code"] = "NO_MRC"
continue
# see if our end of file marker is there, should look like:
# --- END OF FILE (Done!
whole_mrc = mrc.read().splitlines()
last_10 = whole_mrc[-10:]
eof_line = [l for l in last_10 if b"END OF FILE" in l]
server_dict[cimc_id]["response_code"] = "EOF_FOUND" if eof_line else "EOF_MISSING"
if eof_line:
continue
# get DIMM types
hit_inventory_line = False
dimm_info = []
dimm_error_lines = []
equals_count = 0
for line in whole_mrc:
# regex each line... sigh
if b"DIMM Inventory" in line:
hit_inventory_line = True
if not hit_inventory_line:
continue
if hit_inventory_line and b"=========" in line:
equals_count += 1
if equals_count > 2:
break
continue
if equals_count < 2:
continue
# we're in the dimm section and not out of it yet
line = str(line)
reg = dimm_info_re.match(line)
if not reg:
#bad :/
dimm_error_lines.append(line)
continue
dimm_info.append(reg.groupdict())
server_dict[cimc_id]["dimm_info"] = dimm_info
server_dict[cimc_id]["dimm_error_lines"] = dimm_error_lines
return_dict["return_code"] = "COMPLETED"
return_dict["server_dict"] = server_dict
return return_dict
```
the futures are generated like:
futures = client.map(function_name, file_list)
After in this state my goal was to try and recover and have dask release the memory that it had allocated, here were my efforts:
before cancelling futures:
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
21914 atoz 20 0 6257840 4.883g 2324 S 0.0 62.6 121:21.93 dask-scheduler
atoz#atoz-sched:~$ free -h
total used free shared buff/cache available
Mem: 7.8G 7.1G 248M 9.9M 415M 383M
Swap: 8.0G 4.3G 3.7G
while cancelling futures:
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
21914 atoz 20 0 6258864 5.261g 5144 R 60.0 67.5 122:16.38 dask-scheduler
atoz#atoz-sched:~$ free -h
total used free shared buff/cache available
Mem: 7.8G 7.5G 176M 9.4M 126M 83M
Swap: 8.0G 4.1G 3.9G
after cancelling futures:
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
21914 atoz 20 0 6243760 5.217g 4920 S 0.0 66.9 123:13.80 dask-scheduler
atoz#atoz-sched:~$ free -h
total used free shared buff/cache available
Mem: 7.8G 7.5G 186M 9.4M 132M 96M
Swap: 8.0G 4.1G 3.9G
after doing a client.restart()
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
21914 atoz 20 0 6177424 5.228g 4912 S 2.7 67.1 123:20.04 dask-scheduler
atoz#atoz-sched:~$ free -h
total used free shared buff/cache available
Mem: 7.8G 7.5G 196M 9.4M 136M 107M
Swap: 8.0G 4.0G 4.0G
Regardless of what I ran through the distributed system, my expectation was that after cancelling the futures it would be back to at least close to normal... and after doing a client.restart() we would definitely be near our normal baseline. Am I wrong here?
--- second repro ----
Reproduced the behavior (although not total memory exhaustion) using these steps:
Here's my worker function
def get_fault_list_v2(file_dict):
import libarchive
return_dict = file_dict
filename = "{file_path}{file_sha1}/{file_name}".format(**file_dict)
with libarchive.file_reader(filename) as arc:
for e in arc:
pn = e.pathname
return return_dict
I ran that across 68617 iterations / files
before running we saw this much memory being utilized:
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
12256 atoz 20 0 1345848 1.107g 7972 S 1.7 14.2 47:15.24 dask-scheduler
atoz#atoz-sched:~$ free -h
total used free shared buff/cache available
Mem: 7.8G 3.1G 162M 22M 4.5G 4.3G
Swap: 8.0G 3.8G 4.2G
After running we saw this much:
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
12256 atoz 20 0 2461004 2.133g 8024 S 1.3 27.4 66:41.46 dask-scheduler
After doing a client.restart we saw:
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
12256 atoz 20 0 2462756 2.134g 8144 S 6.6 27.4 66:42.61 dask-scheduler

Generally a task should take up less than a kilobyte on the scheduler. There are a few things you can trip up on that result in storing significantly more, the most common of which is including data within the task graph, which is shown below.
Data included directly in a task graph is stored on the scheduler. This commonly occurs when using large data directly in calls like submit:
Bad
x = np.random.random(1000000) # some large array
future = client.submit(np.add, 1, x) # x gets sent along with the task
Good
x = np.random.random(1000000) # some large array
x = client.scatter(x) # scatter data explicitly to worker, get future back
future = client.submit(np.add, 1, x) # only send along the future
This same principle exists using other APIs as well. For more information, I recommend providing an mcve. It's quite hard to help otherwise.

Related

In Lua, is there a limitation to the number of objects/tables that can be deterministically managed by the garbage collector?

For reference, the Scenarios described below were executed using the following:
Ubuntu 18.04.4 LTS (Bionic Beaver)
top "procps-ng 3.3.12"
Lua 5.3.3 Copyright (C) 1994-2016 Lua.org, PUC-Rio
Each "Scenario" below demonstrates results of the following steps:
Capture "lua5.3" initial process resource usage from "top"
In "lua5.3", allocate ~16GB from the heap
Capture "lua5.3" new process resource usage from "top" to show that 16GB was allocated
In "lua5.3", mark root object for deletion and then call "collectgarbage()"
Capture "lua5.3" final process resource usage from "top" to show effect of "collectgarbage()"
For each scenario, compare the initial process resource usage, "allocated" process resource usage, and the final process resource usage after "collectgarbage(). Scenario 1 shows that the "allocated" 16GB is NOT freed after garbage collection. Scenario 2 and 3 show that some of the "allocated" 16GB is freed after garbage collection. Scenario 4 shows that all of the "allocated" 16GB is freed after garbage collection. Scenarios 1 through 4 differ only in the number of objects (i.e. tables) used during allocation, where Scenario 1 uses the most tables (128*1024), and Scenario 4 uses the least (1). Hence, it appears that the Lua 5.3.3 garbage collection has a limitation on the number of objects it can truly manage, or there is a defect in Lua's memory management.
In short, Lua's garbage collection is not showing deterministic behavior with memory management. If a large number of tables are created, the memory usage associated with those tables may never be returned to the heap as managed by the operating system, when the memory is no longer needed in Lua. However, if a small number of tables are created, the memory usage associated with those tables seem to ALWAYS be returned to the heap as managed by the operating system, when the memory is no longer needed in Lua.
Why is the memory management behavior in Lua not consistent/deterministic?
NOTE: For each scenario, calling collectgarbage("count") after collectgarbage() shows that lua5.3 has properly cleaned up all garbage, whereas using "top" shows that is not always the case. Hence, "top" is used here to show the true behavior of Lua's garbage collection.
Scenario 1: Lua collectgarbage() does not free memory used
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
22866 user 20 0 18152 3140 2840 S 0.0 0.0 0:00.00 lua5.3
> collectgarbage("count")
22.8759765625
> a = {} for i=0,256*1024 do a[i] = {} for j=0,4*1024*1024 do a[i][j] = i*j end end
> collectgarbage("count")
16803927.791016
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
22866 user 20 0 16.053g 0.016t 2868 S 0.0 51.3 0:38.97 lua5.3
> a = nil collectgarbage()
> collectgarbage("count")
25.29296875
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
22866 user 20 0 16.049g 0.016t 2868 S 0.0 51.3 0:39.08 lua5.3
Scenario 2: Lua collectgarbage() frees half of memory used
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
22958 user 20 0 18152 3200 2916 S 0.0 0.0 0:00.00 lua5.3
> collectgarbage("count")
22.8759765625
> a = {} for i=0,128*1024 do a[i] = {} for j=0,8*1024*1024 do a[i][j] = i*j end end
> collectgarbage("count")
16790679.791016
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
22958 user 20 0 16.284g 0.016t 2944 S 0.0 52.0 0:39.79 lua5.3
> a = nil collectgarbage()
> collectgarbage("count")
23.1826171875
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
22958 user 20 0 8422324 8.018g 2944 S 0.0 25.6 0:40.50 lua5.3
Scenario 3: Lua collectgarbage() frees almost all memory used
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
23127 user 20 0 18152 3192 2904 S 0.0 0.0 0:00.00 lua5.3
> collectgarbage("count")
22.8759765625
> a = {} for i=0,64*1024 do a[i] = {} for j=0,16*1024*1024 do a[i][j] = i*j end end
> collectgarbage("count")
16784151.791016
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
23127 user 20 0 16.275g 0.016t 2932 S 0.0 52.0 0:41.22 lua5.3
> a = nil collectgarbage()
> collectgarbage("count")
23.1826171875
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
23127 user 20 0 25900 10992 2932 S 0.0 0.0 0:41.81 lua5.3
Scenario 4: Lua collectgarbage() frees all memory used
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
23771 user 20 0 18152 3204 2920 S 0.0 0.0 0:00.00 lua5.3
> collectgarbage("count")
22.8759765625
> a = {} for i=0,0 do a[i] = {} for j=0,1024*1024*1024 do a[i][j] = i*j end end
> collectgarbage("count")
16777241.969727
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
23771 user 20 0 16.017g 0.016t 2948 S 0.0 51.2 0:38.97 lua5.3
> a = nil collectgarbage()
> collectgarbage("count")
23.17578125
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
23771 user 20 0 18152 3364 2948 S 0.0 0.0 0:40.80 lua5.3
Update
After additional testing, it does appear that the number of objects (i.e. tables) managed in Lua affects the behavior of Lua's garbage collection. Specifically, if the number of tables being managed by Lua is less than 256 (this seems to be the threshold), the memory will be returned to the operating system's heap after garbage collection. As the number of objects grows above 256 (inclusive), Lua's memory management begins to exhibit different behavior as it holds on to more memory for internal reuse after its garbage collection. Scenario 1 above shows the most dramatic affect of Lua holding onto memory after garbage collection, and Scenario 4 shows the least affect.
Note: The Example results are not deterministic; that is, other execution in Lua may affect the behavior of Lua's garbage collection. For the most deterministic results, launch a new Lua process for each test.
For illustration, the following examples continuously allocate, populate, and deallocate a table of tables, printing out the following from the Lua process: iteration, preMemoryUsage (after allocation), preMemoryUsagePercent, postMemoryUsage (after deallocation), postMemoryUsagePercent.
Example Program: Change numberOfTables as Desired
numberOfTables = 255 -- Example 1: 255; Example 2: 256
function getMemoryPercent()
handle = io.popen("v=`ps -p $PPID -o pmem= | tr -d '\n' | sed 's/ *//g'`; echo -n $v")
memoryPercent = handle:read("*a")
handle:close()
return memoryPercent
end
socket = require("socket")
print("#", "preMemUsage", "%", "postMemUsage", "%")
iteration = 0
while true
do
a = {}
for i = 1,numberOfTables
do
a[i] = {}
for j = 1,math.floor(256*1024*1024/numberOfTables)
do
a[i][j] = i*j
end
end
preMemoryUsage = collectgarbage("count")
preMemoryUsagePercent = getMemoryPercent()
a = nil
collectgarbage()
postMemoryUsage = collectgarbage("count")
postMemoryUsagePercent = getMemoryPercent()
iteration = iteration + 1
print(iteration, preMemoryUsage, preMemoryUsagePercent, postMemoryUsage, postMemoryUsagePercent)
end
Example 1: Results for numberOfTables=255
# preMemUsage % postMemUsage %
1 8355905.0869141 25.4 47.240234375 0.0
2 8355905.1972656 25.4 47.322265625 0.0
3 8355905.1972656 25.4 47.322265625 0.0
4 8355905.1972656 25.4 47.322265625 0.0
5 8355905.1972656 25.4 47.322265625 0.0
6 8355905.1972656 25.4 47.322265625 0.0
7 8355905.1972656 25.4 47.322265625 0.0
8 8355905.1972656 25.4 47.322265625 0.0
9 8355905.1972656 25.4 47.322265625 0.0
10 8355905.1972656 25.4 47.322265625 0.0
Example 2: Results for numberOfTables=256
# preMemUsage % postMemUsage %
1 4194369.0205078 12.8 47.119140625 0.0
2 4194369.1308594 12.8 47.201171875 12.8
3 4194369.1035156 12.8 47.173828125 12.5
4 4194369.1318359 12.8 47.2021484375 12.4
5 4194369.1318359 12.8 47.2021484375 12.4
6 4194369.1318359 12.8 47.2021484375 12.4
7 4194369.1318359 12.8 47.2021484375 12.4
8 4194369.1318359 12.8 47.2021484375 12.4
9 4194369.1318359 12.8 47.2021484375 12.4
10 4194369.1318359 12.8 47.2021484375 12.4

How to change the thread priority in linux OS which is built using yocto project?

I am using linux os (4.9.126) built with yocto (sumo). It consists of glibc v2.27, util-linux v2.32.1.
The pthread library is located at:
ldconfig -p | grep pthread.so
libpthread.so.0 (libc6,x86-64, OS ABI: Linux 3.2.0) => /lib/libpthread.so.0
I want to set the priority SCHD_RR inside docker containerized C++ application.
..
..
...
int set_max_thread_priority()
{
int policy;
struct sched_param param {
};
pthread_getschedparam(pthread_self(), &policy, &param);
policy = SCHED_RR;
param.sched_priority = sched_get_priority_max(policy);
return pthread_setschedparam(pthread_self(), policy, &param);
}
...
I want to achieve
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
2083 root rt 0 709812 8100 6456 S 44.4 0.2 0:32.24 my-application
But it shows,
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
2083 root 20 0 709812 8100 6456 S 44.4 0.2 0:32.24 my-application
When I run docker containerized my-application on Ubuntu/Debian, then it works fine.
It does not work with yocto os. When I try to change the thread priority on yocto os (using chrt command) it shows following error.
sudo chrt -r -p 99 2083
chrt: failed to set pid 18465's policy: Operation not permitted
I am not sure if I am missing any package inside OS.
Why don't you use cgroups and create rules for your application's priority.

Why does docker see the container is hitting the rss limit?

I'm trying to understand why the limits have decided a task needs to be killed, and how it's doing the accounting. When my GCE Docker container kills a process, it shows something like:
Task in /404daacfcf6b9e55f71b3d7cac358f0dc921a2d580eed460c2826aea8e43f05e killed as a result of limit of /404daacfcf6b9e55f71b3d7cac358f0dc921a2d580eed460c2826aea8e43f05e
memory: usage 2097152kB, limit 2097152kB, failcnt 74571
memory+swap: usage 0kB, limit 18014398509481983kB, failcnt 0
kmem: usage 0kB, limit 18014398509481983kB, failcnt 0
Memory cgroup stats for /404daacfcf6b9e55f71b3d7cac358f0dc921a2d580eed460c2826aea8e43f05e: cache:368KB rss:2096784KB rss_huge:0KB mapped_file:0KB writeback:0KB inactive_anon:16KB active_anon:2097040KB inactive_file:60KB active_file:36KB unevictable:0KB
[ pid ] uid tgid total_vm rss nr_ptes swapents oom_score_adj name
[ 4343] 0 4343 5440 65 15 0 0 bash
[ 4421] 0 4421 265895 6702 77 0 0 npm
[ 4422] 0 4422 12446 2988 28 0 0 gunicorn
[ 4557] 0 4557 739241 346035 1048 0 0 gunicorn
[ 4560] 0 4560 1086 24 8 0 0 sh
[ 4561] 0 4561 5466 103 15 0 0 bash
[14594] 0 14594 387558 168790 672 0 0 node
Memory cgroup out of memory: Kill process 4557 (gunicorn) score 662 or sacrifice child
Killed process 4557 (gunicorn) total-vm:2956964kB, anon-rss:1384140kB, file-rss:0kB
Supposedly the memory hit a 2GB usage limit, and something needs to die. According to the cgroup stats, I appear to have 2GB of usage in active_anon and rss.
When I look at the table of process stats, I don't see where the 2GB is:
For rss, I see the two major processes 346035 + 168790 = 514MB?
For total_vm, I see three major processes 265895 + 739241 + 387558 = 1.4GB?
But when it decides to kill the gunicorn process, it says it had 3GB of Total VM and 1.4GB of Anon RSS. I don't see how this follows from the above numbers at all...
For most of it's life, according to top, the gunicorn process appears to hum along with 555m RES and 2131m VIRT and 22% MEM * 2.5GB box = 550MB of memory usage. (I haven't yet been able to time it properly to peek at top values at the time it dies...)
Can someone help me understand this?
Under what accounting, do these sum to 2GB of usage? (virtual? rss? something else?)
Is there something else besides top/ps I should use to track how much memory a process is using for the purposes of docker's killing it?
From what I know, the total_vm and rss are counted in 4kB (refer to: https://stackoverflow.com/a/43611576), instead of kB.
So for pid<4557>:
rss=346035, means anon-rss:1384140kB (=346035*4kB)
total_vm=739241, means total-vm:2956964kB(=739241*4kB)
This will explain your mem usage very well.

Docker user executing a process cannot be removed

I have some resources used by a specific user that I had to delete because it was taking a lot of resources from the server. When I listed the processes in the server the deleted user now shows as “1001” instead of the name it used to show before I deleted it.
%Cpu(s): 19.8 us, 29.5 sy, 0.0 ni, 50.7 id, 0.0 wa, 0.0 hi, 0.1 si, 0.0 st
KiB Mem : 3882456 total, 183568 free, 2003808 used, 1695080 buff/cache
KiB Swap: 1679356 total, 1155300 free, 524056 used. 1463480 avail Mem
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
9192 1001 20 0 2068436 74700 10284 S 0.3 1.9 3:02.86 node
By using systemd status I found the process and the docker container ID that the user is executing is in.
├─docker
│ ├─42b40e73687acb7fcd9a0e43372ced7588b5568c942f740d06510ab0e85b1462
│ │ ├─17156 /bin/sh -e /usr/local/sbin/start.sh
└─11148 node --debug --nolazy dist-release/server
So I went into the container and I look to the start.sh file but it’s just an executable file, there’s no indication inside of the file that the user is getting called inside of the executable file.
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
apiassets_1
42b40e73687a local.io/api-statements:development "start.sh" 21 hours ago Up 18 hours 0.0.0.0:32785->3000/tcp, 0.0.0.0:5966->5858/tcp
What I want to do is stopping this user to use this resources, so I was just curious how can I either find how this user is calling this script to stop it or how can I stop it.

Cassandra eats memory

I have Cassandra 2.1 and following properties set:
MAX_HEAP_SIZE="5G"
HEAP_NEWSIZE="800M"
memtable_allocation_type: heap_buffers
top utility shows that cassandra eats 14.6G virtual memory:
KiB Mem: 16433148 total, 16276592 used, 156556 free, 22920 buffers
KiB Swap: 16777212 total, 0 used, 16777212 free. 9295960 cached Mem
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
23120 cassand+ 20 0 14.653g 5.475g 29132 S 318.8 34.9 27:07.43 java
It also dies with various OutOfMemoryError exceptions when I am accessing it from Spark.
How I can prevent this "OutOfMemoryErrors" and reduce memory usage?
Cassandra do eat to much memory but it can be controlled but tuning the GC [Garbage Collection] setting.
GC parameters are contained in the bin/cassandra.in.sh file in the JAVA_OPTS variable.
you can apply these settings in JAVA_OPTS
-XX:+UseConcMarkSweepGC
-XX:ParallelCMSThreads=1
-XX:+CMSIncrementalMode
-XX:+CMSIncrementalPacing
-XX:CMSIncrementalDutyCycleMin=0
-XX:CMSIncrementalDutyCycle=10
Or instead of specifying MAX_HEAP_SIZE and HEAP_NEWSIZE these parameter let cassandra'script specify these parameter Because it will assign best values for these parameter.

Resources