Out of memory errors when executing Jest on Ubuntu - node.js

I'm trying to execute Jest on Ubuntu 14.04.02, in a virtual machine with 4gb of RAM. node version 0.12.2, npm 2.0.0-alpha-5
free shows me:
total used free shared buffers cached
Mem: 3.8G 199M 3.6G 976K 1.1M 18M
When I run npm test, I keep getting a variety of out of memory errors:
Error: FATAL ERROR: CALL_AND_RETRY_LAST Allocation failed - process out of memory
FATAL ERROR: Committing semi space failed. Allocation failed - process out of memory
# Fatal error in ../deps/v8/src/heap/store-buffer.cc, line 132
# CHECK(old_virtual_memory_->Commit(reinterpret_cast<void*>(old_limit_), grow * kPointerSize, false)) failed
Any idea what the minimum memory requirement is...or if I have misconfiguration something that is leading to this?

It turns out downgrading to node version 0.10.32, installed via npm, healed the issue.

Related

How to reduce I/O and memory usage when installing node.js on shared hosting?

I've tried to install node.js via Putty on my shared hosting account with cPanel and CloudLinux. But at some moment I/O and physical memory usage reached to their limits and the installation process was stopped. My I/O usage limit is 10 MB and physical memory limit is 512 MB.
This happens when Putty displays the lines:
-c -o /home/vikna211/node/out/Release/obj.target/v8_base/deps/v8/src/api.o
../deps/v8/src/api.cc
After that I see:
make[1]:
[/home/vikna211/node/out/Release/obj.target/v8_base/deps/v8/src/api.o]
Interrupt make[1]: Deleting intermediate file
`4095d8cbfa2eff613349def330937d91ee5aa9c9.intermediate' make: [node]
Interrupt
Is it possible to reduce the usage of both resources when installing node.js to successfully finish the process?
And maybe it's not a problem of memory. Maybe the process tries to delete that intermediate file, but can't do it and causes the memory crash.

i have just begun to use android studio and i cant seem to get my gradle to sync with my application. here is what it shows :

7:46:20 PM Gradle sync started
7:46:35 PM Gradle sync failed: Unable to start the daemon process.
This problem might be caused by incorrect configuration of the daemon.
For example, an unrecognized jvm option is used.
Please refer to the user guide chapter on the daemon at https://docs.gradle.org/2.10/userguide/gradle_daemon.html
Please read the following process output to find out more:
-----------------------
Error occurred during initialization of VM
Could not reserve enough space for object heap
Error: Could not create the Java Virtual Machine.
Error: A fatal exception has occurred. Program will exit.
Consult IDE log for more details (Help | Show Log)
the jvm version is 1.7.0_79
and the studio version is 2.1.1
Error occurred during initialization of VM Could not reserve enough space for object heap Error: Could not create the Java Virtual Machine.
There's no space available in RAM. To fix go to /android-studio-dir/bin and edit studio.vmoptions and studio64.vmoptions to increment the -Xmx and to reserve more memory to Java. Note that the number of processes active may influence on that.
Probably, the /tmp location is full..
Found this somewhere..
Use df command
df
You should see an output with a line like this:
tmpfs 102400 102312 88 100% /tmp
So to change the size of the tmp file:
sudo mount -o remount,size=2G /tmp
Done! Now, It should work..

Starting node server with increased memory

On server startup (using command supervisor ./bin/www) I'm exporting a huge data from mongodb to redis and getting an error:
FATAL ERROR: CALL_AND_RETRY_LAST Allocation failed - process out of memory
However, if I start the command using node --max-old-space-size=4076 ./bin/www it works fine.
How to configure supervisor to start node with 4gb memory?

Spark - UbuntuVM - insufficient memory for the Java Runtime Environment

I'm trying to install Spark1.5.1 on Ubuntu14.04 VM. After un-tarring the file, I changed the directory to the extracted folder and executed the command "./bin/pyspark" which should fire up the pyspark shell. But I got an error message as follows:
[ OpenJDK 64-Bit Server VM warning: INFO: os::commit_memory(0x00000000c5550000, 715849728, 0) failed;
error='Cannot allocate memory' (errno=12) There is insufficient
memory for the Java Runtime Environment to continue.
Native memory allocation (malloc) failed to allocate 715849728 bytes
for committing reserved memory.
An error report file with more information is saved as:
/home/datascience/spark-1.5.1-bin-hadoop2.6/hs_err_pid2750.log ]
Could anyone please give me some directions to sort out the problem?
We need to set spark.executor.memory in conf/spark-defaults.conf file to a value specific to your machine. For example,
usr1#host:~/spark-1.6.1$ cp conf/spark-defaults.conf.template conf/spark-defaults.conf
nano conf/spark-defaults.conf
spark.driver.memory 512m
For more information, refer to the official documentation: http://spark.apache.org/docs/latest/configuration.html
Pretty much what it says. It wants 7GB of RAM. So give the VM ~ 8GB of RAM.

FATAL ERROR: JS Allocation failed - process out of memory Aborted (core dumped) - Node.js

I have been getting 'FATAL ERROR: JS Allocation failed - process out of memory
Aborted (core dumped)' error while running a nodejs process although I am using the command :
node --max-old-space-size=8192 run.js
I am using v10.25.
The code simply downloads data(size=2gb) from aws s3 and there are some data manipulation associated.
Why would nodejs run out of memory?
How to run this nodejs process without the fatal error?
Any help is appreciated.
Edit 1:-
inspecting with
console.log(util.inspect(process.memoryUsage()));
Just before crashing gives this:-
{ rss: 1351979008, heapTotal: 1089684736, heapUsed: 1069900560 }
For anyone who is facing this issue.
I installed nodejs v12.02 to use --max-old-space-size=8192.
It was not working in v10.25.

Resources