I use Hadoop 2.0.5 (Alpha) to do relatively large jobs, and I came across these errors:
The container [pid = 15023, containerID = container_1378641992707_0002_01_000029] going beyond the limits of virtual memory. Current usage: 492.4 MB 1 GB used physical memory; 3.3 GB virtual memory used 2.1 GB . Killing container.
Then I found out about these two parameters:
property yarn.nodemanager.vmem-pmem-ratio , which is set to 2.1 by default.
yarn.app.mapreduce.am.command-opts , which by default is set to -Xmx1024mb (= 1 GB).
This explains the limits noted above.
Setting these parameters to a larger value helped, but then I found this parameter: yarn.app.mapreduce.am.resource.mb , which is set to 1536 by default.
And I canβt tell the difference between 3 of the description given in the Hadoop XML standards, and how I should install them correctly for optimization purposes.
An explanation or a good recommendation would be very helpful.
mapreduce hadoop yarn
itzhaki
source share