in Education by
I would like to know the relation between the mapreduce.map.memory.mb and mapred.map.child.java.opts parameters. Is mapreduce.map.memory.mb > mapred.map.child.java.opts? Select the correct answer from above options

1 Answer

0 votes
by
 
Best answer
mapreduce.map.memory.mb is the physical memory for your map process produced by YARN container. While mapred.map.child.java.opts is the JVM heap size for your map and process. Most common errors that we get nowadays occurs when we run any MapReduce job: Application application_1409135750325_48141 failed 2 times due to AM Container for appattempt_1409135750325_48141_000002 exited with exitCode: 143 due to: Container [pid=4733,containerID=container_1409135750325_48141_02_000001] is running beyond physical memory limits. Current usage: 2.0 GB of 2 GB physical memory used; 6.0 GB of 4.2 GB virtual memory used. Killing container. YARN monitors memory of your running containers. In MapReduce container is either map or reduce process. Whenever the allocated memory of any mapper process exceeds the default memory limit. Hadoop kills the mapper while giving the error: Container[pid=container_1406552545451_0009_01_000002,containerID=container_234132_0001_01_000001] is running beyond physical memory limits. Current usage: 569.1 MB of 512 MB physical memory used; 970.1 MB of 1.0 GB virtual memory used. Killing container. So to overcome these problems increment in the memory available to your MapReduce job is done. Here, we have two memory settings that needs to be configured at the same time: The physical memory for your YARN map and reduce processes(mapreduce.map.memory.mb and mapreduce.reduce.memory.mb) The JVM heap size for your map and reduce processes (mapreduce.map.java.opts and mapreduce.reduce.java.opts) Here, we set the YARN container physical memory limits for your map and reduce processes by configuring mapreduce.map.memory.mb and mapreduce.reduce.memory.mb, respectively. For example, if you want to limit your map process and reduce process to 2GB and 4GB, respectively and you want to make this the default limit in your cluster, then you have to set the mapred-site.xml in the following way: mapreduce.map.memory.mb 2048 mapreduce.reduce.memory.mb 4096 The physical memory configured for your job must fall within the minimum and maximum memory allowed for containers in your cluster. Now, just after configuring your physical memory of map and reduce processes, you need to configure the JVM heap size for your map and reduce processes. The sizes of these processes needs to be less than the physical memory you configured in the previous section. As a general rule, they should be 80% the size of the YARN physical memory settings. To set the map and reduce heap size you need to configure mapreduce.map.java.opts and mapreduce.reduce.java.opts respectively. Now while continuing with the previous section example, we’ll arrive at our Java heap sizes by taking the 2GB and 4GB physical memory limits and multiple by 0.8 to. The changes will be in mapred-site.xml as shown below(assuming you wanted these to be the defaults for your cluster): mapreduce.map.java.opts -Xmx1638m mapreduce.reduce.java.opts -Xmx3278m If you want more information regarding the same, refer to the following link:

Related questions

0 votes
    Why are there two separate packages map-reduce package in Apache's hadoop package tree: org.apache.hadoop.mapred ... separated out? Select the correct answer from above options...
asked Jan 26, 2022 in Education by JackTerrance
0 votes
    I am having some trouble writing the code for wordcount program using MapReduce, I could use some help. Select the correct answer from above options...
asked Jan 21, 2022 in Education by JackTerrance
0 votes
    Hi, I am using MapReduce and there are so many steps in it. I want the value of the last reduced step as ... . Thank you in advance. Select the correct answer from above options...
asked Jan 21, 2022 in Education by JackTerrance
0 votes
    So, I've been looking at Hadoop with keen interest, and to be honest I'm fascinated, things don't get much ... anyone know of one? Select the correct answer from above options...
asked Feb 2, 2022 in Education by JackTerrance
0 votes
    Is spark dependent on Hadoop? If not, then I can run Spark without Hadoop right? Will I miss any features if I do Select the correct answer from above options...
asked Jan 21, 2022 in Education by JackTerrance
0 votes
    The Hadoop list includes the HBase database, the Apache Mahout ________ system, and matrix operations. ... recognition 3. Statistical classification 4. Artificial intelligence...
asked Dec 7, 2022 in Education by JackTerrance
0 votes
    I do not understand the differences between using S3 and S3n with my Hadoop cluster, can someone explain? Select the correct answer from above options...
asked Jan 26, 2022 in Education by JackTerrance
0 votes
    Are they same? In my view "hadoop fs" command shows the hdfs files and the "hdfs dfs" command shows the local files, why is that? Select the correct answer from above options...
asked Jan 21, 2022 in Education by JackTerrance
0 votes
    I tried reading from various sources but I am still not very clear about their difference. What majorly seperates Mongodb and Hadoop? Select the correct answer from above options...
asked Jan 21, 2022 in Education by JackTerrance
0 votes
    Can someone tell me what is the basic difference between HBase and Hadoop? I have done my own research but I ... in Layman terms. Select the correct answer from above options...
asked Jan 20, 2022 in Education by JackTerrance
0 votes
    I am trying to understand how spark runs on YARN cluster/client. I have the following question in my mind. Is ... the job to cluster? Select the correct answer from above options...
asked Jan 28, 2022 in Education by JackTerrance
0 votes
    Suppose you have designed a Big Data batch using the MapReduce framework. Now you want to execute it on a cluster ... Run view? Name Node Data Node Resource Manager Job Tracker...
asked Mar 23, 2021 in Technology by JackTerrance
0 votes
    Is there any way or any command which I can use in command prompt to know the version of Hadoop? Also, how ... the version of Hive? Select the correct answer from above options...
asked Jan 21, 2022 in Education by JackTerrance
0 votes
    What is the difference between Hadoop, HBase, Hive and Pig? I know the basic Definitions of all these terms, But ... can these be used? Select the correct answer from above options...
asked Jan 20, 2022 in Education by JackTerrance
0 votes
    Sun also has the Hadoop Live CD ________ project, which allows running a fully functional Hadoop cluster using a live CD. 1. OpenOffice.org 2. OpenSolaris 3. GNU 4. Linux...
asked Dec 6, 2022 in Education by JackTerrance
...