You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: h2o-docs/source/faq/hadoop_related.rst
+44-31
Original file line number
Diff line number
Diff line change
@@ -43,28 +43,35 @@ description of your Hadoop environment, including the Hadoop distribution and ve
43
43
44
44
Common Hadoop Questions
45
45
""""""""""""""""""""""""
46
-
*What's the syntax for the file path of a data set sitting in hdfs?*
46
+
**What's the syntax for the file path of a data set sitting in hdfs?**
47
47
48
48
To locate an HDFS file, go to **Data > Import** and enter **hdfs://** in the **path** field. H\ :sub:`2`\ O automatically detects any HDFS paths. This is a good way to verify the path to your data set before importing through R or any other non-web API.
49
49
50
-
*When interacting with an H2O cluster launched on multiple Hadoop nodes, is it necessary for R to be installed on all the data nodes?*
50
+
---
51
+
52
+
**When interacting with an H\ :sub:`2`\ O cluster launched on multiple Hadoop nodes, is it necessary for R to be installed on all the data nodes?**
51
53
52
54
No - as long as the R instance can communicate with one of the nodes in the network, R can be installed on any of the nodes, or even on a local machine that will securely tunnel into the cluster.
53
55
54
-
*Is it possible to launch the H2O cluster on Hadoop nodes using R’s h2o.init() command?*
56
+
---
57
+
58
+
**Is it possible to launch the H\ :sub:`2`\ O cluster on Hadoop nodes using R’s** `h2o.init()` **command?**
59
+
60
+
No - follow the instructions in :ref:`Hadoop_Tutorial` and add the IP address to the `h2o.init()` function to connect to the cluster.
55
61
56
-
No - follow the instructions in :ref:`Hadoop_Tutorial` and add the IP address to the **h2o.init()** function to connect to the cluster.
Each mapper task gets its own output directory in HDFS. To prevent overwriting multiple users' files, each mapper task must have a unique output directory name. Change
61
-
the **-output hdfsOutputDir** argument to **-output hdfsOutputDir1** and the task should launch.
67
+
the `-output hdfsOutputDir` argument to `-output hdfsOutputDir1` and the task should launch.
62
68
63
-
*What should I do if H2O starts to launch but times out in 120 seconds?*
69
+
---
64
70
71
+
**What should I do if H\ :sub:`2`\ O starts to launch but times out in 120 seconds?**
65
72
66
-
1.YARN or MapReduce's configuration is not configured correctly. Enable launching for mapper tasks of specified memory sizes. If YARN only allows mapper tasks with a maximum memory size of 1g and the request requires 2g, then the request will timeout at the default of 120 seconds. Read `Configuration Setup <http://hortonworks.com/blog/how-to-plan-and-configure-yarn-in-hdp-2-0/>`_ to make sure your setup will run.
67
73
74
+
1. YARN or MapReduce's configuration is not configured correctly. Enable launching for mapper tasks of specified memory sizes. If YARN only allows mapper tasks with a maximum memory size of 1g and the request requires 2g, then the request will timeout at the default of 120 seconds. Read `Configuration Setup <http://hortonworks.com/blog/how-to-plan-and-configure-yarn-in-hdp-2-0/>`_ to make sure your setup will run.
68
75
69
76
2. The nodes are not communicating with each other. If you request a cluster of two nodes and the output shows a stall in reporting the other nodes and forming a cluster (as shown in the following example), check that the security settings for the network connection between the two nodes are not preventing the nodes from communicating with each other. You should also check to make sure that the flatfile that is generated and being passed has the correct home address; if there are multiple local IP addresses, this could be an issue.
70
77
@@ -94,17 +101,21 @@ the **-output hdfsOutputDir** argument to **-output hdfsOutputDir1** and the tas
*What should I do if the H2O job launches but terminates after 600 seconds?*
107
+
**What should I do if the H2O job launches but terminates after 600 seconds?**
99
108
100
109
The likely cause is a driver mismatch - check to make sure the Hadoop distribution matches the driver jar file used to launch H\ :sub:`2`\ O. If your distribution is not currently
101
110
available in the package, `email us <support@0xdata>`_ for a new driver file.
102
111
103
-
*What should I do if I want to create a job with a bigger heap size but YARN doesn't launch and H2O times out?*
112
+
---
104
113
105
-
First, try the job again but with a smaller heap size (**-mapperXmx**) and a smaller number of nodes (**-nodes**) to verify that a small launch can proceed at all.
114
+
**What should I do if I want to create a job with a bigger heap size but YARN doesn't launch and H\ :sub:`2`\ O times out?**
106
115
107
-
If the cluster manager settings are configured for the default maximum memory size but the memory required for the request exceeds that amount, YARN will not launch and H2O will time out.
116
+
First, try the job again but with a smaller heap size (`-mapperXmx`) and a smaller number of nodes (`-nodes`) to verify that a small launch can proceed at all.
117
+
118
+
If the cluster manager settings are configured for the default maximum memory size but the memory required for the request exceeds that amount, YARN will not launch and H\ :sub:`2`\ O will time out.
108
119
If you have a default configuration, change the configuration settings in your cluster manager to enable launching of mapper tasks for specific memory sizes. Use the following formula to calculate the amount of memory required:
109
120
110
121
::
@@ -134,51 +145,53 @@ Output from an H2O launch is shown below:
134
145
mapreduce.map.memory.mb: 36864
135
146
136
147
137
-
**mapreduce.map.memory.mb** must be less than the YARN memory configuration values for the launch to succeed. See the examples below for how to change the memory configuration values for your version of Hadoop.
148
+
`mapreduce.map.memory.mb` must be less than the YARN memory configuration values for the launch to succeed. See the examples below for how to change the memory configuration values for your version of Hadoop.
138
149
139
150
140
151
**For Cloudera, configure the settings in Cloudera Manager. Depending on how the cluster is configured, you may need to change the settings for more than one role group.**
141
152
142
-
1. Click **Configuration** and enter the following search term in quotes: **yarn.nodemanager.resource.memory-mb**.
153
+
1. Click **Configuration** and enter the following search term in quotes: **yarn.nodemanager.resource.memory-mb**.
143
154
144
-
2. Enter the amount of memory (in GB) to allocate in the **Value** field. If more than one group is listed, change the values for all listed groups.
155
+
2. Enter the amount of memory (in GB) to allocate in the **Value** field. If more than one group is listed, change the values for all listed groups.
For Hortonworks, `configure <http://docs.hortonworks.com/HDPDocuments/Ambari-1.6.0.0/bk_Monitoring_Hadoop_Book/content/monitor-chap2-3-3_2x.html>`_ the settings in Ambari.
158
-
1. Select **YARN**, then click the **Configs** tab.
159
-
2. Select the group.
160
-
3. In the **Node Manager** section, enter the amount of memory (in MB) to allocate in the **yarn.nodemanager.resource.memory-mb** entry field.
168
+
**For Hortonworks,** `configure <http://docs.hortonworks.com/HDPDocuments/Ambari-1.6.0.0/bk_Monitoring_Hadoop_Book/content/monitor-chap2-3-3_2x.html>`_ **the settings in Ambari.**
169
+
170
+
1. Select **YARN**, then click the **Configs** tab.
171
+
2. Select the group.
172
+
3. In the **Node Manager** section, enter the amount of memory (in MB) to allocate in the **yarn.nodemanager.resource.memory-mb** entry field.
161
173
162
174
.. image:: TroubleshootingHadoopAmbariNodeMgr.png
163
175
:width:100 %
164
176
165
-
4. In the **Scheduler** section, enter the amount of memory (in MB)to allocate in the **yarn.scheduler.maximum-allocation-mb** entry field.
177
+
4. In the **Scheduler** section, enter the amount of memory (in MB)to allocate in the **yarn.scheduler.maximum-allocation-mb** entry field.
0 commit comments