Skip to content

Latest commit

 

History

History

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

安装

配置hosts

vim /etc/hosts C:\Windows\System32\drivers\etc\hosts

JDK

安装

hadoop

  • namenode: IP1
  • secondary-namenode: IP2
  • resourcemanager: IP3
  • jobhistory: IP1
  1. 解压
  2. 到根目录
  3. 配置
    1. etc/hadoop/core-site.xml
    <configuration>
        <property>
            <name>fs.defaultFS</name>
            <value>hdfs://IP1:8020</value>
        </property>
        <!-- 数据路径 -->
        <property>
            <name>hadoop.tmp.dir</name>
            <value>/opt/module/hadoop/data</value>
        </property>
        <property>
            <name>hadoop.http.staticuser.user</name>
            <value>root</value>
        </property>
    </configuration>
    1. etc/hadoop/hdfs-site.xml
     <configuration>
         <!-- namenode -->
         <property>
             <name>dfs.namenode.http-address</name>
             <value>IP1:9870</value>
         </property>
         <!-- secondary-namenode -->
         <property>
             <name>dfs.namenode.secondary.http-address</name>
             <value>IP2:9870</value>
         </property>
     </configuration>   
    1. etc/hadoop/yarn-site.xml
     <configuration>
         <property>
             <name>yarn.nodemanager.aux-services</name>
             <value>mapreduce_shuffle</value>
         </property>
         <property>
             <name>yarn.resourcemanager.hostname</name>
             <value>IP3</value>
         </property>
         <!-- 日志聚集 -->
         <property>
             <name>yarn.log-aggregation-enable</name>
             <value>true</value>
         </property>
         <property>
             <name>yarn.log.server.url</name>
             <value>http://IP1:19888/jobhistory/logs</value>
         </property>
         <property>
             <name>yarn.log-aggregation.retain-seconds</name>
             <value>259200</value>
         </property>   
     </configuration>   
    1. etc/hadoop/mapred-site.xml
     <configuration>
         <!-- 管理方式 -->
         <property>
             <name>mapreduce.framework.name</name>
             <value>yarn</value>
         </property>
         <!-- 历史服务器 -->
         <property>
             <name>mapreduce.jobhistory.address</name>
             <value>IP1:10020</value>
         </property>
         <property>
             <name>mapreduce.jobhistory.webapp.address</name>
             <value>IP1:19888</value>
         </property>
         <!-- 任务运行环境 -->
         <property>
             <name>yarn.app.mapreduce.am.env</name>
             <value>HADOOP_MAPRED_HOME=/opt/module/hadoop/</value>
         </property>
         <property>
             <name>mapreduce.map.env</name>
             <value>HADOOP_MAPRED_HOME=/opt/module/hadoop/</value>
         </property>
         <property>
             <name>mapreduce.reduce.env</name>
             <value>HADOOP_MAPRED_HOME=/opt/module/hadoop/</value>
         </property>
     </configuration>   
    1. etc/hadoop/workers
    IP1
    IP2
    IP3
    
  4. 初始化(仅在namenode执行) hdfs namenode -format

其他

启动HDFS(IP1) sbin/start-dfs.sh 启动YARN(IP3) sbin/start-yarn.sh 历史服务器(IP1) bin/mapred --daemon start historyserver 执行jar hadoop jar xxx.jar xxx.YYY ...args 网页地址: http://IP1:9870/

报错说明

ERROR: Attempting to operate on hdfs namenode as root

无ssh访问权限导致。

  1. 生成: ssh-keygen -t rsa
  2. 分发: ssh-copy-id IP
  3. 配置环境: vim /etc/profile
    export HDFS_NAMENODE_USER=root
    export HDFS_DATANODE_USER=root
    export HDFS_SECONDARYNAMENODE_USER=root
    export YARN_RESOURCEMANAGER_USER=root
    export YARN_NODEMANAGER_USER=root
    
  4. 应用配置 source /etc/profile

ERROR: JAVA_HOME is not set and could not be found.

JAVA配置未生效。编辑配置: vim etc/hadoop/hadoop-env.sh

export JAVA_HOME=

IP1:9870 页面报错Failed to retrieve data from /webhdfs/v1

JDK版本过高。降级为JDK8即可。

FileNotFoundException: HADOOP_HOME and hadoop.home.dir are unset.

Windows无Hadoop环境。

  1. 下载环境包 https://github.com/cdarlint/winutils
  2. 配置环境变量 HADOOP_HOME
  3. (不确定是否必须) 复制 hadoop.dllC:\Windows\System32
  4. 重启电脑
  5. (每次)启动 winutils.exe