最小系统下安装hadoop

1、创建haddop用户,并设置密码

useradd -m hadoop -s /bin/bash

passwd hadoop

 

2、hadoop用户添加管理员权限

visudo

找到 root ALL=(ALL) ALL 并在下面添加

hadoop ALL=(ALL) ALL

 

3、退出当前登陆并登陆hadoop用户

exit

 

4、准备工作

(1) 关闭防火墙

systemctl stop firewalld

systemctl disable firewalld

systemctl status firewalld

 

vi /etc/sysconfig/selinux

设置SELINUX=disabled

 

(2) 设置hostname

hostnamectl set-hostname huatec37

vi /etc/hosts

结尾添加 ip地址   huatec37

 

(3) 安装软件

ifconfig:yum install net-tools

自动补全:yum install bash

vim:yum -y install vim

wget:yum -y install wget

hadoop安装包: wget 

http://mirrors.hust.edu.cn/apache/hadoop/core/hadoop-2.7.6/hadoop-2.7.6.tar.gz

(先进入目录 /usr/local 再执行命令)

ssh: sudo yum install openssh-clients

sudo yum install openssh-server

测试命令是否可用,输入yes和密码

ssh localhost

 

配置ssh免密登陆

exit                           # 退出刚才的 ssh localhost

cd ~/.ssh/                     # 若没有该目录,请先执行一次ssh localhost

ssh-keygen -t rsa              # 会有提示,都按回车就可以

cat id_rsa.pub >> authorized_keys  # 加入授权

chmod 600 ./authorized_keys    # 修改文件权限

 

再次输入ssh localhost ,若成功则不需要密码直接登陆

5、安装JAVA

sudo yum install java-1.7.0-openjdk java-1.7.0-openjdk-devel

默认安装目录在 /usr/lib/jvm/java-1.7.0-openjdk

配置JAVA_HOME环境变量

vim ~/.bashrc

在最后一行添加

export JAVA_HOME=/usr/lib/jvm/java-1.7.0-openjdk

 

使环境变量生效

source ~/.bashrc

检验设置是否正确,如果正确,则java -version与$JAVA_HOME/bin/java -version 输出结果一样

echo $JAVA_HOME    

java -version

$JAVA_HOME/bin/java -version

 

6、安装hadoop

进入目录 /usr/local ,解压hadoop压缩包

tar -zxvf hadoop-2.7.6.tar.gz

hadoop-2.7.6改名为hadoop

mv hadoop-2.7.6 hadoop

修改文件权限

sudo chown -R hadoop:hadoop ./hadoop

检查hadoop是否可用

cd /usr/local/hadoop

./bin/hadoop version

 

6、hadoop伪分布式配置

设置hadoop环境变量

vim ~/.bashrc

在结尾添加

export HADOOP_HOME=/usr/local/hadoop

export HADOOP_INSTALL=$HADOOP_HOME

export HADOOP_MAPRED_HOME=$HADOOP_HOME

export HADOOP_COMMON_HOME=$HADOOP_HOME

export HADOOP_HDFS_HOME=$HADOOP_HOME

export YARN_HOME=$HADOOP_HOME

export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native

export PATH=$PATH:$HADOOP_HOME/sbin:$HADOOP_HOME/bin

 

使配置生效

source ~/.bashrc

进入目录 /usr/local/hadoop/etc/hadoop

cd /usr/local/hadoop/etc/hadoop

修改配置文件 core-site.xml  hdfs-site.xml

core-site.xml

<configuration>

    <property>

        <name>hadoop.tmp.dir</name>

        <value>file:/usr/local/hadoop/tmp</value>

        <description>Abase for other temporary directories.</description>

    </property>

    <property>

        <name>fs.defaultFS</name>

        <value>hdfs://localhost:9000</value>

    </property>

</configuration>

 

hdfs-site.xml

<configuration>

    <property>

        <name>dfs.replication</name>

        <value>1</value>

    </property>

    <property>

        <name>dfs.namenode.name.dir</name>

        <value>file:/usr/local/hadoop/tmp/dfs/name</value>

    </property>

    <property>

        <name>dfs.datanode.data.dir</name>

        <value>file:/usr/local/hadoop/tmp/dfs/data</value>

    </property>

</configuration>

 

执行NameNode

hdfs namenode -format

成功的话,会看到 “successfully formatted” 和 “Exitting with status 0” 的提示,若为 “Exitting with status 1” 则是出错。

 

打开NameNode和DataNode守护进程

start-dfs.sh

判断是否打开成功

jps

 

访问192.168.9.135:50070

 

7、启动yarn

修改配置文件mapred-site.xml和yarn-site.xml

mapred-site.xml.template 重命名为mapred-site.xml

mv mapred-site.xml.template mapred-site.xml

mapred-site.xml

<configuration>

    <property>

        <name>mapreduce.framework.name</name>

        <value>yarn</value>

    </property>

</configuration>

yarn-site.xml

<configuration>

    <property>

        <name>yarn.nodemanager.aux-services</name>

        <value>mapreduce_shuffle</value>

        </property>

</configuration>

启动yarn

start-dfs.sh

查看后台进程

jps

 

访问192.168.9.135:8088

 

 参考http://dblab.xmu.edu.cn/blog/install-hadoop-in-centos/

猜你喜欢

转载自www.cnblogs.com/zhida/p/10844397.html