hadoop伪分布式环境搭建记录

Hadoop快3个月没接触了,这阶段全是阿里云的东西,乘着领导让我给同事讲讲Hadoop的机会,复习下Hadoop的知识。

hadoop伪分布式环境搭建记录20160511
1.解压hadoop、jdk压缩包

tar -xvzf hadoop-2.6.0.tar.gz
tar -xvzf jdk-7u79-linux-x64.tar.gz

2.环境变量添加

vi ~/.bash_profile
export JAVA_HOME=/home/hadoop/jdk1.7.0_79
export PATH=$PATH:$JAVA_HOME/bin
export HADOOP_HOME=/home/hadoop/hadoop-2.6.0
export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$HADOOP_HOME/lib
export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
export HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib"


修改完成,生效配置文件
source ~/.bash_profile

检查是否成功
java -version
hadoop version

3.修改hadoop配置文件

<?xml version="1.0"?>
<!-- core-site.xml -->
<configuration>
 <property>
  <name>fs.default.name</name>
  <value>hdfs://127.0.0.1/</value>
 </property>
</configuration>
<?xml version="1.0"?>
<!-- hdfs-site.xml -->
<configuration>
 <property>
  <name>dfs.replication</name>
  <value>1</value>
 </property>
</configuration>
<?xml version="1.0"?>
<!-- mapred-site.xml -->
<configuration>
 <property>
  <name>mapred.job.tracker</name>
  <value>127.0.0.1:8021</value>
 </property>
<property>
  <name>mapreduce.jobhistory.address</name> 
  <value>127.0.0.1:10020</value> 
</property> 
<property> 
  <name>mapreduce.jobhistory.webapp.address</name> 
  <value>127.0.0.1:19888</value> 
</property>
</configuration>
<?xml version="1.0"?>
<!-- yarn-site.xml -->
<configuration>
 <property>
  <name>yarn.resourcemanager.address</name>
  <value>127.0.0.1:8032</value>
 </property>
 <property>
  <name>yarn.nodemanager.aux-services</name>
  <value>mapreduce.shuffle</value>
 </property>
</configuration>


4.安装ssh并配置免密码登录

yum install ssh
#或者
apt-get install ssh
ssh-keygen -t rsa -P '' -f ~/.ssh/id_rsa
cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
ssh localhost

5.格式化hdfs

hadoop namenode -format
#此处遇到报错:
java.net.UnknownHostException
#修改
vi /etc/sysconfig/network
#HOSTNAME修改为localhost(随自己喜好,我这里设置的是Hadoop,确认/etc/hosts 里有对应的host配置就行)
#重启
reboot


6.启动hadoop

#start-all.sh 不推荐使用,用这个nodemanager总是起不起来
start-dfs.sh
start-yarn.sh

此处遇到报错:
localhost: Error: JAVA_HOME is not set and could not be found.

#修改/etc/hadoop/hadoop-env.sh中设JAVA_HOME
source  hadoop-env.sh


重新执行,此处依旧报错:
Cannot find configuration directory: /etc/hadoop
在hadoop-env.sh 配置一条hadoop配置文件所在目录:

export HADOOP_CONF_DIR=/home/hadoop/hadoop-2.6.0/etc/hadoop
source  hadoop-env.sh


7.访问web
http://hadoop:8088/
http://hadoop:50070/
如果无法访问,关闭防火墙

service iptables stop

8.执行hdfs语句

hdfs dfs -ls


至此hadoop环境搭建完毕!


hadoop伪分布式环境搭建记录
https://www.920929.xyz/posts/e9c070d6.html
作者
DELIN
发布于
2016年5月12日
许可协议