Hadoop2.5.2安装与配置
1、一、SSH配置:1、首先在三台服务器上安装SSH,服务器IP地址为:192.168.217.128; 192.168.217.129;192.168.217.130sudo apt-get install openssh-server openssh-client
2、然后分别在三台服务器上执行以下命令,配置SSH免秘钥:$ ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa$ cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys输入命令:ssh localhost第一次需要输入密码,然后再次输入:ssh localhost能无密码登陆,说明配置成功
3、手动复制ubuntu2,ubuntu3~/.ssh/id_dsa.pub 文件内容添加到ubuntu1的~/.ssh/authorized_keys 文件中;或者分别执行以下代码(注:zhou是用户名):在ubuntu2中执行$ scp ~/.ssh/id_dsa.pub zhou@ubuntu1:~/.ssh/authorized_keys在ubuntu3中执行$ scp ~/.ssh/id_dsa.pub zhou@ubuntu1:~/.ssh/authorized_keys
4、将三台服务器中~/.ssh/authorized_keys 文件内容复制成完全相同;然后将三台服务器中文件authorized_keys权限设置为:chmod 600 authorized_keys测试ssh ubuntu1 ssh ubuntu2 ssh ubuntu3第一次需要密码,之后就能实现三台机器无密码访问。
5、二、java1.7安装配置1、执行:sudo apt-get install default-jdk2、修改/etc/environment文件如下:PATH="/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:$JAVA_HOME/bin"export JAVA_HOME=/usr/java/jdk1.7.0_79export CLASSPATH=.:$JAVA_HOME/lib:$JAVA_HOME/jre/lib3、测试:java -version
6、三、hadoop安装配置1、下载解压hadoop:$ wget http://www.trieuvan.com/apache/hadoop/common/hadoop-2.5.2/hadoop-2.5.2.tar.gz$ sudo tar -xzvf hadoop-2.5.2.tar.gz -C /zhou/App
7、配置:etc/hadoop/core-site.xml<property> <name>fs.defaultFS</name> <value>hdfs://ubuntu1:9000</value> </property> <property> <name>hadoop.tmp.dir</name> <value>file:/home/yy/hadoop-2.5.2/tmp</value> </property> <property> <name>io.file.buffer.size</name> <value>131702</value> </property>
8、配置etc/hadoop/hdfs-site.xml<property> <na罪焐芡拂me>dfs.namenode.name.dir</name> <value>/home/zhou/hadoop-2.5.2/dfs/name</value> </property> <property> <name>dfs.datanode.data.dir</name> <value>/home/zhou/hadoop-2.5.2/dfs/data</value> </property> <property> <name>dfs.replication</name> <value>2</value> </property> <property> <name>dfs.namenode.secondary.http-address</name> <value>ubuntu1:9001</value> </property> <property><name>dfs.webhdfs.enabled</name><value>true</value> </property>
9、配置etc/hadoop/mapred-site.xml<property> <name>mapredu艘早祓胂ce.framework.name</name> <value>yarn</value> </property> <property> <name>mapreduce.jobhistory.address</name> <value>ubuntu1:10020</value> </property> <property> <name>mapreduce.jobhistory.webapp.address</name> <value>ubuntu1:19888</value> </property>
10、配置et罕铞泱殳c/hadoop/yarn-site.xml<property> <荏鱿胫协name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> <property> <name>yarn.nodemanager.auxservices.mapreduce.shuffle.class</name> <value>org.apache.hadoop.mapred.ShuffleHandler</value> </property> <property> <name>yarn.resourcemanager.address</name> <value>ubuntu1:8032</value> </property> <property> <name>yarn.resourcemanager.scheduler.address</name> <value>ubuntu1:8030</value> </property> <property> <name>yarn.resourcemanager.resource-tracker.address</name> <value>ubuntu1:8031</value> </property> <property> <name>yarn.resourcemanager.admin.address</name> <value>ubuntu1:8033</value> </property> <property> <name>yarn.resourcemanager.webapp.address</name> <value>ubuntu1:8088</value> </property> <property> <name>yarn.nodemanager.resource.memory-mb</name> <value>768</value> </property>
11、配置etc/hadoop/salvesubuntu2ubuntu3
12、将ubuntu1中hadoop整个目录复制到ubunt2/ubunt3相同目录下,必须保证三个服务器路径配置完全一致
13、配置.bashrc文件sudo gedit ~/.bashrc该命令会打开该文件的编辑窗口,在文件末尾追加下面内容,然后保存,关闭粕盘镱嘧编辑窗口。#HADOOP VARIABLES STARTexport JAVA_HOME=/usr/lib/jvm/java-7-openjdk-amd64export HADOOP_INSTALL=/zhou/App/hadoopexport PATH=$PATH:$HADOOP_INSTALL/binexport PATH=$PATH:$HADOOP_INSTALL/sbinexport HADOOP_MAPRED_HOME=$HADOOP_INSTALLexport HADOOP_COMMON_HOME=$HADOOP_INSTALLexport HADOOP_HDFS_HOME=$HADOOP_INSTALLexport YARN_HOME=$HADOOP_INSTALLexport HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_INSTALL/lib/nativeexport HADOOP_OPTS="-Djava.library.path=$HADOOP_INSTALL/lib"#HADOOP VARIABLES END
14、sbin目录下执行./start-all.sh访问http://localhost:50070/,到此安装配置成功。