-
hadoop 2.0 安装大数据平台 / dataplatform 2dot5 写在2017年04月25日Hadoop2.0配置SSH安装在线安装ssh #sudo apt-get install openssh-serveropenssh-client手工安装ssh存储ssh密码#ssh-keygen -t dsa -P ” -f ~/.ssh/id_dsa#cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys#sudo chmod go-w $HOME $HOME/.ssh#sudo chmod 600 $HOME/.ssh/authorized_keys#sudo chown `whoami` $HOME/.ssh/authorized_keys测试连接本地服务,无密码登陆,则说明ssh服务安装配置正确#ssh localhost#exit安装JDK安装必须1.6或者1.6以上版本。#sudo mkdir /usr/java#cd /usr/java#sudo wgethttp://download.oracle.com/otn-pub/java/jdk/6u31-b04/jdk-6u31-linux-i586.bin#sudo chmod o+w jdk-6u31-linux-i586.bin#sudo chmod +x jdk-6u31-linux-i586.bin#sudo ./jdk-6u31-linux-i586.bin修改环境变量/etc/profile文件中增加如下代码export JAVA_HOME=/usr/java/jdk1.6.0_24export PATH=$PATH:/usr/java/jdk1.6.0_24/binexport CLASSPATH=/usr/java/jdk1.6.0_24/lib/dt.jar:/usr/java/jdk1.6.0_24/lib/tools.jar#source /etc/profile测试# java -version显示java版本,则证明安装配置正确安装hadoop选择一个linux系统,下载并解压hadoop2.0.x并解压到/home/hadoop-2.0.0-alpha。设置hadoop环境变量/etc/profile文件增加如下内容export HADOOP_PREFIX=”/home/hadoop-2.0.0-alpha”export PATH=$PATH:$HADOOP_PREFIX/binexport PATH=$PATH:$HADOOP_PREFIX/sbinexport HADOOP_MAPRED_HOME=${HADOOP_PREFIX}export HADOOP_COMMON_HOME=${HADOOP_PREFIX}export HADOOP_HDFS_HOME=${HADOOP_PREFIX}export YARN_HOME=${HADOOP_PREFIX}#source /etc/profile修改hadoop配置文件hadoop 2.x 版本中 /etc/hadoop 是默认的配置文件夹1、编辑 core-site.xml<configuration><property><name>fs.default.name</name><value>hdfs://localhost:8020</value><description>The name of the default file system. Either the literal string “local” or a host:port for NDFS.</description><final>true</final></property></configuration>2、编辑hdfs-site.xml<configuration><property><name>dfs.namenode.name.dir</name><value>file:/home/hadoop_space/dfs/name</value><description>Determines where on the local filesystem the DFS name node should store the name table. If this is a comma-delimited list of directories then the name table is replicated in all of the directories, for redundancy. </description><final>true</final></property><property><name>dfs.datanode.data.dir</name><value>file:/home/hadoop_space/dfs/data</value><description>Determines where on the local filesystem an DFS data node should store its blocks. If this is a comma-delimited list of directories, then data will be stored in all named directories, typically on different devices.Directories that do not exist are ignored.</description><final>true</final></property><property><name>dfs.replication</name><value>1</value></property><property><name>dfs.permissions</name><value>false</value></property></configuration>说明:file:/home/hadoop_space/dfs/name和file:/home/hadoop_space/dfs/data是计算机中的一些文件夹,用于存放数据和编辑文件。3、创建文件mapred-site.xml<configuration><property><name>mapreduce.framework.name</name><value>yarn</value></property><property><name>mapred.system.dir</name><value>file:/home/hadoop_space/mapred/system</value><final>true</final></property><property><name>mapred.local.dir</name><value>file:/home/hadoop_space/mapred/local</value><final>true</final></property></configuration>说明:file:/home/hadoop_space/mapred/system和file:/home/hadoop_space/mapred/local用于存放数据。4、编辑yarn-site.xml<configuration><property><name>yarn.nodemanager.aux-services</name><value>mapreduce_shuffle</value></property><property><name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name><value>org.apache.hadoop.mapred.ShuffleHandler</value></property></configuration>5、创建hadoop-env.sh并添加:export JAVA_HOME=/usr/java/jdk1.6.0_24启动hadoop1、格式化 namenode# hdfs namenode -format2、开始守护进程# hadoop-daemon.sh start namenode# hadoop-daemon.sh start datanode或者用以下命令替换上述命令:# start-dfs.sh3、开始 yarn 守护进程# yarn-daemon.sh start resourcemanager# yarn-daemon.sh start nodemanager或者用以下命令替换上述命令:# start-yarn.sh检查守护进程是否启动# jps2539 NameNode2744 NodeManager3075 Jps3030 DataNode2691 ResourceManager如果上述五个进程都存在,则证明hadoop启动成功。查看hadoop资源管理页面http://localhost:8088上传文件#列出HDFS文件hdfs dfs -ls /hdfs dfs -ls /tmp#创建存储文件夹hadoop fs -mkdir /tmp创建一个文件test.txt,文件内容为hello world#上传文件/root/tmp到/tmp/test.txthadoop fs -copyFromLocal /home/test.txt /tmp/test.txt或hafs dfs -copyFromLocal /home/test.txt /tmp/test.txt#查看test.txt文件内容hadoop fs -cat /tmp/test.txt或者hdfs dfs -cat /tmp/test.txthadoop fs -cat hdfs://localhost:8020/tmp/test.txt或者hdfs dfs -cat hdfs://localhost:8020/tmp/test.txtDFSShellHDFS允许用户数据由文件和文件夹式的管理,它提供一个接口叫DFSShell,让用户和HDFS中的数据交互 命令集的语法跟其他的shells(bash,csh)相似创建目录foodir : hadoop fs -mkdir /foodir查看文件 /foodir/myfile.txt : hadoop dfs -cat /foodir/myfile.txt删除文件/foodir/myfile.txt : hadoop dfs -rm /foodir myfile.txtDFSAdminDFSAdmin命令集是用于管理dfs集群的,这些命令只由HDFS管理员使用将集群设置成安全模式 : hadoop dfsadmin -safemode enter产生一个数据节点的列表 : hadoop dfsadmin -report或者hdfs dfsadmin -report去掉一个数据节点: hadoop dfsadmin -decommission datanodename
下一篇: HBase shell 基本命令
-
Hbase安装
上篇文章我们已经讲了Hadoop的安装, 这...
-
zookeeper安装
下载地址: http://zookeeper.apache.or...
-
java Socket 简单的即时聊天 ...
我们上面讲了多线程,今天我们用Socket...
-
Hadoop安装(Linux环境)
Hadoop俨然已经成为大数据的标准,搭建...
-
HBase shell 基本命令
我在前面已经分享了hadoop和hbase的安装...
-
kafka安装
下载地址: http://kafka.apache.org 1...
相关推荐
您好!请登录
已有0评论