在裝Hadoop之前首先需要: 1.java1.6.x 最好是sun的,1.5.x也可以 2.ssh 安裝ssh
$ sudo apt-get install ssh
$ sudo apt-get install rsync
下載Hadoop 從http://hadoop./core/releases.html 下 載最近發(fā)布的版本
最好為hadoop創(chuàng)建一個用戶: 比如創(chuàng)建一個group為hadoop user為hadoop的用戶以及組
$ sudo addgroup hadoop
$ sudo adduser --ingroup hadoop hadoop
解壓下載的hadoop文件,放到/home/hadoop目錄下 名字為hadoop 配置JAVA_HOME:
gedit ~/hadoop/conf/hadoop-env.sh
將
修改成java的安裝目錄:(我的是:/usr/lib/jvm/java-6-sun-1.6.0.15)
# The java implementation to use. Required.
export JAVA_HOME=/usr/lib/jvm/java-6-sun-1.6.0.15
現(xiàn)在可以使用單節(jié)點的方式運行:
$ cd hadoop
$ mkdir input $ cp conf/*.xml input $ bin/hadoop jar hadoop-*-examples.jar grep input output 'dfs[a-z.]+' $ cat output/*
Pseudo-distributed方式跑:
配置ssh
$ su - hadoop
$ ssh-keygen -t rsa -P "" Generating public/private rsa key pair. Enter file in which to save the key (/home/hadoop/.ssh/id_rsa): Created directory '/home/hadoop/.ssh'. Your identification has been saved in /home/hadoop/.ssh/id_rsa. Your public key has been saved in /home/hadoop/.ssh/id_rsa.pub. The key fingerprint is: 9d:47:ab:d7:22:54:f0:f9:b9:3b:64:93:12:75:81:27 hadoop@ubuntu
讓其不輸入密碼就能登錄:
hadoop@ubuntu:~$ cat ~/.ssh/id_rsa.pub >>
~/.ssh/authorized_keys
使用:
$ ssh localhost
看看是不是直接ok了。
hadoop配置文件: conf/core-site.xml
hadoop.tmp.dir配置為你想要的路徑,${user.name} 會自動擴展為運行hadoop的用戶名
conf/hdfs-site.xml
dfs.replication為默認block復(fù)制數(shù)量 conf/mapred-site.xml
執(zhí)行
格式化分布式文件系統(tǒng):
$ bin/hadoop namenode -format
啟動hadoop:
可以從
NameNode - http://localhost:50070/
JobTracker - http://localhost:50030/
查看NameNode和JobTracker
運行例子:
$ bin/hadoop fs -put conf input
$ bin/hadoop jar hadoop-*-examples.jar grep input
output 'dfs[a-z.]+'
look at the run result:
$ bin/hadoop fs -get output output
$ cat output/* 參考: 1、http://hadoop./common/docs/current/quickstart.html 2、http://www./wiki /Running_Hadoop_On_Ubuntu_Linux_%28Single-Node_Cluster%29 |
|
來自: ShangShujie > 《我的圖書館》