- 浏览: 2596998 次
- 来自: 深圳
文章分类
最新评论
-
Wangwei86609:
非常好的规则引擎框架,支持决策树和多线程运行规则https:/ ...
规则引擎 -
comet0515:
dist.lib是什么文件?提示找不到
Jenkins(Hudson)+Findbugs/PMD/Checkstyle的配置要点 -
comet0515:
不是maven工程能不能用checkstyle插件?job是一 ...
HUDSON/Jenkins -
comet0515:
不是maven工程能不能用checkstyle插件?job是一 ...
Jenkins(Hudson)+Findbugs/PMD/Checkstyle的配置要点 -
comet0515:
不是maven工程能不能用checkstyle插件?job是一 ...
Jenkins(Hudson)+Findbugs/PMD/Checkstyle的配备要点
开源云计算技术系列(七)Cloudera (hadoop 0.20)
虚拟一套centos 5.3 os.
下载 jdk-6u16-linux-i586-rpm.bin
[root@hadoop ~]# chmod +x jdk-6u16-linux-i586-rpm.bin
[root@hadoop ~]# ./jdk-6u16-linux-i586-rpm.bin
[root@hadoop ~]# java -version
java version “1.6.0″
OpenJDK Runtime Environment (build 1.6.0-b09)
OpenJDK Client VM (build 1.6.0-b09, mixed mode)
[root@hadoop yum.repos.d]# wget http://archive.cloudera.com/redhat/cdh/cloudera-testing.repo
[root@hadoop yum.repos.d]# ls
CentOS-Base.repo CentOS-Base.repo.bak CentOS-Media.repo cloudera-testing.repo
[root@hadoop ~]# yum install hadoop-0.20 -y
Loaded plugins: fastestmirror
Loading mirror speeds from cached hostfile
Setting up Install Process
Parsing package install arguments
Resolving Dependencies
–> Running transaction check
—> Package hadoop-0.20.noarch 0:0.20.0+69-1 set to be updated
–> Finished Dependency Resolution
Dependencies Resolved
===============================================================================
Package Arch Version Repository Size
===============================================
Installing:
hadoop-0.20 noarch 0.20.0+69-1 cloudera-testing 18 M
Transaction Summary
==================================================
Install 1 Package(s)
Update 0 Package(s)
Remove 0 Package(s)
Total download size: 18 M
Downloading Packages:
hadoop-0.20-0.20.0+69-1.noarch.rpm | 18 MB 01:34
Running rpm_check_debug
Running Transaction Test
Finished Transaction Test
Transaction Test Succeeded
Running Transaction
Installing : hadoop-0.20 [1/1]
Installed: hadoop-0.20.noarch 0:0.20.0+69-1
Complete!
root@hadoop conf]# yum install hadoop-0.20-conf-pseudo -y
Loaded plugins: fastestmirror
Loading mirror speeds from cached hostfile
Setting up Install Process
Parsing package install arguments
Resolving Dependencies
–> Running transaction check
—> Package hadoop-0.20-conf-pseudo.noarch 0:0.20.0+69-1 set to be updated
–> Finished Dependency Resolution
Dependencies Resolved
========================================================
Package Arch Version Repository Size
========================================================
Installing:
hadoop-0.20-conf-pseudo noarch 0.20.0+69-1 cloudera-testing 11 k
Transaction Summary
==========================================================
Install 1 Package(s)
Update 0 Package(s)
Remove 0 Package(s)
Total download size: 11 k
Downloading Packages:
hadoop-0.20-conf-pseudo-0.20.0+69-1.noarch.rpm | 11 kB 00:00
Running rpm_check_debug
Running Transaction Test
Finished Transaction Test
Transaction Test Succeeded
Running Transaction
Installing : hadoop-0.20-conf-pseudo [1/1]
Installed: hadoop-0.20-conf-pseudo.noarch 0:0.20.0+69-1
Complete!
安装完后可以在这个目录下看到。
[root@hadoop conf.pseudo]# rpm -ql hadoop-0.20-conf-pseudo
/etc/hadoop-0.20/conf.pseudo
/etc/hadoop-0.20/conf.pseudo/README
/etc/hadoop-0.20/conf.pseudo/capacity-scheduler.xml
/etc/hadoop-0.20/conf.pseudo/configuration.xsl
/etc/hadoop-0.20/conf.pseudo/core-site.xml
/etc/hadoop-0.20/conf.pseudo/fair-scheduler.xml
/etc/hadoop-0.20/conf.pseudo/hadoop-env.sh
/etc/hadoop-0.20/conf.pseudo/hadoop-metrics.properties
/etc/hadoop-0.20/conf.pseudo/hadoop-policy.xml
/etc/hadoop-0.20/conf.pseudo/hdfs-site.xml
/etc/hadoop-0.20/conf.pseudo/log4j.properties
/etc/hadoop-0.20/conf.pseudo/mapred-site.xml
/etc/hadoop-0.20/conf.pseudo/masters
/etc/hadoop-0.20/conf.pseudo/slaves
/etc/hadoop-0.20/conf.pseudo/ssl-client.xml.example
/etc/hadoop-0.20/conf.pseudo/ssl-server.xml.example
/var/lib/hadoop-0.20
/var/lib/hadoop-0.20/cache
[root@hadoop conf.pseudo]# pwd
/etc/hadoop-0.20/conf.pseudo
[root@hadoop conf.pseudo]# more core-site.xml
<?xml version=”1.0″?>
<?xml-stylesheet type=”text/xsl” href=”configuration.xsl”?>
<configuration>
<property>
<name>fs.default.name</name>
<value>hdfs://localhost:8020</value>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/var/lib/hadoop-0.20/cache/${user.name}</value>
</property>
</configuration>
启动hadoop相关服务:
[root@hadoop conf.pseudo]# for service in /etc/init.d/hadoop-0.20-*
> do
> sudo $service start
> done
Starting Hadoop datanode daemon (hadoop-datanode): starting datanode, logging to /usr/lib/hadoop-0.20/bin/../logs/hadoop-hadoop-datanode-hadoop.out
[ OK ]
Starting Hadoop jobtracker daemon (hadoop-jobtracker): starting jobtracker, logging to /usr/lib/hadoop-0.20/bin/../logs/hadoop-hadoop-jobtracker-hadoop.out
[ OK ]
Starting Hadoop namenode daemon (hadoop-namenode): starting namenode, logging to /usr/lib/hadoop-0.20/bin/../logs/hadoop-hadoop-namenode-hadoop.out
[ OK ]
Starting Hadoop secondarynamenode daemon (hadoop-secondarynamenode): starting secondarynamenode, logging to /usr/lib/hadoop-0.20/bin/../logs/hadoop-hadoop-secondarynamenode-hadoop.out
[ OK ]
Starting Hadoop tasktracker daemon (hadoop-tasktracker): starting tasktracker, logging to /usr/lib/hadoop-0.20/bin/../logs/hadoop-hadoop-tasktracker-hadoop.out
[ OK ]
验证一下启动成功:
hadoop 3503 1 8 18:33 ? 00:00:03 /usr/java/jdk1.6.0_16/bin/java -Xmx1000m -Dcom.sun.manage
hadoop 3577 1 10 18:33 ? 00:00:04 /usr/java/jdk1.6.0_16/bin/java -Xmx1000m -Dcom.sun.manage
hadoop 3657 1 15 18:33 ? 00:00:05 /usr/java/jdk1.6.0_16/bin/java -Xmx1000m -Dcom.sun.manage
hadoop 3734 1 11 18:33 ? 00:00:04 /usr/java/jdk1.6.0_16/bin/java -Xmx1000m -Dcom.sun.manage
hadoop 3827 1 7 18:33 ? 00:00:02 /usr/java/jdk1.6.0_16/bin/java -Xmx1000m -Dhadoop.log.di
测试几个例子:
root@hadoop conf.pseudo]# hadoop-0.20 fs -mkdir input
[root@hadoop conf.pseudo]# hadoop-0.20 fs -put /etc/hadoop-0.20/conf/*.xml input
[root@hadoop conf.pseudo]# hadoop-0.20 fs -ls input
Found 6 items
-rw-r–r– 1 root supergroup 6275 2009-08-25 18:34 /user/root/input/capacity-scheduler.xml
-rw-r–r– 1 root supergroup 338 2009-08-25 18:34 /user/root/input/core-site.xml
-rw-r–r– 1 root supergroup 3032 2009-08-25 18:34 /user/root/input/fair-scheduler.xml
-rw-r–r– 1 root supergroup 4190 2009-08-25 18:34 /user/root/input/hadoop-policy.xml
-rw-r–r– 1 root supergroup 496 2009-08-25 18:34 /user/root/input/hdfs-site.xml
-rw-r–r– 1 root supergroup 213 2009-08-25 18:34 /user/root/input/mapred-site.xml
[root@hadoop conf.pseudo]# hadoop-0.20 jar /usr/lib/hadoop-0.20/hadoop-*-examples.jar grep input output ‘dfs[a-z.]+’
09/08/25 18:34:59 INFO mapred.FileInputFormat: Total input paths to process : 6
09/08/25 18:35:00 INFO mapred.JobClient: Running job: job_200908251833_0001
09/08/25 18:35:01 INFO mapred.JobClient: map 0% reduce 0%
09/08/25 18:35:20 INFO mapred.JobClient: map 33% reduce 0%
09/08/25 18:35:33 INFO mapred.JobClient: map 66% reduce 11%
09/08/25 18:35:42 INFO mapred.JobClient: map 66% reduce 22%
09/08/25 18:35:45 INFO mapred.JobClient: map 100% reduce 22%
09/08/25 18:35:57 INFO mapred.JobClient: map 100% reduce 100%
09/08/25 18:35:59 INFO mapred.JobClient: Job complete: job_200908251833_0001
09/08/25 18:35:59 INFO mapred.JobClient: Counters: 18
09/08/25 18:35:59 INFO mapred.JobClient: Job Counters
09/08/25 18:35:59 INFO mapred.JobClient: Launched reduce tasks=1
09/08/25 18:35:59 INFO mapred.JobClient: Launched map tasks=6
09/08/25 18:35:59 INFO mapred.JobClient: Da<wbr>ta-local map tasks=6<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: FileSystemCounters<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: FILE_BYTES_READ=100<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: HDFS_BYTES_READ=14544<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: FILE_BYTES_WRITTEN=422<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: HDFS_BYTES_WRITTEN=204<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: Map-Reduce Framework<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: Reduce input groups=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: Combine output records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: Map input records=364<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: Reduce shuffle bytes=124<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: Reduce output records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: Spilled Records=8<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: Map output bytes=86<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: Map input bytes=14544<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: Combine input records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: Map output records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.JobClient: Reduce input records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:35:59 INFO mapred.FileInputFormat: Total input paths to process : 1<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:00 INFO mapred.JobClient: Running job: job_200908251833_0002<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:01 INFO mapred.JobClient: map 0% reduce 0%<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:12 INFO mapred.JobClient: map 100% reduce 0%<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:24 INFO mapred.JobClient: map 100% reduce 100%<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Job complete: job_200908251833_0002<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Counters: 18<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Job Counters<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Launched reduce tasks=1<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Launched map tasks=1<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Da<wbr>ta-local map tasks=1<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: FileSystemCounters<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: FILE_BYTES_READ=100<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: HDFS_BYTES_READ=204<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: FILE_BYTES_WRITTEN=232<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: HDFS_BYTES_WRITTEN=62<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Map-Reduce Framework<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Reduce input groups=1<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Combine output records=0<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Map input records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Reduce shuffle bytes=0<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Reduce output records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Spilled Records=8<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Map output bytes=86<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Map input bytes=118<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Combine input records=0<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Map output records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">
09/08/25 18:36:26 INFO mapred.JobClient: Reduce input records=4</wbr></wbr>
[root@hadoop conf.pseudo]# hadoop-0.20 fs -ls
Found 2 items
drwxr-xr-x – root supergroup 0 2009-08-25 18:34 /user/root/input
drwxr-xr-x – root supergroup 0 2009-08-25 18:36 /user/root/output
[root@hadoop conf.pseudo]# hadoop-0.20 fs -ls output
Found 2 items
drwxr-xr-x – root supergroup 0 2009-08-25 18:36 /user/root/output/_logs
-rw-r–r– 1 root supergroup 62 2009-08-25 18:36 /user/root/output/part-00000
[root@hadoop conf.pseudo]# hadoop-0.20 fs -cat output/part-00000 | head
1 dfs.name.dir
1 dfs.permissions
1 dfs.replication
1 dfsadmin
转自:http://rdc.taobao.com/blog/dw/archives/436
相关推荐
Cloudera Hadoop 安装指南
Cloudera Hadoop 5&Hadoop高阶管理及调优课程,完整版,提供课件代码资料下载。 内容简介 本教程针对有一定Hadoop基础的学员,深入讲解如下方面的内容: 1、Hadoop2.0高阶运维,包括Hadoop节点增加删除、HDFS和...
Cloudera hadoop开发员培训教材,掌握hadoop,cdh平台的不错资料
作为业界最领先的企业级数据平台软件,Cloudera企业版除了包含业界最流行的基于开源Hadoop及其生态组件构建的CDH核心,还包含了大量为支撑企业级业务的高级管理特性。 借助于Cloudera企业版的整体解决方案,企业可以...
Cloudera hadoop官方文档 Cloudera Enterprise 5.11.x Documentation https://www.cloudera.com/documentation/enterprise/latest.html
Cloudera培训题目,主要关于spark,语言用的scala和python
cdh5-vagrant, 准备使用,手动调整 Cloudera Hadoop发行版 5已经置备集群 Hadoop发行版 5的 Vagrant脚本这个脚本提供了一个完整的自动方式来在完全分布式的node 中安装 Cloudera,无需用户干预。 因为我在互联网上...
该文档来自CCTC 2016中国云计算技术大会。Cloudera大数据技术专家韩军发表的题为“Hadoop架构在云上的应用实践”的主题演讲,欢迎下载!
cloudera连接自定义MySQL数据库,cloudera集成hadoop以及所有组件
最近学习hadoop,综述了大家的博客,终于弄出来一篇自己认为比较完成的文档。 centos6.2 mini hadoop-2.0.0-cdh4.3.0.tar.gz zookeeper-3.4.5-cdh4.3.0.tar.gz hbase-0.94.6-cdh4.3.0.tar.gz
公司使用的hadoop一般都不是apache下的hadoop,而是cloudera公司的hadoop,因为它除了提供hadoop外还提供了集成工具,所以,今天也采用cloudera hadoop进行安装学习。 一、先来看看hadoop对环境的要求 我选择的...
七. 参数调整 33 1. Mapreduce 33 2. Hdfs 34 3. Hbase 34 4. hive 35 5. Oozie 35 八. 常用命令 37 1. 安全模式 37 2. 查看block大小 37 九. 其他 37 1. 修改修改centos bogon主机名 37 2. 安装ganglia 38 3...
Hadoop在蓝汛 说说Cloudera和它的产品们 运维杂记——几次重大事故
Centos7 安装Cloudera。本文适用于安装最新版本的Cloudera Manger,在1.2 节中请选择最近版本的 Cloudera manager相关rpm安装包下载,并在后面的步骤中使用对应的rpm安装。否则在3.3节中安装 cm服务到40%时会卡住不...
下载修改(双计数“cloudera”)字数: 执行: 拉取 Docker 镜像: docker pull karthicks/cdh5-docker 在后台运行图像并挂载 wordCount.java : docker run -i -t -v /root/WordCount.java:/root/WordCount.java...
妳那伊抹微笑_云计算之ClouderaManager以及CHD5.1.0集群部署安装文档V1.0.docx 该文档与教程http://blog.csdn.net/u012185296/article/details/38084107 配套使用的,主要讲的是妳那伊抹微笑_云计算之Cloudera...
hadoop海量数据处理平台,大数据时代的到来,海量数据已成为下一步的发展趋势,所以我们要把数据做好,分析好!
出此书的目的就是为了帮助新人快速进入大数据行业,市面上有很多类似的书籍都是重理论少实践,特别缺少一线企业实践经验的传授,而这个手册会让您少走弯路、快速入门...里面包含有hadoop生态圈介绍以及机器学习算法介绍
在第四届中国云计算大会中,Cloudera创始人兼CTO Amr Awadallah发表了《Hadoop的剖析及第一手经验》的主题演讲。他不仅展示了最新的Hadoop实践应用案例,还为大家带来第一手的Hadoop经验分享。