本帖最后由 howtodown 于 2014-8-19 11:49 编辑
问题导读:
1.如何修改本地时区?
2.需要安装那些软件?
3.带dos参数与不带dos编译有什么区别?
-- 参考:http://hadoop.apache.org/docs/r2 ... ativeLibraries.html
-- hadoop compile for CentOS
-- 编译hadoop之前,先修改源码中如下文件以支持本地时区
cd hadoop-2.4.1-src
vi ./hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/resources/webapps/static/yarn.dt.plugins.js
-- 将如下行
return new Date(parseInt(data)).toUTCString();
-- 修改为:
return new Date(parseInt(data)).toLocaleString();
----------------------------------------------------------------------------------
Requirements:
* Unix System
* JDK 1.6+
* Maven 3.0 or later
* Findbugs 1.3.9 (if running findbugs)
* ProtocolBuffer 2.5.0
* CMake 2.6 or newer (if compiling native code)
* Internet connection for first build (to fetch all Maven and Hadoop dependencies)
-- 安装如下包:
[root@funshion-hadoop192 hadoop-2.4.1-src]# rpm -qa|grep gzip
gzip-1.3.12-19.el6_4.x86_64
[root@funshion-hadoop192 hadoop-2.4.1-src]# rpm -qa|grep zlib
zlib-1.2.3-29.el6.x86_64
zlib-devel-1.2.3-29.el6.x86_64
zlib-static-1.2.3-29.el6.x86_64
-- 安装protobuf 2.5.0
-- 参考:http://www.cnblogs.com/Anker/p/3209764.html
mkdir -p /usr/local/protobuf
tar -xvf protobuf-2.5.0.tar.gz
cd protobuf-2.5.0
./configure --prefix=/usr/local/protobuf
make
make check
make install
vi + /etc/profile
export PATH=$PATH:/usr/local/protobuf/bin
export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/usr/local/protobuf/lib
source /etc/profile
-- 开始编译:
[root@funshion-hadoop192 hadoop-2.4.1-src]# mvn package -Pdist,native -DskipTests -Dtar
...
main:
[exec] $ tar cf hadoop-2.4.1.tar hadoop-2.4.1
[exec] $ gzip -f hadoop-2.4.1.tar
[exec]
[exec] Hadoop dist tar available at: /opt/software/hadoop-2.4.1-src/hadoop-dist/target/hadoop-2.4.1.tar.gz
[exec]
[INFO] Executed tasks
[INFO]
[INFO] --- maven-javadoc-plugin:2.8.1:jar (module-javadocs) @ hadoop-dist ---
[INFO] Building jar: /opt/software/hadoop-2.4.1-src/hadoop-dist/target/hadoop-dist-2.4.1-javadoc.jar
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Hadoop Main ................................ SUCCESS [ 2.125 s]
[INFO] Apache Hadoop Project POM ......................... SUCCESS [ 2.349 s]
[INFO] Apache Hadoop Annotations ......................... SUCCESS [ 7.223 s]
[INFO] Apache Hadoop Assemblies .......................... SUCCESS [ 0.494 s]
[INFO] Apache Hadoop Project Dist POM .................... SUCCESS [ 57.967 s]
[INFO] Apache Hadoop Maven Plugins ....................... SUCCESS [ 5.738 s]
[INFO] Apache Hadoop MiniKDC ............................. SUCCESS [ 6.730 s]
[INFO] Apache Hadoop Auth ................................ SUCCESS [01:24 min]
[INFO] Apache Hadoop Auth Examples ....................... SUCCESS [ 5.100 s]
[INFO] Apache Hadoop Common .............................. SUCCESS [07:35 min]
[INFO] Apache Hadoop NFS ................................. SUCCESS [ 11.409 s]
[INFO] Apache Hadoop Common Project ...................... SUCCESS [ 0.131 s]
[INFO] Apache Hadoop HDFS ................................ SUCCESS [15:32 min]
[INFO] Apache Hadoop HttpFS .............................. SUCCESS [ 34.246 s]
[INFO] Apache Hadoop HDFS BookKeeper Journal ............. SUCCESS [ 14.162 s]
[INFO] Apache Hadoop HDFS-NFS ............................ SUCCESS [ 7.359 s]
[INFO] Apache Hadoop HDFS Project ........................ SUCCESS [ 0.104 s]
[INFO] hadoop-yarn ....................................... SUCCESS [ 0.093 s]
[INFO] hadoop-yarn-api ................................... SUCCESS [01:57 min]
[INFO] hadoop-yarn-common ................................ SUCCESS [ 50.802 s]
[INFO] hadoop-yarn-server ................................ SUCCESS [ 0.063 s]
[INFO] hadoop-yarn-server-common ......................... SUCCESS [ 16.060 s]
[INFO] hadoop-yarn-server-nodemanager .................... SUCCESS [ 26.625 s]
[INFO] hadoop-yarn-server-web-proxy ...................... SUCCESS [ 5.495 s]
[INFO] hadoop-yarn-server-applicationhistoryservice ...... SUCCESS [ 9.923 s]
[INFO] hadoop-yarn-server-resourcemanager ................ SUCCESS [ 24.955 s]
[INFO] hadoop-yarn-server-tests .......................... SUCCESS [ 1.274 s]
[INFO] hadoop-yarn-client ................................ SUCCESS [ 10.712 s]
[INFO] hadoop-yarn-applications .......................... SUCCESS [ 0.080 s]
[INFO] hadoop-yarn-applications-distributedshell ......... SUCCESS [ 4.875 s]
[INFO] hadoop-yarn-applications-unmanaged-am-launcher .... SUCCESS [ 4.060 s]
[INFO] hadoop-yarn-site .................................. SUCCESS [ 0.062 s]
[INFO] hadoop-yarn-project ............................... SUCCESS [ 8.983 s]
[INFO] hadoop-mapreduce-client ........................... SUCCESS [ 0.209 s]
[INFO] hadoop-mapreduce-client-core ...................... SUCCESS [ 36.397 s]
[INFO] hadoop-mapreduce-client-common .................... SUCCESS [ 29.828 s]
[INFO] hadoop-mapreduce-client-shuffle ................... SUCCESS [ 5.075 s]
[INFO] hadoop-mapreduce-client-app ....................... SUCCESS [ 16.686 s]
[INFO] hadoop-mapreduce-client-hs ........................ SUCCESS [ 14.421 s]
[INFO] hadoop-mapreduce-client-jobclient ................. SUCCESS [ 11.174 s]
[INFO] hadoop-mapreduce-client-hs-plugins ................ SUCCESS [ 4.673 s]
[INFO] Apache Hadoop MapReduce Examples .................. SUCCESS [ 12.380 s]
[INFO] hadoop-mapreduce .................................. SUCCESS [ 7.728 s]
[INFO] Apache Hadoop MapReduce Streaming ................. SUCCESS [ 9.712 s]
[INFO] Apache Hadoop Distributed Copy .................... SUCCESS [ 12.677 s]
[INFO] Apache Hadoop Archives ............................ SUCCESS [ 4.079 s]
[INFO] Apache Hadoop Rumen ............................... SUCCESS [ 10.651 s]
[INFO] Apache Hadoop Gridmix ............................. SUCCESS [ 8.470 s]
[INFO] Apache Hadoop Data Join ........................... SUCCESS [ 5.563 s]
[INFO] Apache Hadoop Extras .............................. SUCCESS [ 6.085 s]
[INFO] Apache Hadoop Pipes ............................... SUCCESS [ 13.207 s]
[INFO] Apache Hadoop OpenStack support ................... SUCCESS [ 10.080 s]
[INFO] Apache Hadoop Client .............................. SUCCESS [ 9.622 s]
[INFO] Apache Hadoop Mini-Cluster ........................ SUCCESS [ 0.245 s]
[INFO] Apache Hadoop Scheduler Load Simulator ............ SUCCESS [ 10.502 s]
[INFO] Apache Hadoop Tools Dist .......................... SUCCESS [ 9.243 s]
[INFO] Apache Hadoop Tools ............................... SUCCESS [ 0.086 s]
[INFO] Apache Hadoop Distribution ........................ SUCCESS [01:42 min]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD SUCCESS
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 37:45 min
[INFO] Finished at: 2014-08-05T14:23:22+08:00
[INFO] Final Memory: 180M/917M
[INFO] ------------------------------------------------------------------------
-- 注意:带doc参数编译出来有近200M,不带doc编译出来的只有120M左右
mvn package -Pdist,native,docs -DskipTests -Dtar
mvn package -Pdist,native -DskipTests -Dtar
-- ############################################################################################ --
--------------------------------------------------------------------------------------------------
-- lzo 支持:
-- 参考:http://www.iteblog.com/archives/992
wget http://www.oberhumer.com/opensource/lzo/download/lzo-2.06.tar.gz
tar -zxvf lzo-2.06.tar.gz
cd lzo-2.06
export CFLAGS=-m64
./configure -enable-shared -prefix=/usr/local/hadoop/lzo/
make && make install
git clone https://github.com/twitter/hadoop-lzo.git
cd hadoop-lzo
export CFLAGS=-m64
export CXXFLAGS=-m64
export C_INCLUDE_PATH=/usr/local/hadoop/lzo/include
export LIBRARY_PATH=/usr/local/hadoop/lzo/lib
mvn clean package -Dmaven.test.skip=true
cd target/native/Linux-amd64-64
tar -cBf - -C lib . | tar -xBvf - -C ~
cp ~/libgplcompression* /usr/local/hadoop/lib/native/
cp /opt/software/hadoop-lzo/target/hadoop-lzo-0.4.20-SNAPSHOT.jar /usr/local/hadoop/share/hadoop/common/
hadoop2.4.1的64位centos的native包
链接:http://pan.baidu.com/s/1eQioTd4 密码:4pgb
hadoop2.4.1的64位编译包
链接:http://pan.baidu.com/s/1mgxBN3U 密码:
hadoop2.4.1的32位编译包
链接:http://pan.baidu.com/s/1mg2v4es 密码:krdx
官网源码包:未下载代码
http://pan.baidu.com/s/1dDgphJz
更多资源:
hadoop家族、strom、spark、Linux、flume等jar包、安装包汇总下载(持续更新)
|