summaryrefslogtreecommitdiffstats
path: root/pkgs/applications/networking/cluster/hadoop
diff options
context:
space:
mode:
authorvolth <volth@volth.com>2017-12-20 13:22:00 +0000
committervolth <volth volth@volth.com>2017-12-22 13:25:15 +0000
commit74588bf88adc304a49144874077d718202415408 (patch)
tree7f6fe7554c3786f447e1520009c7ba1bd1734a7a /pkgs/applications/networking/cluster/hadoop
parente6c3dcfcf7ef39335a2462040b364e653dfb5574 (diff)
hadoop: 2.2.0 -> 2.7.5, 2.8.3, 2.9.0, 3.0.0
Diffstat (limited to 'pkgs/applications/networking/cluster/hadoop')
-rw-r--r--pkgs/applications/networking/cluster/hadoop/default.nix182
1 files changed, 138 insertions, 44 deletions
diff --git a/pkgs/applications/networking/cluster/hadoop/default.nix b/pkgs/applications/networking/cluster/hadoop/default.nix
index 9af4cf3f0ed6..f96ee42fb1c1 100644
--- a/pkgs/applications/networking/cluster/hadoop/default.nix
+++ b/pkgs/applications/networking/cluster/hadoop/default.nix
@@ -1,51 +1,145 @@
-{ stdenv, fetchurl, makeWrapper, which, jre, bash }:
+{ stdenv, fetchurl, makeWrapper, pkgconfig, which, maven, cmake, jre, bash, coreutils, glibc, protobuf2_5, fuse, snappy, zlib, bzip2, openssl }:
-stdenv.mkDerivation rec {
- name = "hadoop-2.2.0";
+let
+ common = { version, sha256, dependencies-sha256, tomcat }:
+ let
+ # compile the hadoop tarball from sources, it requires some patches
+ binary-distributon = stdenv.mkDerivation rec {
+ name = "hadoop-${version}-bin";
+ src = fetchurl {
+ url = "mirror://apache/hadoop/common/hadoop-${version}/hadoop-${version}-src.tar.gz";
+ inherit sha256;
+ };
- src = fetchurl {
- url = "mirror://apache/hadoop/common/${name}/${name}.tar.gz";
- sha256 = "0r0kx8arsrvmcfy0693hpv4cz3i0razvk1xa3yhlf3ybb80a8106";
+ # perform fake build to make a fixed-output derivation of dependencies downloaded from maven central (~100Mb in ~3000 files)
+ fetched-maven-deps = stdenv.mkDerivation {
+ name = "hadoop-${version}-maven-deps";
+ inherit src nativeBuildInputs buildInputs configurePhase;
+ buildPhase = ''
+ while mvn package -Dmaven.repo.local=$out/.m2 ${mavenFlags} -Dmaven.wagon.rto=5000; [ $? = 1 ]; do
+ echo "timeout, restart maven to continue downloading"
+ done
+ '';
+ # keep only *.{pom,jar,xml,sha1,so,dll,dylib} and delete all ephemeral files with lastModified timestamps inside
+ installPhase = ''find $out/.m2 -type f -regex '.+\(\.lastUpdated\|resolver-status\.properties\|_remote\.repositories\)' -delete'';
+ outputHashAlgo = "sha256";
+ outputHashMode = "recursive";
+ outputHash = dependencies-sha256;
+ };
+
+ nativeBuildInputs = [ maven cmake pkgconfig ];
+ buildInputs = [ fuse snappy zlib bzip2 openssl protobuf2_5 ];
+ # most of the hardcoded pathes are fixed in 2.9.x and 3.0.0, this list of patched files might be reduced when 2.7.x and 2.8.x will be deprecated
+ postPatch = ''
+ for file in hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/HardLink.java \
+ hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/Shell.java \
+ hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DefaultContainerExecutor.java \
+ hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DockerContainerExecutor.java \
+ hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/launcher/ContainerLaunch.java \
+ hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java; do
+ if [ -f "$file" ]; then
+ substituteInPlace "$file" \
+ --replace '/usr/bin/stat' 'stat' \
+ --replace '/bin/bash' 'bash' \
+ --replace '/bin/ls' 'ls' \
+ --replace '/bin/mv' 'mv'
+ fi
+ done
+ '';
+ configurePhase = "true"; # do not trigger cmake hook
+ mavenFlags = "-Drequire.snappy -Drequire.bzip2 -DskipTests -Pdist,native -e";
+ # prevent downloading tomcat during the build
+ preBuild = stdenv.lib.optionalString (tomcat != null) ''
+ install -D ${tomcat.src} hadoop-hdfs-project/hadoop-hdfs-httpfs/downloads/apache-tomcat-${tomcat.version}.tar.gz
+ install -D ${tomcat.src} hadoop-common-project/hadoop-kms/downloads/apache-tomcat-${tomcat.version}.tar.gz
+ '';
+ buildPhase = ''
+ # 'maven.repo.local' must be writable
+ mvn package --offline -Dmaven.repo.local=$(cp -dpR ${fetched-maven-deps}/.m2 ./ && chmod +w -R .m2 && pwd)/.m2 ${mavenFlags}
+ # remove runtime dependency on $jdk/jre/lib/amd64/server/libjvm.so
+ patchelf --set-rpath ${stdenv.lib.makeLibraryPath [glibc]} hadoop-dist/target/hadoop-${version}/lib/native/libhadoop.so.1.0.0
+ patchelf --set-rpath ${stdenv.lib.makeLibraryPath [glibc]} hadoop-dist/target/hadoop-${version}/lib/native/libhdfs.so.0.0.0
+ '';
+ installPhase = "mv hadoop-dist/target/hadoop-${version} $out";
+ };
+ in
+ stdenv.mkDerivation rec {
+ name = "hadoop-${version}";
+
+ src = binary-distributon;
+
+ nativeBuildInputs = [ makeWrapper ];
+
+ installPhase = ''
+ mkdir -p $out/share/doc/hadoop/
+ cp -dpR * $out/
+ mv $out/*.txt $out/share/doc/hadoop/
+
+ for n in $out/bin/*; do
+ if [ -f "$n" ]; then # only regular files
+ wrapProgram "$n" \
+ --prefix PATH : "${stdenv.lib.makeBinPath [ which jre bash coreutils ]}" \
+ --prefix JAVA_LIBRARY_PATH : "${stdenv.lib.makeLibraryPath [ openssl snappy zlib bzip2 ]}" \
+ --set JAVA_HOME "${jre}" \
+ --set HADOOP_PREFIX "$out"
+ fi
+ done
+ '';
+
+ meta = with stdenv.lib; {
+ homepage = "http://hadoop.apache.org/";
+ description = "Framework for distributed processing of large data sets across clusters of computers";
+ license = licenses.asl20;
+
+ longDescription = ''
+ The Apache Hadoop software library is a framework that allows for
+ the distributed processing of large data sets across clusters of
+ computers using a simple programming model. It is designed to
+ scale up from single servers to thousands of machines, each
+ offering local computation and storage. Rather than rely on
+ hardware to deliver high-avaiability, the library itself is
+ designed to detect and handle failures at the application layer,
+ so delivering a highly-availabile service on top of a cluster of
+ computers, each of which may be prone to failures.
+ '';
+ maintainers = with maintainers; [ volth ];
+ platforms = [ "x86_64-linux" ];
+ };
+ };
+
+ tomcat_6_0_48 = rec {
+ version = "6.0.48";
+ src = fetchurl {
+ # do not use "mirror://apache/" here, tomcat-6 is legacy and has been removed from the mirrors
+ url = "https://archive.apache.org/dist/tomcat/tomcat-6/v${version}/bin/apache-tomcat-${version}.tar.gz";
+ sha256 = "1w4jf28g8p25fmijixw6b02iqlagy2rvr57y3n90hvz341kb0bbc";
+ };
};
- buildInputs = [ makeWrapper ];
-
- buildPhase = ''
- for n in "bin/"* "sbin/"*; do
- sed -i $n -e "s|#!/usr/bin/env bash|#! ${bash}/bin/bash|"
- done
- '' + stdenv.lib.optionalString (!stdenv.isDarwin) ''
- patchelf --set-interpreter "$(cat $NIX_CC/nix-support/dynamic-linker)" bin/container-executor;
- '';
-
- installPhase = ''
- mkdir -p $out
- mv *.txt share/doc/hadoop/
- mv * $out
-
- for n in $out/{bin,sbin}"/"*; do
- wrapProgram $n --prefix PATH : "${stdenv.lib.makeBinPath [ which jre bash ]}" --set JAVA_HOME "${jre}" --set HADOOP_PREFIX "$out"
- done
- '';
-
- meta = {
- homepage = http://hadoop.apache.org/;
- description = "Framework for distributed processing of large data sets across clusters of computers";
- license = stdenv.lib.licenses.asl20;
-
- longDescription = ''
- The Apache Hadoop software library is a framework that allows for
- the distributed processing of large data sets across clusters of
- computers using a simple programming model. It is designed to
- scale up from single servers to thousands of machines, each
- offering local computation and storage. Rather than rely on
- hardware to deliver high-avaiability, the library itself is
- designed to detect and handle failures at the application layer,
- so delivering a highly-availabile service on top of a cluster of
- computers, each of which may be prone to failures.
- '';
-
- platforms = stdenv.lib.platforms.linux;
+in {
+ hadoop_2_7 = common {
+ version = "2.7.5";
+ sha256 = "0dvgi4z2pi45b0yzzm9zxjp0lwvd8lwqn4903mxlqh7xv2g3llnp";
+ dependencies-sha256 = "137v077q022997ih41n412l2xq4ynypk5dl5p4n16i16745k9lk2";
+ tomcat = tomcat_6_0_48;
+ };
+ hadoop_2_8 = common {
+ version = "2.8.3";
+ sha256 = "0ziairmgfi05iqmib66lysanlabrp1cq6palnmwkiss098b5dcjd";
+ dependencies-sha256 = "1qmdmiprr0l0i0q3w9nv964db5cm3615iqczywrs5v8qaaf580b3";
+ tomcat = tomcat_6_0_48;
+ };
+ hadoop_2_9 = common {
+ version = "2.9.0";
+ sha256 = "03mh193bccm4bi7cq6ym651zal82d0kvl00yhnzzrsnwq10j2h62";
+ dependencies-sha256 = "0773lw4d03r89dcrrsglbr4zgcz2m9b15587nxdz4clm2f5gqiw5";
+ tomcat = tomcat_6_0_48;
+ };
+ hadoop_3_0 = common {
+ version = "3.0.0";
+ sha256 = "0h22mdpgvg9v4rbc8gn4m8vrpkq45lwk7m8f5c25fvnzdyrc8fzc";
+ dependencies-sha256 = "0ghbl768al3wpgz0rx2mqbnf6p7cdqpkq6yk6hgf65b4qab4aw9v";
+ tomcat = null;
};
}