forked from mirrors/nixpkgs
Merge pull request #31861 from kamilchm/spark
spark: provided-hadoop -> nixpkgs hadoop
This commit is contained in:
commit
7ab91da929
|
@ -1,20 +1,16 @@
|
||||||
{ stdenv, fetchzip, makeWrapper, jre, pythonPackages, coreutils
|
{ stdenv, fetchzip, makeWrapper, jre, pythonPackages, coreutils, hadoop
|
||||||
, RSupport? true, R
|
, RSupport? true, R
|
||||||
, mesosSupport ? true, mesos
|
, mesosSupport ? true, mesos
|
||||||
, version
|
, version
|
||||||
}:
|
}:
|
||||||
|
|
||||||
let
|
let
|
||||||
versionMap = {
|
sha256 = {
|
||||||
"2.2.1" = {
|
"1.6.3" = "142hw73wf20d846l83ydx0yg7qj5qxywm4h7qrhwnd7lsy2sbnjf";
|
||||||
hadoopVersion = "hadoop2.7";
|
"2.2.1" = "10nxsf9a6hj1263sxv0cbdqxdb8mb4cl6iqq32ljq9ydvk32s99c";
|
||||||
sparkSha256 = "10nxsf9a6hj1263sxv0cbdqxdb8mb4cl6iqq32ljq9ydvk32s99c";
|
}.${version};
|
||||||
};
|
|
||||||
};
|
|
||||||
in
|
in
|
||||||
|
|
||||||
with versionMap.${version};
|
|
||||||
|
|
||||||
with stdenv.lib;
|
with stdenv.lib;
|
||||||
|
|
||||||
stdenv.mkDerivation rec {
|
stdenv.mkDerivation rec {
|
||||||
|
@ -22,15 +18,15 @@ stdenv.mkDerivation rec {
|
||||||
name = "spark-${version}";
|
name = "spark-${version}";
|
||||||
|
|
||||||
src = fetchzip {
|
src = fetchzip {
|
||||||
url = "mirror://apache/spark/${name}/${name}-bin-${hadoopVersion}.tgz";
|
inherit sha256;
|
||||||
sha256 = sparkSha256;
|
url = "mirror://apache/spark/${name}/${name}-bin-without-hadoop.tgz";
|
||||||
};
|
};
|
||||||
|
|
||||||
buildInputs = [ makeWrapper jre pythonPackages.python pythonPackages.numpy ]
|
buildInputs = [ makeWrapper jre pythonPackages.python pythonPackages.numpy ]
|
||||||
++ optional RSupport R
|
++ optional RSupport R
|
||||||
++ optional mesosSupport mesos;
|
++ optional mesosSupport mesos;
|
||||||
|
|
||||||
untarDir = "${name}-bin-${hadoopVersion}";
|
untarDir = "${name}-bin-without-hadoop";
|
||||||
installPhase = ''
|
installPhase = ''
|
||||||
mkdir -p $out/{lib/${untarDir}/conf,bin,/share/java}
|
mkdir -p $out/{lib/${untarDir}/conf,bin,/share/java}
|
||||||
mv * $out/lib/${untarDir}
|
mv * $out/lib/${untarDir}
|
||||||
|
@ -42,6 +38,7 @@ stdenv.mkDerivation rec {
|
||||||
cat > $out/lib/${untarDir}/conf/spark-env.sh <<- EOF
|
cat > $out/lib/${untarDir}/conf/spark-env.sh <<- EOF
|
||||||
export JAVA_HOME="${jre}"
|
export JAVA_HOME="${jre}"
|
||||||
export SPARK_HOME="$out/lib/${untarDir}"
|
export SPARK_HOME="$out/lib/${untarDir}"
|
||||||
|
export SPARK_DIST_CLASSPATH=$(${hadoop}/bin/hadoop classpath)
|
||||||
export PYSPARK_PYTHON="${pythonPackages.python}/bin/${pythonPackages.python.executable}"
|
export PYSPARK_PYTHON="${pythonPackages.python}/bin/${pythonPackages.python.executable}"
|
||||||
export PYTHONPATH="\$PYTHONPATH:$PYTHONPATH"
|
export PYTHONPATH="\$PYTHONPATH:$PYTHONPATH"
|
||||||
${optionalString RSupport
|
${optionalString RSupport
|
||||||
|
@ -63,7 +60,7 @@ stdenv.mkDerivation rec {
|
||||||
homepage = "http://spark.apache.org";
|
homepage = "http://spark.apache.org";
|
||||||
license = stdenv.lib.licenses.asl20;
|
license = stdenv.lib.licenses.asl20;
|
||||||
platforms = stdenv.lib.platforms.all;
|
platforms = stdenv.lib.platforms.all;
|
||||||
maintainers = with maintainers; [ thoughtpolice offline ];
|
maintainers = with maintainers; [ thoughtpolice offline kamilchm ];
|
||||||
repositories.git = git://git.apache.org/spark.git;
|
repositories.git = git://git.apache.org/spark.git;
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue