summarylogtreecommitdiffstats
path: root/PKGBUILD
diff options
context:
space:
mode:
Diffstat (limited to 'PKGBUILD')
-rw-r--r--PKGBUILD72
1 files changed, 39 insertions, 33 deletions
diff --git a/PKGBUILD b/PKGBUILD
index d5424f361816..cb5bf5f1d634 100644
--- a/PKGBUILD
+++ b/PKGBUILD
@@ -2,66 +2,72 @@
# Contributor: Christian Krause ("wookietreiber") <kizkizzbangbang@gmail.com>
pkgname=apache-spark
-pkgver=1.6.1
+pkgver=1.6.2
pkgrel=1
pkgdesc="fast and general engine for large-scale data processing"
arch=('any')
url="http://spark.apache.org"
license=('APACHE')
-depends=('maven>=3.3.3' 'java-environment>=6' 'scala' 'python2>=2.7' 'hadoop>=2.6')
-optdepends=('python: PYSPARK_PYTHON=python3 pyspark'
- 'ipython: PYSPARK_DRIVER_PYTHON=ipython pyspark; IPYTHON=1 pyspark')
+depends=('java-environment>=6' 'hadoop')
+optdepends=('python2: python2 support for pyspark'
+ 'ipython2: ipython2 support for pyspark'
+ 'python: python3 support for pyspark'
+ 'ipython: ipython3 support for pyspark'
+ 'r: support for sparkR'
+ 'rsync: support rsync hadoop binaries from master')
install=apache-spark.install
-source=("http://d3kbcqa49mib13.cloudfront.net/spark-$pkgver.tgz"
- 'apache-spark-standalone.service'
- 'spark-env.sh')
-md5sums=('12e1368138840b62f08ed22a8637955d'
- 'bb7d8b85366e6f9cc0b2777eaea161a8'
- '0913001583e607849270090555dbd309')
+source=("http://d3kbcqa49mib13.cloudfront.net/spark-${pkgver}-bin-without-hadoop.tgz"
+ 'apache-spark-master.service'
+ 'apache-spark-slave@.service'
+ 'spark-env.sh'
+ 'spark-daemon-run.sh'
+ 'run-master.sh'
+ 'run-slave.sh')
+md5sums=('304394fbe2899211217f0cd9e9b2b5d9'
+ '9ffe1f9c4bb2ea4e5a75ab6469fe76d4'
+ '8d34bd4cc946f46625597ca606da8ab6'
+ 'f8cc449543df418b8adfcc36a3afb384'
+ '8ff953f0436209b6190add59703a34f0'
+ '028472b82e9def7d5d409f008d064fe2'
+ '99115eedc453c9b8ca04cca2e32e4537')
backup=('etc/apache-spark/spark-env.sh')
PKGEXT=${PKGEXT:-'.pkg.tar.xz'}
prepare() {
- mkdir -p "$srcdir/spark-$pkgver"
- cd "$srcdir/spark-$pkgver"
-
- sed -i 's|pid=$SPARK_PID_DIR/spark-$SPARK_IDENT_STRING-$command-$instance.pid|pid=/var/lib/apache-spark/spark-daemon.pid|' sbin/spark-daemon.sh
-}
-
-build() {
- cd "$srcdir/spark-$pkgver"
-
- export MAVEN_OPTS="-Xmx2g -XX:MaxPermSize=512M -XX:ReservedCodeCacheSize=512m"
-
- dev/change-scala-version.sh 2.11
-
- JAVA_HOME=/usr/lib/jvm/default-runtime ./make-distribution.sh -Pscala-2.11 -DskipTests -Dmaven.repo.local=/tmp -DautoVersionSubmodules=true -U -Djline.version=2.13 -Djline.groupid=jline -Pyarn -Phadoop-2.6
+ cd "$srcdir/spark-${pkgver}-bin-without-hadoop"
}
package() {
- cd "$srcdir/spark-$pkgver"
+ cd "$srcdir/spark-${pkgver}-bin-without-hadoop"
- install -d "$pkgdir/usr/bin" "$pkgdir/usr/share"
+ install -d "$pkgdir/usr/bin" "$pkgdir/opt" "$pkgdir/var/log/apache-spark"
- cp -r "$srcdir/spark-$pkgver/dist" "$pkgdir/usr/share/apache-spark/"
+ cp -r "$srcdir/spark-${pkgver}-bin-without-hadoop" "$pkgdir/opt/apache-spark/"
cd "$pkgdir/usr/bin"
for binary in beeline pyspark sparkR spark-class spark-shell spark-sql spark-submit load-spark-env.sh; do
- binpath="/usr/share/apache-spark/bin/$binary"
+ binpath="/opt/apache-spark/bin/$binary"
ln -s "$binpath" $binary
- sed -i 's|^export SPARK_HOME=.*$|export SPARK_HOME=/usr/share/apache-spark|' "$pkgdir/$binpath"
+ sed -i 's|^export SPARK_HOME=.*$|export SPARK_HOME=/opt/apache-spark|' "$pkgdir/$binpath"
done
mkdir -p $pkgdir/etc/profile.d
echo '#!/bin/sh' > $pkgdir/etc/profile.d/apache-spark.sh
- echo 'SPARK_HOME=/usr/share/apache-spark' >> $pkgdir/etc/profile.d/apache-spark.sh
+ echo 'SPARK_HOME=/opt/apache-spark' >> $pkgdir/etc/profile.d/apache-spark.sh
echo 'export SPARK_HOME' >> $pkgdir/etc/profile.d/apache-spark.sh
chmod 755 $pkgdir/etc/profile.d/apache-spark.sh
- install -Dm644 "$srcdir/apache-spark-standalone.service" "$pkgdir/usr/lib/systemd/system/apache-spark-standalone.service"
+ install -Dm644 "$srcdir/apache-spark-master.service" "$pkgdir/usr/lib/systemd/system/apache-spark-master.service"
+ install -Dm644 "$srcdir/apache-spark-slave@.service" "$pkgdir/usr/lib/systemd/system/apache-spark-slave@.service"
install -Dm644 "$srcdir/spark-env.sh" "$pkgdir/etc/apache-spark/spark-env.sh"
+ for script in run-master.sh run-slave.sh spark-daemon-run.sh; do
+ install -Dm755 "$srcdir/$script" "$pkgdir/opt/apache-spark/sbin/$script"
+ done
+ install -Dm644 "$srcdir/spark-${pkgver}-bin-without-hadoop/conf"/* "$pkgdir/etc/apache-spark"
- cd "$pkgdir/usr/share/apache-spark/conf"
- ln -sf "/etc/apache-spark/spark-env.sh" .
+ cd "$pkgdir/opt/apache-spark"
+ mv conf conf-templates
+ ln -sf "/etc/apache-spark" conf
+ ln -sf "/var/lib/apache-spark/work" .
}