Package Details: apache-spark 3.5.1-1

Git Clone URL: https://aur.archlinux.org/apache-spark.git (read-only, click to copy)
Package Base: apache-spark
Description: A unified analytics engine for large-scale data processing
Upstream URL: http://spark.apache.org
Keywords: spark
Licenses: Apache
Submitter: huitseeker
Maintainer: aakashhemadri
Last Packager: aakashhemadri
Votes: 57
Popularity: 0.000044
First Submitted: 2015-10-04 09:31 (UTC)
Last Updated: 2024-05-07 17:40 (UTC)

Dependencies (2)

Required by (2)

Sources (4)

Latest Comments

1 2 3 4 5 6 7 Next › Last »

Sol_Maxguy commented on 2025-02-07 08:00 (UTC)

Following Appuchia's comment, the first sha256sum should be : d8e08877ed428bf9fcd44dbbec8cecadad34785a05513e5020ae74ffdabcbc83

If needed, you can type the following command to update to the right sha256 (if I mistyped the key) : sha256sum spark-3.5.4-bin-hadoop3.tgz

Appuchia commented on 2025-02-06 11:00 (UTC)

Also change the first sha256sum to 8e08877ed428bf9fcd44dbbec8cecadad34785a05513e5020ae74ffdabcbc83. The sha512 of the file matches the one on the cdn.

This makes the install work

Appuchia commented on 2025-02-06 10:56 (UTC)

The file version 3.5.1 is no longer available in https://dlcdn.apache.org/spark/ Changing the version in the PKGBUILD to 3.5.4 makes the download work

freebench commented on 2024-10-22 13:54 (UTC)

A proper url is required I assume

==> Making package: apache-spark 3.5.1-1 (Tue 22 Oct 2024 04:52:06 PM EEST)
==> Checking runtime dependencies...
==> Checking buildtime dependencies...
==> Retrieving sources...
  -> Downloading spark-3.5.1-bin-hadoop3.tgz...
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed
  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
curl: (22) The requested URL returned error: 404
==> ERROR: Failure while downloading https://dlcdn.apache.org/spark/spark-3.5.1/spark-3.5.1-bin-hadoop3.tgz
    Aborting...

krutoileshii commented on 2024-03-01 01:27 (UTC) (edited on 2024-03-01 01:29 (UTC) by krutoileshii)

diff of the update to the latest + adding python back in

--- PKGBUILD.orig   2023-09-29 07:49:21.000000000 -0600
+++ PKGBUILD    2024-02-29 18:22:53.703306712 -0700
@@ -5,7 +5,7 @@
 # Contributor: Emanuel Fontelles ("emanuelfontelles") <emanuelfontelles@hotmail.com>

 pkgname=apache-spark
-pkgver=3.5.0
+pkgver=3.5.1
 pkgrel=1
 pkgdesc="A unified analytics engine for large-scale data processing"
 arch=('any')
@@ -17,7 +17,7 @@
         'apache-spark.sh'
         'apache-spark-master.service'
         'apache-spark-worker@.service')
-sha256sums=('ac064ec620129b4b9630e180e25ac2b71a8ccd4c6912bf5b5631943b742f0379'
+sha256sums=('5df15f8027067c060fe47ebd351a1431a61dbecc9c28b8dd29e2c6e1935c23eb'
             '0cc82baad4d878d4e2bc5864a00b99d38f2906781ea47ee6282546788e797049'
             'de54c025ca8ce34a7b4fd95ec7b8d5dec44582787a0bd8da09232f26e2182c9a'
             '47e6c154daecf7631ac9a33fe53a76888070c823c4381fcbde8d98377e586505')
@@ -28,7 +28,7 @@
     cd spark-${pkgver}-bin-hadoop3

     # Remove Python and R files
-    rm -rf python R
+    #rm -rf python R

     # Remove windows batch files
     rm bin/*.cmd

ViToni commented on 2024-02-10 19:59 (UTC)

Pyspark is also broken for me. Seems this package is supposed to be only for cluster installation which is inconvenient, when one want to work also locally...

lllf commented on 2023-10-09 12:39 (UTC) (edited on 2023-10-09 12:42 (UTC) by lllf)

Why are the Python and R files removed? This breaks the pyspark command.

prepare() {
    ...
    # Remove Python and R files
    rm -rf python R
}
$ pyspark
Python 3.11.3 (main, Jun  5 2023, 09:32:32) [GCC 13.1.1 20230429] on linux
Type "help", "copyright", "credits" or "license" for more information.
Could not open PYTHONSTARTUP
FileNotFoundError: [Errno 2] No such file or directory: '/opt/apache-spark/python/pyspark/shell.py'
>>> 

ttc0419 commented on 2023-06-24 12:38 (UTC)

@agaskell Updated

agaskell commented on 2023-06-23 18:43 (UTC)

@tcc0419 @PolarianDev the latest version of Spark is 3.4.1.

PolarianDev commented on 2023-03-26 12:03 (UTC)

@ttc0419 thank you, I might update it later today but I am also busy at the moment :)