Package Details: apache-spark 3.5.1-1

Git Clone URL: https://aur.archlinux.org/apache-spark.git (read-only, click to copy)
Package Base: apache-spark
Description: A unified analytics engine for large-scale data processing
Upstream URL: http://spark.apache.org
Keywords: spark
Licenses: Apache
Submitter: huitseeker
Maintainer: aakashhemadri
Last Packager: aakashhemadri
Votes: 57
Popularity: 0.000050
First Submitted: 2015-10-04 09:31 (UTC)
Last Updated: 2024-05-07 17:40 (UTC)

Dependencies (2)

Required by (2)

Sources (4)

Latest Comments

« First ‹ Previous 1 2 3 4 5 6 7 Next › Last »

huitseeker commented on 2016-12-29 18:40 (UTC)

@steph.schie updated!

steph.schie commented on 2016-12-27 14:33 (UTC)

Why is hadoop a dependency? I don't need and want to use hadoop with spark.

huitseeker commented on 2016-11-18 07:45 (UTC)

There should be a file /etc/profile.d/apache-spark.sh that sets the $SPARK_HOME env variable for you correctly. If this is set up right, you should need no patching of binaries. To test that assumption, first check the value of $SPARK_HOME (should be /opt/apache-spark) and then run sh /opt/apache-spark/bin/load-spark-env.sh. Report (with as much complete information as possible) if you see an error then.

TaXules commented on 2016-11-08 16:30 (UTC)

To fix the error "ls: cannot access '/usr/assembly/target/scala-2.10': No such …", you must patch spark bins by running: "sed -i 's/`dirname "$0"`/`dirname "$(readlink -f $0)"`/g' /opt/apache-spark/bin/*" (readlink is in coreutils)

mtrokic commented on 2016-08-13 17:20 (UTC)

I think there are dependencies which are missing. After installing gcc-fortran and postgresql-libs I was able to compile successfully.

sidec commented on 2016-07-10 20:16 (UTC) (edited on 2016-07-10 20:19 (UTC) by sidec)

After successful makepkg -sri I fail to run spark-shell, I get this message insteed: ls: cannot access '/usr/assembly/target/scala-2.10': No such file or directory Failed to find Spark assembly in /usr/assembly/target/scala-2.10. You need to build Spark before running this program.

axelmagn commented on 2016-06-29 19:24 (UTC)

I can confirm discord's issue. I am having the same problem.

huitseeker commented on 2016-05-03 15:40 (UTC)

I have trouble reproducing your issue, sorry.

discord commented on 2016-04-26 18:31 (UTC)

Considering removing hive from the pom build and re-building, since I don't use it. However not sure why this works for anyone except myself.

discord commented on 2016-04-16 21:11 (UTC)

I resolved the previous issue by connecting to my maven repository. However I'm having build issues now: [INFO] [INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ spark-hive_2.11 --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] Copying 1 resource [INFO] Copying 3 resources [INFO] [INFO] --- scala-maven-plugin:3.2.2:compile (scala-compile-first) @ spark-hive_2.11 --- [WARNING] Failed to build parent project for org.spark-project.hive:hive-cli:jar:1.2.1.spark [WARNING] Invalid POM for org.spark-project.hive:hive-cli:jar:1.2.1.spark, transitive dependencies (if any) will not be available, enable debug logging for more details [WARNING] Failed to build parent project for org.spark-project.hive:hive-exec:jar:1.2.1.spark [WARNING] Invalid POM for org.spark-project.hive:hive-exec:jar:1.2.1.spark, transitive dependencies (if any) will not be available, enable debug logging for more details [WARNING] Failed to build parent project for org.spark-project.hive:hive-metastore:jar:1.2.1.spark [WARNING] Invalid POM for org.spark-project.hive:hive-metastore:jar:1.2.1.spark, transitive dependencies (if any) will not be available, enable debug logging for more details [INFO] Using zinc server for incremental compilation [info] Compiling 28 Scala sources and 1 Java source to /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/target/scala-2.11/classes... [warn] Class org.antlr.runtime.tree.CommonTree not found - continuing with a stub. [warn] Class org.antlr.runtime.Token not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.Tree not found - continuing with a stub. [error] Class org.antlr.runtime.tree.CommonTree not found - continuing with a stub. [warn] Class org.antlr.runtime.Parser not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.TreeAdaptor not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. .... [warn] Class org.antlr.runtime.TokenStream not found - continuing with a stub. [warn] Class org.antlr.runtime.TokenStream not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.TreeAdaptor not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.TreeAdaptor not found - continuing with a stub. [warn] Class org.antlr.runtime.IntStream not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [error] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.CommonTree not found - continuing with a stub. [warn] Class org.antlr.runtime.Parser not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.TreeAdaptor not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. ... [warn] Class org.antlr.runtime.Parser not found - continuing with a stub. [warn] Class org.antlr.runtime.TokenStream not found - continuing with a stub. [warn] Class org.antlr.runtime.TokenStream not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.TreeAdaptor not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.TreeAdaptor not found - continuing with a stub. [warn] Class org.antlr.runtime.IntStream not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [error] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. ... [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.Parser not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.TreeAdaptor not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. ... [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. [warn] Class org.antlr.runtime.Parser not found - continuing with a stub. [warn] Class org.antlr.runtime.TokenStream not found - continuing with a stub. [warn] Class org.antlr.runtime.TokenStream not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.TreeAdaptor not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.TreeAdaptor not found - continuing with a stub. [warn] Class org.antlr.runtime.IntStream not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [error] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.Parser not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.TreeAdaptor not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. [warn] Class org.antlr.runtime.Parser not found - continuing with a stub. [warn] Class org.antlr.runtime.TokenStream not found - continuing with a stub. [warn] Class org.antlr.runtime.TokenStream not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.TreeAdaptor not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.TreeAdaptor not found - continuing with a stub. [warn] Class org.antlr.runtime.IntStream not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [warn] Class org.antlr.runtime.BitSet not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [error] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognitionException not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [warn] Class org.antlr.runtime.RecognizerSharedState not found - continuing with a stub. [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:221: value token is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] n.token.asInstanceOf[org.antlr.runtime.CommonToken].setText(newText) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:221: object antlr is not a member of package org [error] n.token.asInstanceOf[org.antlr.runtime.CommonToken].setText(newText) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:229: value getChildCount is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] (1 to n.getChildCount).foreach(_ => n.deleteChild(0)) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:229: value deleteChild is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] (1 to n.getChildCount).foreach(_ => n.deleteChild(0)) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:230: value addChildren is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] n.addChildren(newChildren.asJava) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:246: value getType is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] check("type", _.getType) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:247: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] check("text", _.getText) [error] ^ [warn] Class org.antlr.runtime.TokenRewriteStream not found - continuing with a stub. [warn] Class org.antlr.runtime.TokenRewriteStream not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.TreeAdaptor not found - continuing with a stub. [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:304: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] val plan = if (nativeCommands contains tree.getText) { [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:345: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] assert(tree.asInstanceOf[ASTNode].getText == "TOK_CREATETABLE", "Only CREATE TABLE supported.") [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:349: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] .find(_.asInstanceOf[ASTNode].getText == "TOK_TABCOLLIST") [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:360: value getLine is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] CurrentOrigin.setPosition(t.getLine, t.getCharPositionInLine) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:360: value getCharPositionInLine is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] CurrentOrigin.setPosition(t.getLine, t.getCharPositionInLine) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:361: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Some((t.getText, [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:372: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] val (matches, nonMatches) = remainingNodes.partition(_.getText.toUpperCase == clauseName) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:373: value ++ is not a member of Any [error] remainingNodes = nonMatches ++ (if (matches.nonEmpty) matches.tail else Nil) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:373: value nonEmpty is not a member of Any [error] remainingNodes = nonMatches ++ (if (matches.nonEmpty) matches.tail else Nil) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:373: value tail is not a member of Any [error] remainingNodes = nonMatches ++ (if (matches.nonEmpty) matches.tail else Nil) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:374: value headOption is not a member of Any [error] matches.headOption [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:390: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] nodeList.filter { case ast: ASTNode => ast.getText == clauseName } match { [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:407: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] DecimalType(precision.getText.toInt, scale.getText.toInt) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:407: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] DecimalType(precision.getText.toInt, scale.getText.toInt) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:409: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] DecimalType(precision.getText.toInt, 0) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:480: value toStringTree is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] val keyMap = keyASTs.map(_.toStringTree).zipWithIndex.toMap [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:480: Cannot construct a collection of type That with elements of type (A1, Int) based on a collection of type Seq[Nothing]. [error] val keyMap = keyASTs.map(_.toStringTree).zipWithIndex.toMap [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:486: value toStringTree is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] val colString = col.asInstanceOf[ASTNode].toStringTree() [error] ^ [error] Class org.antlr.runtime.TokenRewriteStream not found - continuing with a stub. [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:565: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] if noExplainCommands.contains(explainArgs.head.getText) => [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:568: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] if "TOK_CREATETABLE" == explainArgs.head.getText => [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:606: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] UnresolvedRelation(TableIdentifier(tableName.getText), None), [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:672: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] val comment = BaseSemanticAnalyzer.unescapeSQLString(child.getText) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:752: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] val comment = BaseSemanticAnalyzer.unescapeSQLString(child.getText) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:767: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] val fieldDelim = BaseSemanticAnalyzer.unescapeSQLString (rowChild1.getText()) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:771: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] val fieldEscape = BaseSemanticAnalyzer.unescapeSQLString (rowChild2(0).getText) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:775: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] val collItemDelim = BaseSemanticAnalyzer.unescapeSQLString(rowChild.getText) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:778: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] val mapKeyDelim = BaseSemanticAnalyzer.unescapeSQLString(rowChild.getText) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:781: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] val lineDelim = BaseSemanticAnalyzer.unescapeSQLString(rowChild.getText) [error] ^ [warn] Class org.antlr.runtime.tree.Tree not found - continuing with a stub. [warn] Class org.antlr.runtime.tree.Tree not found - continuing with a stub. [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:790: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] val nullFormat = BaseSemanticAnalyzer.unescapeSQLString(rowChild.getText) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:797: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] var location = BaseSemanticAnalyzer.unescapeSQLString(child.getText) [error] ^ [warn] Class org.json.JSONObject not found - continuing with a stub. [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:802: value getChild is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] serde = Option(BaseSemanticAnalyzer.unescapeSQLString(child.getChild(0).getText))) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:803: value getChildCount is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] if (child.getChildCount == 2) { [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:806: value getChild is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] (child.getChild(1).getChild(0)).asInstanceOf[ASTNode], serdeParams) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:811: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] child.getText().toLowerCase(Locale.ENGLISH) match { [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:866: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] s"Unrecognized file format in STORED AS clause: ${child.getText}") [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:885: value getChild is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Option(BaseSemanticAnalyzer.unescapeSQLString(list.getChild(0).getText)), [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:887: value getChild is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Option(BaseSemanticAnalyzer.unescapeSQLString(list.getChild(1).getText))) [error] ^ [error] Class org.antlr.runtime.tree.Tree not found - continuing with a stub. [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:903: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] if Seq("TOK_FROM", "TOK_INSERT").contains(queryArgs.head.getText) => [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1085: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] .asInstanceOf[ASTNode].getText [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1232: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] .asInstanceOf[ASTNode].getText [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1247: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] if (clauses.last.getText.startsWith("TOK")) { [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1289: value getType is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] s"""No parse rules for sampling clause: ${a.getType}, text: ${a.getText} : [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1289: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] s"""No parse rules for sampling clause: ${a.getType}, text: ${a.getText} : [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1297: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] case (arg, i) => arg.getText == "TOK_TABREF" [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1541: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Cast(nodeToExpr(arg), DecimalType(precision.getText.toInt, scale.getText.toInt)) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1541: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Cast(nodeToExpr(arg), DecimalType(precision.getText.toInt, scale.getText.toInt)) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1543: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Cast(nodeToExpr(arg), DecimalType(precision.getText.toInt, 0)) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1652: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Literal(strings.map(s => BaseSemanticAnalyzer.unescapeSQLString(s.getText)).mkString) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1656: value getType is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] case ast: ASTNode if numericAstTypes contains ast.getType => [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1659: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] if (ast.getText.endsWith("L")) { [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1661: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] v = Literal.create(ast.getText.substring(0, ast.getText.length() - 1).toLong, LongType) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1661: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] v = Literal.create(ast.getText.substring(0, ast.getText.length() - 1).toLong, LongType) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1662: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] } else if (ast.getText.endsWith("S")) { [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1664: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] v = Literal.create(ast.getText.substring(0, ast.getText.length() - 1).toShort, ShortType) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1664: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] v = Literal.create(ast.getText.substring(0, ast.getText.length() - 1).toShort, ShortType) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1665: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] } else if (ast.getText.endsWith("Y")) { [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1667: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] v = Literal.create(ast.getText.substring(0, ast.getText.length() - 1).toByte, ByteType) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1667: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] v = Literal.create(ast.getText.substring(0, ast.getText.length() - 1).toByte, ByteType) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1668: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] } else if (ast.getText.endsWith("BD") || ast.getText.endsWith("D")) { [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1668: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] } else if (ast.getText.endsWith("BD") || ast.getText.endsWith("D")) { [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1670: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] val strVal = ast.getText.stripSuffix("D").stripSuffix("B") [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1673: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] v = Literal.create(ast.getText.toDouble, DoubleType) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1674: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] v = Literal.create(ast.getText.toLong, LongType) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1675: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] v = Literal.create(ast.getText.toInt, IntegerType) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1682: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] sys.error(s"Failed to parse number '${ast.getText}'.") [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1687: value getType is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] case ast: ASTNode if ast.getType == HiveParser.StringLiteral => [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1688: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Literal(BaseSemanticAnalyzer.unescapeSQLString(ast.getText)) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1690: value getType is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] case ast: ASTNode if ast.getType == HiveParser.TOK_DATELITERAL => [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1691: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Literal(Date.valueOf(ast.getText.substring(1, ast.getText.length - 1))) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1691: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Literal(Date.valueOf(ast.getText.substring(1, ast.getText.length - 1))) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1693: value getType is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] case ast: ASTNode if ast.getType == HiveParser.TOK_CHARSETLITERAL => [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1694: value getChild is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Literal(BaseSemanticAnalyzer.charSetString(ast.getChild(0).getText, ast.getChild(1).getText)) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1694: value getChild is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Literal(BaseSemanticAnalyzer.charSetString(ast.getChild(0).getText, ast.getChild(1).getText)) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1696: value getType is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] case ast: ASTNode if ast.getType == HiveParser.TOK_INTERVAL_YEAR_MONTH_LITERAL => [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1697: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Literal(CalendarInterval.fromYearMonthString(ast.getText)) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1699: value getType is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] case ast: ASTNode if ast.getType == HiveParser.TOK_INTERVAL_DAY_TIME_LITERAL => [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1700: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Literal(CalendarInterval.fromDayTimeString(ast.getText)) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1702: value getType is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] case ast: ASTNode if ast.getType == HiveParser.TOK_INTERVAL_YEAR_LITERAL => [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1703: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Literal(CalendarInterval.fromSingleUnitString("year", ast.getText)) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1705: value getType is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] case ast: ASTNode if ast.getType == HiveParser.TOK_INTERVAL_MONTH_LITERAL => [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1706: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Literal(CalendarInterval.fromSingleUnitString("month", ast.getText)) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1708: value getType is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] case ast: ASTNode if ast.getType == HiveParser.TOK_INTERVAL_DAY_LITERAL => [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1709: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Literal(CalendarInterval.fromSingleUnitString("day", ast.getText)) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1711: value getType is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] case ast: ASTNode if ast.getType == HiveParser.TOK_INTERVAL_HOUR_LITERAL => [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1712: value getText is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] Literal(CalendarInterval.fromSingleUnitString("hour", ast.getText)) [error] ^ [error] /home/colin/build/AUR/apache-spark/src/spark-1.6.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveQl.scala:1714: value getType is not a member of org.apache.hadoop.hive.ql.parse.ASTNode [error] case ast: ASTNode if ast.getType == HiveParser.TOK_INTERVAL_MINUTE_LITERAL => [error] ^ [warn] 1778 warnings found [error] 112 errors found [error] Compile failed at Apr 16, 2016 4:51:19 AM [3.663s] [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Spark Project Parent POM ........................... SUCCESS [03:14 min] [INFO] Spark Project Test Tags ............................ SUCCESS [01:00 min] [INFO] Spark Project Launcher ............................. SUCCESS [01:22 min] [INFO] Spark Project Networking ........................... SUCCESS [ 13.477 s] [INFO] Spark Project Shuffle Streaming Service ............ SUCCESS [ 9.245 s] [INFO] Spark Project Unsafe ............................... SUCCESS [ 14.595 s] [INFO] Spark Project Core ................................. SUCCESS [04:17 min] [INFO] Spark Project Bagel ................................ SUCCESS [ 4.719 s] [INFO] Spark Project GraphX ............................... SUCCESS [ 22.595 s] [INFO] Spark Project Streaming ............................ SUCCESS [ 37.540 s] [INFO] Spark Project Catalyst ............................. SUCCESS [ 51.130 s] [INFO] Spark Project SQL .................................. SUCCESS [01:16 min] [INFO] Spark Project ML Library ........................... SUCCESS [02:02 min] [INFO] Spark Project Tools ................................ SUCCESS [ 1.967 s] [INFO] Spark Project Hive ................................. FAILURE [ 55.443 s] [INFO] Spark Project Docker Integration Tests ............. SKIPPED [INFO] Spark Project REPL ................................. SKIPPED [INFO] Spark Project YARN Shuffle Service ................. SKIPPED [INFO] Spark Project YARN ................................. SKIPPED [INFO] Spark Project Assembly ............................. SKIPPED [INFO] Spark Project External Twitter ..................... SKIPPED [INFO] Spark Project External Flume Sink .................. SKIPPED [INFO] Spark Project External Flume ....................... SKIPPED [INFO] Spark Project External Flume Assembly .............. SKIPPED [INFO] Spark Project External MQTT ........................ SKIPPED [INFO] Spark Project External MQTT Assembly ............... SKIPPED [INFO] Spark Project External ZeroMQ ...................... SKIPPED [INFO] Spark Project External Kafka ....................... SKIPPED [INFO] Spark Project Examples ............................. SKIPPED [INFO] Spark Project External Kafka Assembly .............. SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 16:46 min [INFO] Finished at: 2016-04-16T04:51:19+00:00 [INFO] Final Memory: 58M/529M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal net.alchim31.maven:scala-maven-plugin:3.2.2:compile (scala-compile-first) on project spark-hive_2.11: Execution scala-compile-first of goal net.alchim31.maven:scala-maven-plugin:3.2.2:compile failed. CompileFailed -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/PluginExecutionException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn <goals> -rf :spark-hive_2.11 ==> ERROR: A failure occurred in build(). Aborting...