org.scalatest.exceptions.TestFailedException: spark-submit returned with exit code 1. Command line: './bin/spark-submit' '--name' 'prepare testing tables' '--master' 'local[2]' '--conf' 'spark.ui.enabled=false' '--conf' 'spark.master.rest.enabled=false' '--conf' 'spark.sql.hive.metastore.version=1.2.1' '--conf' 'spark.sql.hive.metastore.jars=maven' '--conf' 'spark.sql.warehouse.dir=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-33022bdf-15f2-47cb-a690-d1d48e7ed427' '--conf' 'spark.sql.test.version.index=2' '--driver-java-options' '-Dderby.system.home=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-33022bdf-15f2-47cb-a690-d1d48e7ed427' '/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test5233324416890106231.py' 2019-09-02 19:32:57.354 - stderr> 19/09/02 19:32:57 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2019-09-02 19:32:57.713 - stdout> Traceback (most recent call last): 2019-09-02 19:32:57.713 - stdout> File "/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test5233324416890106231.py", line 2, in <module> 2019-09-02 19:32:57.713 - stdout> from pyspark.sql import SparkSession 2019-09-02 19:32:57.713 - stdout> File "<frozen importlib._bootstrap>", line 971, in _find_and_load 2019-09-02 19:32:57.713 - stdout> File "<frozen importlib._bootstrap>", line 955, in _find_and_load_unlocked 2019-09-02 19:32:57.714 - stdout> File "<frozen importlib._bootstrap>", line 656, in _load_unlocked 2019-09-02 19:32:57.714 - stdout> File "<frozen importlib._bootstrap>", line 626, in _load_backward_compatible 2019-09-02 19:32:57.714 - stdout> File "/tmp/test-spark/spark-2.4.4/python/lib/pyspark.zip/pyspark/__init__.py", line 51, in <module> 2019-09-02 19:32:57.714 - stdout> File "<frozen importlib._bootstrap>", line 971, in _find_and_load 2019-09-02 19:32:57.714 - stdout> File "<frozen importlib._bootstrap>", line 955, in _find_and_load_unlocked 2019-09-02 19:32:57.714 - stdout> File "<frozen importlib._bootstrap>", line 656, in _load_unlocked 2019-09-02 19:32:57.714 - stdout> File "<frozen importlib._bootstrap>", line 626, in _load_backward_compatible 2019-09-02 19:32:57.714 - stdout> File "/tmp/test-spark/spark-2.4.4/python/lib/pyspark.zip/pyspark/context.py", line 29, in <module> 2019-09-02 19:32:57.714 - stdout> File "<frozen importlib._bootstrap>", line 971, in _find_and_load 2019-09-02 19:32:57.714 - stdout> File "<frozen importlib._bootstrap>", line 951, in _find_and_load_unlocked 2019-09-02 19:32:57.714 - stdout> File "<frozen importlib._bootstrap>", line 894, in _find_spec 2019-09-02 19:32:57.714 - stdout> File "<frozen importlib._bootstrap_external>", line 1157, in find_spec 2019-09-02 19:32:57.715 - stdout> File "<frozen importlib._bootstrap_external>", line 1131, in _get_spec 2019-09-02 19:32:57.715 - stdout> File "<frozen importlib._bootstrap_external>", line 1112, in _legacy_get_spec 2019-09-02 19:32:57.715 - stdout> File "<frozen importlib._bootstrap>", line 441, in spec_from_loader 2019-09-02 19:32:57.715 - stdout> File "<frozen importlib._bootstrap_external>", line 544, in spec_from_file_location 2019-09-02 19:32:57.715 - stdout> FileNotFoundError: [Errno 2] No such file or directory: '/tmp/test-spark/spark-2.4.4/python/lib/py4j-0.10.7-src.zip' 2019-09-02 19:32:57.726 - stderr> log4j:WARN No appenders could be found for logger (org.apache.spark.util.ShutdownHookManager). 2019-09-02 19:32:57.726 - stderr> log4j:WARN Please initialize the log4j system properly. 2019-09-02 19:32:57.726 - stderr> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.

sbt.ForkMain$ForkError: org.scalatest.exceptions.TestFailedException: spark-submit returned with exit code 1.
Command line: './bin/spark-submit' '--name' 'prepare testing tables' '--master' 'local[2]' '--conf' 'spark.ui.enabled=false' '--conf' 'spark.master.rest.enabled=false' '--conf' 'spark.sql.hive.metastore.version=1.2.1' '--conf' 'spark.sql.hive.metastore.jars=maven' '--conf' 'spark.sql.warehouse.dir=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-33022bdf-15f2-47cb-a690-d1d48e7ed427' '--conf' 'spark.sql.test.version.index=2' '--driver-java-options' '-Dderby.system.home=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-33022bdf-15f2-47cb-a690-d1d48e7ed427' '/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test5233324416890106231.py'

2019-09-02 19:32:57.354 - stderr> 19/09/02 19:32:57 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
2019-09-02 19:32:57.713 - stdout> Traceback (most recent call last):
2019-09-02 19:32:57.713 - stdout>   File "/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test5233324416890106231.py", line 2, in <module>
2019-09-02 19:32:57.713 - stdout>     from pyspark.sql import SparkSession
2019-09-02 19:32:57.713 - stdout>   File "<frozen importlib._bootstrap>", line 971, in _find_and_load
2019-09-02 19:32:57.713 - stdout>   File "<frozen importlib._bootstrap>", line 955, in _find_and_load_unlocked
2019-09-02 19:32:57.714 - stdout>   File "<frozen importlib._bootstrap>", line 656, in _load_unlocked
2019-09-02 19:32:57.714 - stdout>   File "<frozen importlib._bootstrap>", line 626, in _load_backward_compatible
2019-09-02 19:32:57.714 - stdout>   File "/tmp/test-spark/spark-2.4.4/python/lib/pyspark.zip/pyspark/__init__.py", line 51, in <module>
2019-09-02 19:32:57.714 - stdout>   File "<frozen importlib._bootstrap>", line 971, in _find_and_load
2019-09-02 19:32:57.714 - stdout>   File "<frozen importlib._bootstrap>", line 955, in _find_and_load_unlocked
2019-09-02 19:32:57.714 - stdout>   File "<frozen importlib._bootstrap>", line 656, in _load_unlocked
2019-09-02 19:32:57.714 - stdout>   File "<frozen importlib._bootstrap>", line 626, in _load_backward_compatible
2019-09-02 19:32:57.714 - stdout>   File "/tmp/test-spark/spark-2.4.4/python/lib/pyspark.zip/pyspark/context.py", line 29, in <module>
2019-09-02 19:32:57.714 - stdout>   File "<frozen importlib._bootstrap>", line 971, in _find_and_load
2019-09-02 19:32:57.714 - stdout>   File "<frozen importlib._bootstrap>", line 951, in _find_and_load_unlocked
2019-09-02 19:32:57.714 - stdout>   File "<frozen importlib._bootstrap>", line 894, in _find_spec
2019-09-02 19:32:57.714 - stdout>   File "<frozen importlib._bootstrap_external>", line 1157, in find_spec
2019-09-02 19:32:57.715 - stdout>   File "<frozen importlib._bootstrap_external>", line 1131, in _get_spec
2019-09-02 19:32:57.715 - stdout>   File "<frozen importlib._bootstrap_external>", line 1112, in _legacy_get_spec
2019-09-02 19:32:57.715 - stdout>   File "<frozen importlib._bootstrap>", line 441, in spec_from_loader
2019-09-02 19:32:57.715 - stdout>   File "<frozen importlib._bootstrap_external>", line 544, in spec_from_file_location
2019-09-02 19:32:57.715 - stdout> FileNotFoundError: [Errno 2] No such file or directory: '/tmp/test-spark/spark-2.4.4/python/lib/py4j-0.10.7-src.zip'
2019-09-02 19:32:57.726 - stderr> log4j:WARN No appenders could be found for logger (org.apache.spark.util.ShutdownHookManager).
2019-09-02 19:32:57.726 - stderr> log4j:WARN Please initialize the log4j system properly.
2019-09-02 19:32:57.726 - stderr> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.
           
	at org.scalatest.Assertions.newAssertionFailedException(Assertions.scala:528)
	at org.scalatest.Assertions.newAssertionFailedException$(Assertions.scala:527)
	at org.scalatest.FunSuite.newAssertionFailedException(FunSuite.scala:1560)
	at org.scalatest.Assertions.fail(Assertions.scala:1089)
	at org.scalatest.Assertions.fail$(Assertions.scala:1085)
	at org.scalatest.FunSuite.fail(FunSuite.scala:1560)
	at org.apache.spark.sql.hive.SparkSubmitTestUtils.runSparkSubmit(SparkSubmitTestUtils.scala:94)
	at org.apache.spark.sql.hive.SparkSubmitTestUtils.runSparkSubmit$(SparkSubmitTestUtils.scala:41)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.runSparkSubmit(HiveExternalCatalogVersionsSuite.scala:50)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.$anonfun$prepare$1(HiveExternalCatalogVersionsSuite.scala:200)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.$anonfun$prepare$1$adapted(HiveExternalCatalogVersionsSuite.scala:183)
	at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
	at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
	at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.prepare(HiveExternalCatalogVersionsSuite.scala:183)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.beforeAll(HiveExternalCatalogVersionsSuite.scala:209)
	at org.scalatest.BeforeAndAfterAll.liftedTree1$1(BeforeAndAfterAll.scala:212)
	at org.scalatest.BeforeAndAfterAll.run(BeforeAndAfterAll.scala:210)
	at org.scalatest.BeforeAndAfterAll.run$(BeforeAndAfterAll.scala:208)
	at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:56)
	at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:314)
	at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:507)
	at sbt.ForkMain$Run$2.call(ForkMain.java:296)
	at sbt.ForkMain$Run$2.call(ForkMain.java:286)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)