org.scalatest.exceptions.TestFailedException: spark-submit returned with exit code 1. Command line: './bin/spark-submit' '--name' 'prepare testing tables' '--master' 'local[2]' '--conf' 'spark.ui.enabled=false' '--conf' 'spark.master.rest.enabled=false' '--conf' 'spark.sql.warehouse.dir=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/target/tmp/warehouse-70e11e30-0b52-43d3-bfc6-ac59eeac4cf9' '--conf' 'spark.sql.test.version.index=1' '--driver-java-options' '-Dderby.system.home=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/target/tmp/warehouse-70e11e30-0b52-43d3-bfc6-ac59eeac4cf9' '/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/target/tmp/test1907736328509388370.py' 2019-03-12 15:27:57.163 - stderr> Exception in thread "main" javax.xml.parsers.FactoryConfigurationError: Provider for class javax.xml.parsers.DocumentBuilderFactory cannot be created 2019-03-12 15:27:57.163 - stderr> at javax.xml.parsers.FactoryFinder.findServiceProvider(FactoryFinder.java:311) 2019-03-12 15:27:57.163 - stderr> at javax.xml.parsers.FactoryFinder.find(FactoryFinder.java:267) 2019-03-12 15:27:57.163 - stderr> at javax.xml.parsers.DocumentBuilderFactory.newInstance(DocumentBuilderFactory.java:120) 2019-03-12 15:27:57.163 - stderr> at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2516) 2019-03-12 15:27:57.163 - stderr> at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2492) 2019-03-12 15:27:57.163 - stderr> at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2405) 2019-03-12 15:27:57.163 - stderr> at org.apache.hadoop.conf.Configuration.set(Configuration.java:1143) 2019-03-12 15:27:57.166 - stderr> at org.apache.hadoop.conf.Configuration.set(Configuration.java:1115) 2019-03-12 15:27:57.166 - stderr> at org.apache.spark.deploy.SparkHadoopUtil$.org$apache$spark$deploy$SparkHadoopUtil$$appendS3AndSparkHadoopConfigurations(SparkHadoopUtil.scala:464) 2019-03-12 15:27:57.167 - stderr> at org.apache.spark.deploy.SparkHadoopUtil$.newConfiguration(SparkHadoopUtil.scala:436) 2019-03-12 15:27:57.167 - stderr> at org.apache.spark.deploy.SparkSubmit$$anonfun$2.apply(SparkSubmit.scala:334) 2019-03-12 15:27:57.167 - stderr> at org.apache.spark.deploy.SparkSubmit$$anonfun$2.apply(SparkSubmit.scala:334) 2019-03-12 15:27:57.167 - stderr> at scala.Option.getOrElse(Option.scala:121) 2019-03-12 15:27:57.168 - stderr> at org.apache.spark.deploy.SparkSubmit.prepareSubmitEnvironment(SparkSubmit.scala:334) 2019-03-12 15:27:57.168 - stderr> at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:143) 2019-03-12 15:27:57.168 - stderr> at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86) 2019-03-12 15:27:57.168 - stderr> at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924) 2019-03-12 15:27:57.169 - stderr> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933) 2019-03-12 15:27:57.169 - stderr> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) 2019-03-12 15:27:57.169 - stderr> Caused by: java.lang.RuntimeException: Provider for class javax.xml.parsers.DocumentBuilderFactory cannot be created 2019-03-12 15:27:57.17 - stderr> at javax.xml.parsers.FactoryFinder.findServiceProvider(FactoryFinder.java:308) 2019-03-12 15:27:57.17 - stderr> ... 18 more 2019-03-12 15:27:57.171 - stderr> Caused by: java.util.ServiceConfigurationError: javax.xml.parsers.DocumentBuilderFactory: Error reading configuration file 2019-03-12 15:27:57.171 - stderr> at java.util.ServiceLoader.fail(ServiceLoader.java:232) 2019-03-12 15:27:57.171 - stderr> at java.util.ServiceLoader.parse(ServiceLoader.java:309) 2019-03-12 15:27:57.172 - stderr> at java.util.ServiceLoader.access$200(ServiceLoader.java:185) 2019-03-12 15:27:57.172 - stderr> at java.util.ServiceLoader$LazyIterator.hasNextService(ServiceLoader.java:357) 2019-03-12 15:27:57.172 - stderr> at java.util.ServiceLoader$LazyIterator.hasNext(ServiceLoader.java:393) 2019-03-12 15:27:57.172 - stderr> at java.util.ServiceLoader$1.hasNext(ServiceLoader.java:474) 2019-03-12 15:27:57.173 - stderr> at javax.xml.parsers.FactoryFinder$1.run(FactoryFinder.java:293) 2019-03-12 15:27:57.173 - stderr> at java.security.AccessController.doPrivileged(Native Method) 2019-03-12 15:27:57.174 - stderr> at javax.xml.parsers.FactoryFinder.findServiceProvider(FactoryFinder.java:289) 2019-03-12 15:27:57.174 - stderr> ... 18 more 2019-03-12 15:27:57.175 - stderr> Caused by: java.io.FileNotFoundException: /tmp/test-spark/spark-2.4.0/jars/xercesImpl-2.9.1.jar (No such file or directory) 2019-03-12 15:27:57.175 - stderr> at java.util.zip.ZipFile.open(Native Method) 2019-03-12 15:27:57.175 - stderr> at java.util.zip.ZipFile.<init>(ZipFile.java:225) 2019-03-12 15:27:57.175 - stderr> at java.util.zip.ZipFile.<init>(ZipFile.java:155) 2019-03-12 15:27:57.176 - stderr> at java.util.jar.JarFile.<init>(JarFile.java:166) 2019-03-12 15:27:57.176 - stderr> at java.util.jar.JarFile.<init>(JarFile.java:103) 2019-03-12 15:27:57.176 - stderr> at sun.net.www.protocol.jar.URLJarFile.<init>(URLJarFile.java:93) 2019-03-12 15:27:57.176 - stderr> at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69) 2019-03-12 15:27:57.177 - stderr> at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:84) 2019-03-12 15:27:57.177 - stderr> at sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122) 2019-03-12 15:27:57.177 - stderr> at sun.net.www.protocol.jar.JarURLConnection.getInputStream(JarURLConnection.java:152) 2019-03-12 15:27:57.177 - stderr> at java.net.URL.openStream(URL.java:1045) 2019-03-12 15:27:57.177 - stderr> at java.util.ServiceLoader.parse(ServiceLoader.java:304) 2019-03-12 15:27:57.178 - stderr> ... 25 more

sbt.ForkMain$ForkError: org.scalatest.exceptions.TestFailedException: spark-submit returned with exit code 1.
Command line: './bin/spark-submit' '--name' 'prepare testing tables' '--master' 'local[2]' '--conf' 'spark.ui.enabled=false' '--conf' 'spark.master.rest.enabled=false' '--conf' 'spark.sql.warehouse.dir=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/target/tmp/warehouse-70e11e30-0b52-43d3-bfc6-ac59eeac4cf9' '--conf' 'spark.sql.test.version.index=1' '--driver-java-options' '-Dderby.system.home=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/target/tmp/warehouse-70e11e30-0b52-43d3-bfc6-ac59eeac4cf9' '/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/target/tmp/test1907736328509388370.py'

2019-03-12 15:27:57.163 - stderr> Exception in thread "main" javax.xml.parsers.FactoryConfigurationError: Provider for class javax.xml.parsers.DocumentBuilderFactory cannot be created
2019-03-12 15:27:57.163 - stderr> 	at javax.xml.parsers.FactoryFinder.findServiceProvider(FactoryFinder.java:311)
2019-03-12 15:27:57.163 - stderr> 	at javax.xml.parsers.FactoryFinder.find(FactoryFinder.java:267)
2019-03-12 15:27:57.163 - stderr> 	at javax.xml.parsers.DocumentBuilderFactory.newInstance(DocumentBuilderFactory.java:120)
2019-03-12 15:27:57.163 - stderr> 	at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2516)
2019-03-12 15:27:57.163 - stderr> 	at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2492)
2019-03-12 15:27:57.163 - stderr> 	at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2405)
2019-03-12 15:27:57.163 - stderr> 	at org.apache.hadoop.conf.Configuration.set(Configuration.java:1143)
2019-03-12 15:27:57.166 - stderr> 	at org.apache.hadoop.conf.Configuration.set(Configuration.java:1115)
2019-03-12 15:27:57.166 - stderr> 	at org.apache.spark.deploy.SparkHadoopUtil$.org$apache$spark$deploy$SparkHadoopUtil$$appendS3AndSparkHadoopConfigurations(SparkHadoopUtil.scala:464)
2019-03-12 15:27:57.167 - stderr> 	at org.apache.spark.deploy.SparkHadoopUtil$.newConfiguration(SparkHadoopUtil.scala:436)
2019-03-12 15:27:57.167 - stderr> 	at org.apache.spark.deploy.SparkSubmit$$anonfun$2.apply(SparkSubmit.scala:334)
2019-03-12 15:27:57.167 - stderr> 	at org.apache.spark.deploy.SparkSubmit$$anonfun$2.apply(SparkSubmit.scala:334)
2019-03-12 15:27:57.167 - stderr> 	at scala.Option.getOrElse(Option.scala:121)
2019-03-12 15:27:57.168 - stderr> 	at org.apache.spark.deploy.SparkSubmit.prepareSubmitEnvironment(SparkSubmit.scala:334)
2019-03-12 15:27:57.168 - stderr> 	at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:143)
2019-03-12 15:27:57.168 - stderr> 	at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
2019-03-12 15:27:57.168 - stderr> 	at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924)
2019-03-12 15:27:57.169 - stderr> 	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933)
2019-03-12 15:27:57.169 - stderr> 	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
2019-03-12 15:27:57.169 - stderr> Caused by: java.lang.RuntimeException: Provider for class javax.xml.parsers.DocumentBuilderFactory cannot be created
2019-03-12 15:27:57.17 - stderr> 	at javax.xml.parsers.FactoryFinder.findServiceProvider(FactoryFinder.java:308)
2019-03-12 15:27:57.17 - stderr> 	... 18 more
2019-03-12 15:27:57.171 - stderr> Caused by: java.util.ServiceConfigurationError: javax.xml.parsers.DocumentBuilderFactory: Error reading configuration file
2019-03-12 15:27:57.171 - stderr> 	at java.util.ServiceLoader.fail(ServiceLoader.java:232)
2019-03-12 15:27:57.171 - stderr> 	at java.util.ServiceLoader.parse(ServiceLoader.java:309)
2019-03-12 15:27:57.172 - stderr> 	at java.util.ServiceLoader.access$200(ServiceLoader.java:185)
2019-03-12 15:27:57.172 - stderr> 	at java.util.ServiceLoader$LazyIterator.hasNextService(ServiceLoader.java:357)
2019-03-12 15:27:57.172 - stderr> 	at java.util.ServiceLoader$LazyIterator.hasNext(ServiceLoader.java:393)
2019-03-12 15:27:57.172 - stderr> 	at java.util.ServiceLoader$1.hasNext(ServiceLoader.java:474)
2019-03-12 15:27:57.173 - stderr> 	at javax.xml.parsers.FactoryFinder$1.run(FactoryFinder.java:293)
2019-03-12 15:27:57.173 - stderr> 	at java.security.AccessController.doPrivileged(Native Method)
2019-03-12 15:27:57.174 - stderr> 	at javax.xml.parsers.FactoryFinder.findServiceProvider(FactoryFinder.java:289)
2019-03-12 15:27:57.174 - stderr> 	... 18 more
2019-03-12 15:27:57.175 - stderr> Caused by: java.io.FileNotFoundException: /tmp/test-spark/spark-2.4.0/jars/xercesImpl-2.9.1.jar (No such file or directory)
2019-03-12 15:27:57.175 - stderr> 	at java.util.zip.ZipFile.open(Native Method)
2019-03-12 15:27:57.175 - stderr> 	at java.util.zip.ZipFile.<init>(ZipFile.java:225)
2019-03-12 15:27:57.175 - stderr> 	at java.util.zip.ZipFile.<init>(ZipFile.java:155)
2019-03-12 15:27:57.176 - stderr> 	at java.util.jar.JarFile.<init>(JarFile.java:166)
2019-03-12 15:27:57.176 - stderr> 	at java.util.jar.JarFile.<init>(JarFile.java:103)
2019-03-12 15:27:57.176 - stderr> 	at sun.net.www.protocol.jar.URLJarFile.<init>(URLJarFile.java:93)
2019-03-12 15:27:57.176 - stderr> 	at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69)
2019-03-12 15:27:57.177 - stderr> 	at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:84)
2019-03-12 15:27:57.177 - stderr> 	at sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122)
2019-03-12 15:27:57.177 - stderr> 	at sun.net.www.protocol.jar.JarURLConnection.getInputStream(JarURLConnection.java:152)
2019-03-12 15:27:57.177 - stderr> 	at java.net.URL.openStream(URL.java:1045)
2019-03-12 15:27:57.177 - stderr> 	at java.util.ServiceLoader.parse(ServiceLoader.java:304)
2019-03-12 15:27:57.178 - stderr> 	... 25 more
           
	at org.scalatest.Assertions.newAssertionFailedException(Assertions.scala:528)
	at org.scalatest.Assertions.newAssertionFailedException$(Assertions.scala:527)
	at org.scalatest.FunSuite.newAssertionFailedException(FunSuite.scala:1560)
	at org.scalatest.Assertions.fail(Assertions.scala:1089)
	at org.scalatest.Assertions.fail$(Assertions.scala:1085)
	at org.scalatest.FunSuite.fail(FunSuite.scala:1560)
	at org.apache.spark.sql.hive.SparkSubmitTestUtils.runSparkSubmit(SparkSubmitTestUtils.scala:94)
	at org.apache.spark.sql.hive.SparkSubmitTestUtils.runSparkSubmit$(SparkSubmitTestUtils.scala:41)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.runSparkSubmit(HiveExternalCatalogVersionsSuite.scala:43)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.$anonfun$beforeAll$1(HiveExternalCatalogVersionsSuite.scala:187)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.$anonfun$beforeAll$1$adapted(HiveExternalCatalogVersionsSuite.scala:172)
	at scala.collection.immutable.List.foreach(List.scala:392)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.beforeAll(HiveExternalCatalogVersionsSuite.scala:172)
	at org.scalatest.BeforeAndAfterAll.liftedTree1$1(BeforeAndAfterAll.scala:212)
	at org.scalatest.BeforeAndAfterAll.run(BeforeAndAfterAll.scala:210)
	at org.scalatest.BeforeAndAfterAll.run$(BeforeAndAfterAll.scala:208)
	at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:53)
	at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:314)
	at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:507)
	at sbt.ForkMain$Run$2.call(ForkMain.java:296)
	at sbt.ForkMain$Run$2.call(ForkMain.java:286)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)