org.scalatest.exceptions.TestFailedException: spark-submit returned with exit code 1. Command line: './bin/spark-submit' '--name' 'prepare testing tables' '--master' 'local[2]' '--conf' 'spark.ui.enabled=false' '--conf' 'spark.master.rest.enabled=false' '--conf' 'spark.sql.warehouse.dir=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.6/target/tmp/warehouse-fd2b79e3-2134-46e6-8fcc-345b3b05da9a' '--conf' 'spark.sql.test.version.index=2' '--driver-java-options' '-Dderby.system.home=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.6/target/tmp/warehouse-fd2b79e3-2134-46e6-8fcc-345b3b05da9a' '/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.6/target/tmp/test3628304127848568246.py' 2018-10-10 14:32:46.829 - stderr> log4j:ERROR Could not read configuration file from URL [jar:file:/tmp/test-spark/spark-2.3.2/jars/kubernetes-model-2.0.0.jar!/log4j.properties]. 2018-10-10 14:32:46.829 - stderr> java.io.FileNotFoundException: /tmp/test-spark/spark-2.3.2/jars/kubernetes-model-2.0.0.jar (No such file or directory) 2018-10-10 14:32:46.829 - stderr> at java.util.zip.ZipFile.open(Native Method) 2018-10-10 14:32:46.829 - stderr> at java.util.zip.ZipFile.<init>(ZipFile.java:219) 2018-10-10 14:32:46.829 - stderr> at java.util.zip.ZipFile.<init>(ZipFile.java:149) 2018-10-10 14:32:46.829 - stderr> at java.util.jar.JarFile.<init>(JarFile.java:166) 2018-10-10 14:32:46.829 - stderr> at java.util.jar.JarFile.<init>(JarFile.java:103) 2018-10-10 14:32:46.829 - stderr> at sun.net.www.protocol.jar.URLJarFile.<init>(URLJarFile.java:93) 2018-10-10 14:32:46.829 - stderr> at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69) 2018-10-10 14:32:46.829 - stderr> at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:99) 2018-10-10 14:32:46.829 - stderr> at sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122) 2018-10-10 14:32:46.829 - stderr> at sun.net.www.protocol.jar.JarURLConnection.getInputStream(JarURLConnection.java:150) 2018-10-10 14:32:46.829 - stderr> at org.apache.log4j.PropertyConfigurator.doConfigure(PropertyConfigurator.java:557) 2018-10-10 14:32:46.829 - stderr> at org.apache.log4j.helpers.OptionConverter.selectAndConfigure(OptionConverter.java:526) 2018-10-10 14:32:46.83 - stderr> at org.apache.log4j.LogManager.<clinit>(LogManager.java:127) 2018-10-10 14:32:46.83 - stderr> at org.apache.spark.internal.Logging$class.initializeLogging(Logging.scala:120) 2018-10-10 14:32:46.83 - stderr> at org.apache.spark.internal.Logging$class.initializeLogIfNecessary(Logging.scala:108) 2018-10-10 14:32:46.83 - stderr> at org.apache.spark.deploy.SparkSubmit$.initializeLogIfNecessary(SparkSubmit.scala:71) 2018-10-10 14:32:46.83 - stderr> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:128) 2018-10-10 14:32:46.83 - stderr> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) 2018-10-10 14:32:46.83 - stderr> log4j:ERROR Ignoring configuration file [jar:file:/tmp/test-spark/spark-2.3.2/jars/kubernetes-model-2.0.0.jar!/log4j.properties]. 2018-10-10 14:32:46.831 - stderr> log4j:ERROR Could not read configuration file from URL [jar:file:/tmp/test-spark/spark-2.3.2/jars/spark-core_2.11-2.3.2.jar!/org/apache/spark/log4j-defaults.properties]. 2018-10-10 14:32:46.831 - stderr> java.io.FileNotFoundException: /tmp/test-spark/spark-2.3.2/jars/spark-core_2.11-2.3.2.jar (No such file or directory) 2018-10-10 14:32:46.831 - stderr> at java.util.zip.ZipFile.open(Native Method) 2018-10-10 14:32:46.831 - stderr> at java.util.zip.ZipFile.<init>(ZipFile.java:219) 2018-10-10 14:32:46.831 - stderr> at java.util.zip.ZipFile.<init>(ZipFile.java:149) 2018-10-10 14:32:46.831 - stderr> at java.util.jar.JarFile.<init>(JarFile.java:166) 2018-10-10 14:32:46.831 - stderr> at java.util.jar.JarFile.<init>(JarFile.java:103) 2018-10-10 14:32:46.831 - stderr> at sun.net.www.protocol.jar.URLJarFile.<init>(URLJarFile.java:93) 2018-10-10 14:32:46.831 - stderr> at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69) 2018-10-10 14:32:46.831 - stderr> at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:99) 2018-10-10 14:32:46.831 - stderr> at sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122) 2018-10-10 14:32:46.831 - stderr> at sun.net.www.protocol.jar.JarURLConnection.getInputStream(JarURLConnection.java:150) 2018-10-10 14:32:46.831 - stderr> at org.apache.log4j.PropertyConfigurator.doConfigure(PropertyConfigurator.java:557) 2018-10-10 14:32:46.831 - stderr> at org.apache.log4j.PropertyConfigurator.configure(PropertyConfigurator.java:415) 2018-10-10 14:32:46.831 - stderr> at org.apache.spark.internal.Logging$class.initializeLogging(Logging.scala:127) 2018-10-10 14:32:46.831 - stderr> at org.apache.spark.internal.Logging$class.initializeLogIfNecessary(Logging.scala:108) 2018-10-10 14:32:46.831 - stderr> at org.apache.spark.deploy.SparkSubmit$.initializeLogIfNecessary(SparkSubmit.scala:71) 2018-10-10 14:32:46.832 - stderr> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:128) 2018-10-10 14:32:46.832 - stderr> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) 2018-10-10 14:32:46.832 - stderr> log4j:ERROR Ignoring configuration file [jar:file:/tmp/test-spark/spark-2.3.2/jars/spark-core_2.11-2.3.2.jar!/org/apache/spark/log4j-defaults.properties]. 2018-10-10 14:32:47.113 - stderr> Exception in thread "main" javax.xml.parsers.FactoryConfigurationError: Provider for class javax.xml.parsers.DocumentBuilderFactory cannot be created 2018-10-10 14:32:47.113 - stderr> at javax.xml.parsers.FactoryFinder.findServiceProvider(FactoryFinder.java:311) 2018-10-10 14:32:47.113 - stderr> at javax.xml.parsers.FactoryFinder.find(FactoryFinder.java:267) 2018-10-10 14:32:47.113 - stderr> at javax.xml.parsers.DocumentBuilderFactory.newInstance(DocumentBuilderFactory.java:120) 2018-10-10 14:32:47.113 - stderr> at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2516) 2018-10-10 14:32:47.113 - stderr> at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2492) 2018-10-10 14:32:47.113 - stderr> at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2405) 2018-10-10 14:32:47.113 - stderr> at org.apache.hadoop.conf.Configuration.set(Configuration.java:1143) 2018-10-10 14:32:47.113 - stderr> at org.apache.hadoop.conf.Configuration.set(Configuration.java:1115) 2018-10-10 14:32:47.113 - stderr> at org.apache.spark.deploy.SparkHadoopUtil$.org$apache$spark$deploy$SparkHadoopUtil$$appendS3AndSparkHadoopConfigurations(SparkHadoopUtil.scala:474) 2018-10-10 14:32:47.113 - stderr> at org.apache.spark.deploy.SparkHadoopUtil$.newConfiguration(SparkHadoopUtil.scala:446) 2018-10-10 14:32:47.113 - stderr> at org.apache.spark.deploy.SparkSubmit$$anonfun$1.apply(SparkSubmit.scala:383) 2018-10-10 14:32:47.113 - stderr> at org.apache.spark.deploy.SparkSubmit$$anonfun$1.apply(SparkSubmit.scala:383) 2018-10-10 14:32:47.113 - stderr> at scala.Option.getOrElse(Option.scala:121) 2018-10-10 14:32:47.113 - stderr> at org.apache.spark.deploy.SparkSubmit$.doPrepareSubmitEnvironment(SparkSubmit.scala:383) 2018-10-10 14:32:47.113 - stderr> at org.apache.spark.deploy.SparkSubmit$.prepareSubmitEnvironment(SparkSubmit.scala:250) 2018-10-10 14:32:47.113 - stderr> at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:171) 2018-10-10 14:32:47.113 - stderr> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137) 2018-10-10 14:32:47.113 - stderr> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) 2018-10-10 14:32:47.113 - stderr> Caused by: java.lang.RuntimeException: Provider for class javax.xml.parsers.DocumentBuilderFactory cannot be created 2018-10-10 14:32:47.113 - stderr> at javax.xml.parsers.FactoryFinder.findServiceProvider(FactoryFinder.java:308) 2018-10-10 14:32:47.113 - stderr> ... 17 more 2018-10-10 14:32:47.113 - stderr> Caused by: java.util.ServiceConfigurationError: javax.xml.parsers.DocumentBuilderFactory: Error reading configuration file 2018-10-10 14:32:47.114 - stderr> at java.util.ServiceLoader.fail(ServiceLoader.java:232) 2018-10-10 14:32:47.114 - stderr> at java.util.ServiceLoader.parse(ServiceLoader.java:309) 2018-10-10 14:32:47.114 - stderr> at java.util.ServiceLoader.access$200(ServiceLoader.java:185) 2018-10-10 14:32:47.114 - stderr> at java.util.ServiceLoader$LazyIterator.hasNextService(ServiceLoader.java:357) 2018-10-10 14:32:47.114 - stderr> at java.util.ServiceLoader$LazyIterator.hasNext(ServiceLoader.java:393) 2018-10-10 14:32:47.114 - stderr> at java.util.ServiceLoader$1.hasNext(ServiceLoader.java:474) 2018-10-10 14:32:47.114 - stderr> at javax.xml.parsers.FactoryFinder$1.run(FactoryFinder.java:293) 2018-10-10 14:32:47.114 - stderr> at java.security.AccessController.doPrivileged(Native Method) 2018-10-10 14:32:47.114 - stderr> at javax.xml.parsers.FactoryFinder.findServiceProvider(FactoryFinder.java:289) 2018-10-10 14:32:47.114 - stderr> ... 17 more 2018-10-10 14:32:47.114 - stderr> Caused by: java.io.FileNotFoundException: /tmp/test-spark/spark-2.3.2/jars/xercesImpl-2.9.1.jar (No such file or directory) 2018-10-10 14:32:47.114 - stderr> at java.util.zip.ZipFile.open(Native Method) 2018-10-10 14:32:47.114 - stderr> at java.util.zip.ZipFile.<init>(ZipFile.java:219) 2018-10-10 14:32:47.114 - stderr> at java.util.zip.ZipFile.<init>(ZipFile.java:149) 2018-10-10 14:32:47.114 - stderr> at java.util.jar.JarFile.<init>(JarFile.java:166) 2018-10-10 14:32:47.114 - stderr> at java.util.jar.JarFile.<init>(JarFile.java:103) 2018-10-10 14:32:47.114 - stderr> at sun.net.www.protocol.jar.URLJarFile.<init>(URLJarFile.java:93) 2018-10-10 14:32:47.114 - stderr> at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69) 2018-10-10 14:32:47.114 - stderr> at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:84) 2018-10-10 14:32:47.114 - stderr> at sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122) 2018-10-10 14:32:47.114 - stderr> at sun.net.www.protocol.jar.JarURLConnection.getInputStream(JarURLConnection.java:150) 2018-10-10 14:32:47.114 - stderr> at java.net.URL.openStream(URL.java:1038) 2018-10-10 14:32:47.114 - stderr> at java.util.ServiceLoader.parse(ServiceLoader.java:304) 2018-10-10 14:32:47.114 - stderr> ... 24 more

sbt.ForkMain$ForkError: org.scalatest.exceptions.TestFailedException: spark-submit returned with exit code 1.
Command line: './bin/spark-submit' '--name' 'prepare testing tables' '--master' 'local[2]' '--conf' 'spark.ui.enabled=false' '--conf' 'spark.master.rest.enabled=false' '--conf' 'spark.sql.warehouse.dir=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.6/target/tmp/warehouse-fd2b79e3-2134-46e6-8fcc-345b3b05da9a' '--conf' 'spark.sql.test.version.index=2' '--driver-java-options' '-Dderby.system.home=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.6/target/tmp/warehouse-fd2b79e3-2134-46e6-8fcc-345b3b05da9a' '/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.6/target/tmp/test3628304127848568246.py'

2018-10-10 14:32:46.829 - stderr> log4j:ERROR Could not read configuration file from URL [jar:file:/tmp/test-spark/spark-2.3.2/jars/kubernetes-model-2.0.0.jar!/log4j.properties].
2018-10-10 14:32:46.829 - stderr> java.io.FileNotFoundException: /tmp/test-spark/spark-2.3.2/jars/kubernetes-model-2.0.0.jar (No such file or directory)
2018-10-10 14:32:46.829 - stderr> 	at java.util.zip.ZipFile.open(Native Method)
2018-10-10 14:32:46.829 - stderr> 	at java.util.zip.ZipFile.<init>(ZipFile.java:219)
2018-10-10 14:32:46.829 - stderr> 	at java.util.zip.ZipFile.<init>(ZipFile.java:149)
2018-10-10 14:32:46.829 - stderr> 	at java.util.jar.JarFile.<init>(JarFile.java:166)
2018-10-10 14:32:46.829 - stderr> 	at java.util.jar.JarFile.<init>(JarFile.java:103)
2018-10-10 14:32:46.829 - stderr> 	at sun.net.www.protocol.jar.URLJarFile.<init>(URLJarFile.java:93)
2018-10-10 14:32:46.829 - stderr> 	at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69)
2018-10-10 14:32:46.829 - stderr> 	at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:99)
2018-10-10 14:32:46.829 - stderr> 	at sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122)
2018-10-10 14:32:46.829 - stderr> 	at sun.net.www.protocol.jar.JarURLConnection.getInputStream(JarURLConnection.java:150)
2018-10-10 14:32:46.829 - stderr> 	at org.apache.log4j.PropertyConfigurator.doConfigure(PropertyConfigurator.java:557)
2018-10-10 14:32:46.829 - stderr> 	at org.apache.log4j.helpers.OptionConverter.selectAndConfigure(OptionConverter.java:526)
2018-10-10 14:32:46.83 - stderr> 	at org.apache.log4j.LogManager.<clinit>(LogManager.java:127)
2018-10-10 14:32:46.83 - stderr> 	at org.apache.spark.internal.Logging$class.initializeLogging(Logging.scala:120)
2018-10-10 14:32:46.83 - stderr> 	at org.apache.spark.internal.Logging$class.initializeLogIfNecessary(Logging.scala:108)
2018-10-10 14:32:46.83 - stderr> 	at org.apache.spark.deploy.SparkSubmit$.initializeLogIfNecessary(SparkSubmit.scala:71)
2018-10-10 14:32:46.83 - stderr> 	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:128)
2018-10-10 14:32:46.83 - stderr> 	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
2018-10-10 14:32:46.83 - stderr> log4j:ERROR Ignoring configuration file [jar:file:/tmp/test-spark/spark-2.3.2/jars/kubernetes-model-2.0.0.jar!/log4j.properties].
2018-10-10 14:32:46.831 - stderr> log4j:ERROR Could not read configuration file from URL [jar:file:/tmp/test-spark/spark-2.3.2/jars/spark-core_2.11-2.3.2.jar!/org/apache/spark/log4j-defaults.properties].
2018-10-10 14:32:46.831 - stderr> java.io.FileNotFoundException: /tmp/test-spark/spark-2.3.2/jars/spark-core_2.11-2.3.2.jar (No such file or directory)
2018-10-10 14:32:46.831 - stderr> 	at java.util.zip.ZipFile.open(Native Method)
2018-10-10 14:32:46.831 - stderr> 	at java.util.zip.ZipFile.<init>(ZipFile.java:219)
2018-10-10 14:32:46.831 - stderr> 	at java.util.zip.ZipFile.<init>(ZipFile.java:149)
2018-10-10 14:32:46.831 - stderr> 	at java.util.jar.JarFile.<init>(JarFile.java:166)
2018-10-10 14:32:46.831 - stderr> 	at java.util.jar.JarFile.<init>(JarFile.java:103)
2018-10-10 14:32:46.831 - stderr> 	at sun.net.www.protocol.jar.URLJarFile.<init>(URLJarFile.java:93)
2018-10-10 14:32:46.831 - stderr> 	at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69)
2018-10-10 14:32:46.831 - stderr> 	at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:99)
2018-10-10 14:32:46.831 - stderr> 	at sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122)
2018-10-10 14:32:46.831 - stderr> 	at sun.net.www.protocol.jar.JarURLConnection.getInputStream(JarURLConnection.java:150)
2018-10-10 14:32:46.831 - stderr> 	at org.apache.log4j.PropertyConfigurator.doConfigure(PropertyConfigurator.java:557)
2018-10-10 14:32:46.831 - stderr> 	at org.apache.log4j.PropertyConfigurator.configure(PropertyConfigurator.java:415)
2018-10-10 14:32:46.831 - stderr> 	at org.apache.spark.internal.Logging$class.initializeLogging(Logging.scala:127)
2018-10-10 14:32:46.831 - stderr> 	at org.apache.spark.internal.Logging$class.initializeLogIfNecessary(Logging.scala:108)
2018-10-10 14:32:46.831 - stderr> 	at org.apache.spark.deploy.SparkSubmit$.initializeLogIfNecessary(SparkSubmit.scala:71)
2018-10-10 14:32:46.832 - stderr> 	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:128)
2018-10-10 14:32:46.832 - stderr> 	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
2018-10-10 14:32:46.832 - stderr> log4j:ERROR Ignoring configuration file [jar:file:/tmp/test-spark/spark-2.3.2/jars/spark-core_2.11-2.3.2.jar!/org/apache/spark/log4j-defaults.properties].
2018-10-10 14:32:47.113 - stderr> Exception in thread "main" javax.xml.parsers.FactoryConfigurationError: Provider for class javax.xml.parsers.DocumentBuilderFactory cannot be created
2018-10-10 14:32:47.113 - stderr> 	at javax.xml.parsers.FactoryFinder.findServiceProvider(FactoryFinder.java:311)
2018-10-10 14:32:47.113 - stderr> 	at javax.xml.parsers.FactoryFinder.find(FactoryFinder.java:267)
2018-10-10 14:32:47.113 - stderr> 	at javax.xml.parsers.DocumentBuilderFactory.newInstance(DocumentBuilderFactory.java:120)
2018-10-10 14:32:47.113 - stderr> 	at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2516)
2018-10-10 14:32:47.113 - stderr> 	at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2492)
2018-10-10 14:32:47.113 - stderr> 	at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2405)
2018-10-10 14:32:47.113 - stderr> 	at org.apache.hadoop.conf.Configuration.set(Configuration.java:1143)
2018-10-10 14:32:47.113 - stderr> 	at org.apache.hadoop.conf.Configuration.set(Configuration.java:1115)
2018-10-10 14:32:47.113 - stderr> 	at org.apache.spark.deploy.SparkHadoopUtil$.org$apache$spark$deploy$SparkHadoopUtil$$appendS3AndSparkHadoopConfigurations(SparkHadoopUtil.scala:474)
2018-10-10 14:32:47.113 - stderr> 	at org.apache.spark.deploy.SparkHadoopUtil$.newConfiguration(SparkHadoopUtil.scala:446)
2018-10-10 14:32:47.113 - stderr> 	at org.apache.spark.deploy.SparkSubmit$$anonfun$1.apply(SparkSubmit.scala:383)
2018-10-10 14:32:47.113 - stderr> 	at org.apache.spark.deploy.SparkSubmit$$anonfun$1.apply(SparkSubmit.scala:383)
2018-10-10 14:32:47.113 - stderr> 	at scala.Option.getOrElse(Option.scala:121)
2018-10-10 14:32:47.113 - stderr> 	at org.apache.spark.deploy.SparkSubmit$.doPrepareSubmitEnvironment(SparkSubmit.scala:383)
2018-10-10 14:32:47.113 - stderr> 	at org.apache.spark.deploy.SparkSubmit$.prepareSubmitEnvironment(SparkSubmit.scala:250)
2018-10-10 14:32:47.113 - stderr> 	at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:171)
2018-10-10 14:32:47.113 - stderr> 	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137)
2018-10-10 14:32:47.113 - stderr> 	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
2018-10-10 14:32:47.113 - stderr> Caused by: java.lang.RuntimeException: Provider for class javax.xml.parsers.DocumentBuilderFactory cannot be created
2018-10-10 14:32:47.113 - stderr> 	at javax.xml.parsers.FactoryFinder.findServiceProvider(FactoryFinder.java:308)
2018-10-10 14:32:47.113 - stderr> 	... 17 more
2018-10-10 14:32:47.113 - stderr> Caused by: java.util.ServiceConfigurationError: javax.xml.parsers.DocumentBuilderFactory: Error reading configuration file
2018-10-10 14:32:47.114 - stderr> 	at java.util.ServiceLoader.fail(ServiceLoader.java:232)
2018-10-10 14:32:47.114 - stderr> 	at java.util.ServiceLoader.parse(ServiceLoader.java:309)
2018-10-10 14:32:47.114 - stderr> 	at java.util.ServiceLoader.access$200(ServiceLoader.java:185)
2018-10-10 14:32:47.114 - stderr> 	at java.util.ServiceLoader$LazyIterator.hasNextService(ServiceLoader.java:357)
2018-10-10 14:32:47.114 - stderr> 	at java.util.ServiceLoader$LazyIterator.hasNext(ServiceLoader.java:393)
2018-10-10 14:32:47.114 - stderr> 	at java.util.ServiceLoader$1.hasNext(ServiceLoader.java:474)
2018-10-10 14:32:47.114 - stderr> 	at javax.xml.parsers.FactoryFinder$1.run(FactoryFinder.java:293)
2018-10-10 14:32:47.114 - stderr> 	at java.security.AccessController.doPrivileged(Native Method)
2018-10-10 14:32:47.114 - stderr> 	at javax.xml.parsers.FactoryFinder.findServiceProvider(FactoryFinder.java:289)
2018-10-10 14:32:47.114 - stderr> 	... 17 more
2018-10-10 14:32:47.114 - stderr> Caused by: java.io.FileNotFoundException: /tmp/test-spark/spark-2.3.2/jars/xercesImpl-2.9.1.jar (No such file or directory)
2018-10-10 14:32:47.114 - stderr> 	at java.util.zip.ZipFile.open(Native Method)
2018-10-10 14:32:47.114 - stderr> 	at java.util.zip.ZipFile.<init>(ZipFile.java:219)
2018-10-10 14:32:47.114 - stderr> 	at java.util.zip.ZipFile.<init>(ZipFile.java:149)
2018-10-10 14:32:47.114 - stderr> 	at java.util.jar.JarFile.<init>(JarFile.java:166)
2018-10-10 14:32:47.114 - stderr> 	at java.util.jar.JarFile.<init>(JarFile.java:103)
2018-10-10 14:32:47.114 - stderr> 	at sun.net.www.protocol.jar.URLJarFile.<init>(URLJarFile.java:93)
2018-10-10 14:32:47.114 - stderr> 	at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69)
2018-10-10 14:32:47.114 - stderr> 	at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:84)
2018-10-10 14:32:47.114 - stderr> 	at sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122)
2018-10-10 14:32:47.114 - stderr> 	at sun.net.www.protocol.jar.JarURLConnection.getInputStream(JarURLConnection.java:150)
2018-10-10 14:32:47.114 - stderr> 	at java.net.URL.openStream(URL.java:1038)
2018-10-10 14:32:47.114 - stderr> 	at java.util.ServiceLoader.parse(ServiceLoader.java:304)
2018-10-10 14:32:47.114 - stderr> 	... 24 more
           
	at org.scalatest.Assertions$class.newAssertionFailedException(Assertions.scala:528)
	at org.scalatest.FunSuite.newAssertionFailedException(FunSuite.scala:1560)
	at org.scalatest.Assertions$class.fail(Assertions.scala:1089)
	at org.scalatest.FunSuite.fail(FunSuite.scala:1560)
	at org.apache.spark.sql.hive.SparkSubmitTestUtils$class.runSparkSubmit(SparkSubmitTestUtils.scala:94)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.runSparkSubmit(HiveExternalCatalogVersionsSuite.scala:43)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite$$anonfun$beforeAll$1.apply(HiveExternalCatalogVersionsSuite.scala:187)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite$$anonfun$beforeAll$1.apply(HiveExternalCatalogVersionsSuite.scala:172)
	at scala.collection.immutable.List.foreach(List.scala:392)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.beforeAll(HiveExternalCatalogVersionsSuite.scala:172)
	at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:212)
	at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:210)
	at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:52)
	at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:314)
	at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:480)
	at sbt.ForkMain$Run$2.call(ForkMain.java:296)
	at sbt.ForkMain$Run$2.call(ForkMain.java:286)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)