org.scalatest.exceptions.TestFailedException: spark-submit returned with exit code 1. Command line: './bin/spark-submit' '--name' 'prepare testing tables' '--master' 'local[2]' '--conf' 'spark.ui.enabled=false' '--conf' 'spark.master.rest.enabled=false' '--conf' 'spark.sql.hive.metastore.version=1.2.1' '--conf' 'spark.sql.hive.metastore.jars=maven' '--conf' 'spark.sql.warehouse.dir=/home/jenkins/workspace/spark-master-test-sbt-hadoop-3.2/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-5ab906ca-f1e4-44d8-a45f-12c83e23e08b' '--conf' 'spark.sql.test.version.index=1' '--driver-java-options' '-Dderby.system.home=/home/jenkins/workspace/spark-master-test-sbt-hadoop-3.2/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-5ab906ca-f1e4-44d8-a45f-12c83e23e08b' '/home/jenkins/workspace/spark-master-test-sbt-hadoop-3.2/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test6000080996274345592.py' 2019-07-30 09:34:41.957 - stderr> 19/07/30 09:34:41 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2019-07-30 09:34:42.748 - stderr> Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties 2019-07-30 09:34:42.751 - stderr> 19/07/30 09:34:42 INFO SparkContext: Running Spark version 2.4.3 2019-07-30 09:34:42.781 - stderr> 19/07/30 09:34:42 INFO SparkContext: Submitted application: prepare testing tables 2019-07-30 09:34:42.84 - stderr> 19/07/30 09:34:42 INFO SecurityManager: Changing view acls to: jenkins 2019-07-30 09:34:42.84 - stderr> 19/07/30 09:34:42 INFO SecurityManager: Changing modify acls to: jenkins 2019-07-30 09:34:42.84 - stderr> 19/07/30 09:34:42 INFO SecurityManager: Changing view acls groups to: 2019-07-30 09:34:42.841 - stderr> 19/07/30 09:34:42 INFO SecurityManager: Changing modify acls groups to: 2019-07-30 09:34:42.841 - stderr> 19/07/30 09:34:42 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(jenkins); groups with view permissions: Set(); users with modify permissions: Set(jenkins); groups with modify permissions: Set() 2019-07-30 09:34:43.15 - stderr> 19/07/30 09:34:43 INFO Utils: Successfully started service 'sparkDriver' on port 34599. 2019-07-30 09:34:43.183 - stderr> 19/07/30 09:34:43 INFO SparkEnv: Registering MapOutputTracker 2019-07-30 09:34:43.204 - stderr> 19/07/30 09:34:43 INFO SparkEnv: Registering BlockManagerMaster 2019-07-30 09:34:43.208 - stderr> 19/07/30 09:34:43 INFO BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information 2019-07-30 09:34:43.208 - stderr> 19/07/30 09:34:43 INFO BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up 2019-07-30 09:34:43.219 - stderr> 19/07/30 09:34:43 INFO DiskBlockManager: Created local directory at /tmp/blockmgr-bd6af0e6-04fc-4dfc-bb61-9f5bd92fdb3a 2019-07-30 09:34:43.241 - stderr> 19/07/30 09:34:43 INFO MemoryStore: MemoryStore started with capacity 366.3 MB 2019-07-30 09:34:43.257 - stderr> 19/07/30 09:34:43 INFO SparkEnv: Registering OutputCommitCoordinator 2019-07-30 09:34:43.353 - stderr> 19/07/30 09:34:43 INFO Executor: Starting executor ID driver on host localhost 2019-07-30 09:34:43.425 - stderr> 19/07/30 09:34:43 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 35997. 2019-07-30 09:34:43.426 - stderr> 19/07/30 09:34:43 INFO NettyBlockTransferService: Server created on amp-jenkins-worker-03.amp:35997 2019-07-30 09:34:43.427 - stderr> 19/07/30 09:34:43 INFO BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy 2019-07-30 09:34:43.453 - stderr> 19/07/30 09:34:43 INFO BlockManagerMaster: Registering BlockManager BlockManagerId(driver, amp-jenkins-worker-03.amp, 35997, None) 2019-07-30 09:34:43.458 - stderr> 19/07/30 09:34:43 INFO BlockManagerMasterEndpoint: Registering block manager amp-jenkins-worker-03.amp:35997 with 366.3 MB RAM, BlockManagerId(driver, amp-jenkins-worker-03.amp, 35997, None) 2019-07-30 09:34:43.461 - stderr> 19/07/30 09:34:43 INFO BlockManagerMaster: Registered BlockManager BlockManagerId(driver, amp-jenkins-worker-03.amp, 35997, None) 2019-07-30 09:34:43.461 - stderr> 19/07/30 09:34:43 INFO BlockManager: Initialized BlockManager: BlockManagerId(driver, amp-jenkins-worker-03.amp, 35997, None) 2019-07-30 09:34:43.819 - stderr> 19/07/30 09:34:43 INFO SharedState: Setting hive.metastore.warehouse.dir ('null') to the value of spark.sql.warehouse.dir ('/home/jenkins/workspace/spark-master-test-sbt-hadoop-3.2/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-5ab906ca-f1e4-44d8-a45f-12c83e23e08b'). 2019-07-30 09:34:43.82 - stderr> 19/07/30 09:34:43 INFO SharedState: Warehouse path is '/home/jenkins/workspace/spark-master-test-sbt-hadoop-3.2/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-5ab906ca-f1e4-44d8-a45f-12c83e23e08b'. 2019-07-30 09:34:44.389 - stderr> 19/07/30 09:34:44 INFO StateStoreCoordinatorRef: Registered StateStoreCoordinator endpoint 2019-07-30 09:34:46.933 - stderr> 19/07/30 09:34:46 INFO HiveUtils: Initializing HiveMetastoreConnection version 1.2.1 using maven. 2019-07-30 09:34:46.939 - stderr> http://www.datanucleus.org/downloads/maven2 added as a remote repository with the name: repo-1 2019-07-30 09:34:46.942 - stderr> Ivy Default Cache set to: /home/jenkins/.ivy2/cache 2019-07-30 09:34:46.942 - stderr> The jars for the packages stored in: /home/jenkins/.ivy2/jars 2019-07-30 09:34:46.975 - stderr> :: loading settings :: url = jar:file:/tmp/test-spark/spark-2.4.3/jars/ivy-2.4.0.jar!/org/apache/ivy/core/settings/ivysettings.xml 2019-07-30 09:34:47.028 - stderr> org.apache.hive#hive-metastore added as a dependency 2019-07-30 09:34:47.028 - stderr> org.apache.hive#hive-exec added as a dependency 2019-07-30 09:34:47.028 - stderr> org.apache.hive#hive-common added as a dependency 2019-07-30 09:34:47.029 - stderr> org.apache.hive#hive-serde added as a dependency 2019-07-30 09:34:47.029 - stderr> com.google.guava#guava added as a dependency 2019-07-30 09:34:47.029 - stderr> org.apache.hadoop#hadoop-client added as a dependency 2019-07-30 09:34:47.031 - stderr> :: resolving dependencies :: org.apache.spark#spark-submit-parent-57f852f1-25a6-4c8a-a84a-c38d787a00b7;1.0 2019-07-30 09:34:47.031 - stderr> confs: [default] 2019-07-30 09:34:47.43 - stderr> found org.apache.hive#hive-metastore;1.2.2 in central 2019-07-30 09:34:47.492 - stderr> found org.apache.hive#hive-serde;1.2.2 in central 2019-07-30 09:34:47.529 - stderr> found org.apache.hive#hive-common;1.2.2 in central 2019-07-30 09:34:47.564 - stderr> found org.apache.hive#hive-shims;1.2.2 in central 2019-07-30 09:34:47.605 - stderr> found org.apache.hive.shims#hive-shims-common;1.2.2 in central 2019-07-30 09:34:47.636 - stderr> found commons-logging#commons-logging;1.1.3 in user-list 2019-07-30 09:34:47.667 - stderr> found log4j#log4j;1.2.16 in user-list 2019-07-30 09:34:47.688 - stderr> found log4j#apache-log4j-extras;1.2.17 in user-list 2019-07-30 09:34:47.718 - stderr> found com.google.guava#guava;14.0.1 in user-list 2019-07-30 09:34:47.738 - stderr> found commons-lang#commons-lang;2.6 in user-list 2019-07-30 09:34:47.757 - stderr> found org.apache.thrift#libthrift;0.9.2 in user-list 2019-07-30 09:34:47.781 - stderr> found org.slf4j#slf4j-api;1.7.5 in user-list 2019-07-30 09:34:47.803 - stderr> found org.apache.httpcomponents#httpclient;4.4 in user-list 2019-07-30 09:34:47.826 - stderr> found org.apache.httpcomponents#httpcore;4.4 in user-list 2019-07-30 09:34:47.849 - stderr> found commons-codec#commons-codec;1.4 in user-list 2019-07-30 09:34:47.877 - stderr> found org.apache.zookeeper#zookeeper;3.4.6 in user-list 2019-07-30 09:34:47.904 - stderr> found org.slf4j#slf4j-log4j12;1.7.5 in user-list 2019-07-30 09:34:47.932 - stderr> found jline#jline;2.12 in user-list 2019-07-30 09:34:47.955 - stderr> found io.netty#netty;3.7.0.Final in central 2019-07-30 09:34:47.985 - stderr> found org.apache.hive.shims#hive-shims-0.20S;1.2.2 in central 2019-07-30 09:34:48.015 - stderr> found org.apache.hive.shims#hive-shims-0.23;1.2.2 in central 2019-07-30 09:34:48.048 - stderr> found org.apache.hadoop#hadoop-yarn-server-resourcemanager;2.6.0 in user-list 2019-07-30 09:34:48.077 - stderr> found org.apache.hadoop#hadoop-annotations;2.6.0 in user-list 2019-07-30 09:34:48.631 - stderr> found com.google.inject.extensions#guice-servlet;3.0 in central 2019-07-30 09:34:48.662 - stderr> found com.google.inject#guice;3.0 in user-list 2019-07-30 09:34:48.687 - stderr> found javax.inject#javax.inject;1 in user-list 2019-07-30 09:34:48.707 - stderr> found aopalliance#aopalliance;1.0 in user-list 2019-07-30 09:34:48.727 - stderr> found org.sonatype.sisu.inject#cglib;2.2.1-v20090111 in user-list 2019-07-30 09:34:48.755 - stderr> found asm#asm;3.1 in user-list 2019-07-30 09:34:48.78 - stderr> found com.google.protobuf#protobuf-java;2.5.0 in user-list 2019-07-30 09:34:48.819 - stderr> found commons-io#commons-io;2.4 in user-list 2019-07-30 09:34:48.846 - stderr> found com.sun.jersey#jersey-json;1.14 in user-list 2019-07-30 09:34:48.871 - stderr> found org.codehaus.jettison#jettison;1.1 in user-list 2019-07-30 09:34:48.894 - stderr> found com.sun.xml.bind#jaxb-impl;2.2.3-1 in user-list 2019-07-30 09:34:48.912 - stderr> found javax.xml.bind#jaxb-api;2.2.2 in user-list 2019-07-30 09:34:48.928 - stderr> found javax.xml.stream#stax-api;1.0-2 in user-list 2019-07-30 09:34:48.95 - stderr> found javax.activation#activation;1.1 in user-list 2019-07-30 09:34:48.966 - stderr> found org.codehaus.jackson#jackson-core-asl;1.9.2 in central 2019-07-30 09:34:48.981 - stderr> found org.codehaus.jackson#jackson-mapper-asl;1.9.2 in central 2019-07-30 09:34:48.997 - stderr> found org.codehaus.jackson#jackson-jaxrs;1.9.2 in user-list 2019-07-30 09:34:49.013 - stderr> found org.codehaus.jackson#jackson-xc;1.9.2 in user-list 2019-07-30 09:34:49.03 - stderr> found com.sun.jersey.contribs#jersey-guice;1.9 in user-list 2019-07-30 09:34:49.058 - stderr> found org.apache.hadoop#hadoop-yarn-common;2.6.0 in user-list 2019-07-30 09:34:49.088 - stderr> found org.apache.hadoop#hadoop-yarn-api;2.6.0 in user-list 2019-07-30 09:34:49.138 - stderr> found org.apache.commons#commons-compress;1.4.1 in user-list 2019-07-30 09:34:49.156 - stderr> found org.tukaani#xz;1.0 in user-list 2019-07-30 09:34:49.182 - stderr> found org.mortbay.jetty#jetty-util;6.1.26 in user-list 2019-07-30 09:34:49.201 - stderr> found com.sun.jersey#jersey-core;1.14 in user-list 2019-07-30 09:34:49.227 - stderr> found com.sun.jersey#jersey-client;1.9 in user-list 2019-07-30 09:34:49.26 - stderr> found commons-cli#commons-cli;1.2 in user-list 2019-07-30 09:34:49.297 - stderr> found com.sun.jersey#jersey-server;1.14 in user-list 2019-07-30 09:34:49.333 - stderr> found org.apache.hadoop#hadoop-yarn-server-common;2.6.0 in user-list 2019-07-30 09:34:49.368 - stderr> found org.fusesource.leveldbjni#leveldbjni-all;1.8 in user-list 2019-07-30 09:34:49.383 - stderr> found org.apache.hadoop#hadoop-yarn-server-applicationhistoryservice;2.6.0 in user-list 2019-07-30 09:34:49.421 - stderr> found commons-collections#commons-collections;3.2.2 in central 2019-07-30 09:34:49.437 - stderr> found org.apache.hadoop#hadoop-yarn-server-web-proxy;2.6.0 in user-list 2019-07-30 09:34:49.453 - stderr> found commons-httpclient#commons-httpclient;3.0.1 in central 2019-07-30 09:34:49.466 - stderr> found junit#junit;4.11 in user-list 2019-07-30 09:34:49.481 - stderr> found org.hamcrest#hamcrest-core;1.3 in user-list 2019-07-30 09:34:49.496 - stderr> found org.mortbay.jetty#jetty;6.1.26 in user-list 2019-07-30 09:34:49.53 - stderr> found org.apache.hive.shims#hive-shims-scheduler;1.2.2 in central 2019-07-30 09:34:49.548 - stderr> found joda-time#joda-time;2.5 in user-list 2019-07-30 09:34:49.561 - stderr> found org.apache.ant#ant;1.9.1 in user-list 2019-07-30 09:34:49.57 - stderr> found org.apache.ant#ant-launcher;1.9.1 in user-list 2019-07-30 09:34:49.579 - stderr> found org.json#json;20090211 in user-list 2019-07-30 09:34:49.589 - stderr> found com.google.code.findbugs#jsr305;3.0.0 in user-list 2019-07-30 09:34:49.599 - stderr> found org.apache.avro#avro;1.7.5 in user-list 2019-07-30 09:34:49.614 - stderr> found com.thoughtworks.paranamer#paranamer;2.3 in user-list 2019-07-30 09:34:49.623 - stderr> found org.xerial.snappy#snappy-java;1.0.5 in user-list 2019-07-30 09:34:49.632 - stderr> found net.sf.opencsv#opencsv;2.3 in user-list 2019-07-30 09:34:49.642 - stderr> found com.twitter#parquet-hadoop-bundle;1.6.0 in user-list 2019-07-30 09:34:49.652 - stderr> found com.jolbox#bonecp;0.8.0.RELEASE in user-list 2019-07-30 09:34:49.66 - stderr> found org.apache.derby#derby;10.10.2.0 in user-list 2019-07-30 09:34:49.669 - stderr> found org.datanucleus#datanucleus-api-jdo;3.2.6 in user-list 2019-07-30 09:34:49.678 - stderr> found org.datanucleus#datanucleus-core;3.2.10 in user-list 2019-07-30 09:34:49.689 - stderr> found org.datanucleus#datanucleus-rdbms;3.2.9 in user-list 2019-07-30 09:34:49.699 - stderr> found commons-pool#commons-pool;1.5.4 in user-list 2019-07-30 09:34:49.707 - stderr> found commons-dbcp#commons-dbcp;1.4 in central 2019-07-30 09:34:49.716 - stderr> found javax.jdo#jdo-api;3.0.1 in user-list 2019-07-30 09:34:49.725 - stderr> found javax.transaction#jta;1.1 in user-list 2019-07-30 09:34:49.733 - stderr> found org.antlr#antlr-runtime;3.4 in user-list 2019-07-30 09:34:49.74 - stderr> found org.antlr#stringtemplate;3.2.1 in user-list 2019-07-30 09:34:49.748 - stderr> found antlr#antlr;2.7.7 in user-list 2019-07-30 09:34:49.758 - stderr> found org.apache.thrift#libfb303;0.9.2 in user-list 2019-07-30 09:34:49.771 - stderr> found org.apache.hive#hive-exec;1.2.2 in central 2019-07-30 09:34:49.784 - stderr> found org.apache.hive#hive-ant;1.2.2 in central 2019-07-30 09:34:49.804 - stderr> found org.apache.velocity#velocity;1.5 in user-list 2019-07-30 09:34:49.823 - stderr> found oro#oro;2.0.8 in user-list 2019-07-30 09:34:49.876 - stderr> found org.antlr#ST4;4.0.4 in user-list 2019-07-30 09:34:49.89 - stderr> found org.apache.ivy#ivy;2.4.0 in user-list 2019-07-30 09:34:49.901 - stderr> found org.codehaus.groovy#groovy-all;2.1.6 in user-list 2019-07-30 09:34:49.912 - stderr> found org.apache.calcite#calcite-core;1.2.0-incubating in user-list 2019-07-30 09:34:49.922 - stderr> found org.apache.calcite#calcite-avatica;1.2.0-incubating in user-list 2019-07-30 09:34:49.931 - stderr> found org.apache.calcite#calcite-linq4j;1.2.0-incubating in user-list 2019-07-30 09:34:49.95 - stderr> found net.hydromatic#eigenbase-properties;1.1.5 in central 2019-07-30 09:34:49.959 - stderr> found org.codehaus.janino#janino;2.7.6 in central 2019-07-30 09:34:49.969 - stderr> found org.codehaus.janino#commons-compiler;2.7.6 in central 2019-07-30 09:34:49.98 - stderr> found stax#stax-api;1.0.1 in user-list 2019-07-30 09:34:49.993 - stderr> found org.apache.hadoop#hadoop-client;2.7.3 in central 2019-07-30 09:34:50.004 - stderr> found org.apache.hadoop#hadoop-common;2.7.3 in central 2019-07-30 09:34:50.022 - stderr> found org.apache.hadoop#hadoop-annotations;2.7.3 in central 2019-07-30 09:34:50.042 - stderr> found org.apache.commons#commons-math3;3.1.1 in user-list 2019-07-30 09:34:50.051 - stderr> found xmlenc#xmlenc;0.52 in user-list 2019-07-30 09:34:50.059 - stderr> found commons-httpclient#commons-httpclient;3.1 in user-list 2019-07-30 09:34:50.078 - stderr> found commons-net#commons-net;3.1 in user-list 2019-07-30 09:34:50.089 - stderr> found log4j#log4j;1.2.17 in user-list 2019-07-30 09:34:50.101 - stderr> found commons-configuration#commons-configuration;1.6 in user-list 2019-07-30 09:34:50.113 - stderr> found commons-digester#commons-digester;1.8 in user-list 2019-07-30 09:34:50.122 - stderr> found commons-beanutils#commons-beanutils;1.7.0 in user-list 2019-07-30 09:34:50.131 - stderr> found commons-beanutils#commons-beanutils-core;1.8.0 in user-list 2019-07-30 09:34:50.14 - stderr> found org.slf4j#slf4j-api;1.7.10 in user-list 2019-07-30 09:34:50.149 - stderr> found org.codehaus.jackson#jackson-core-asl;1.9.13 in user-list 2019-07-30 09:34:50.157 - stderr> found org.codehaus.jackson#jackson-mapper-asl;1.9.13 in user-list 2019-07-30 09:34:50.173 - stderr> found com.google.code.gson#gson;2.2.4 in user-list 2019-07-30 09:34:50.186 - stderr> found org.apache.hadoop#hadoop-auth;2.7.3 in central 2019-07-30 09:34:50.203 - stderr> found org.apache.directory.server#apacheds-kerberos-codec;2.0.0-M15 in user-list 2019-07-30 09:34:50.214 - stderr> found org.apache.directory.server#apacheds-i18n;2.0.0-M15 in user-list 2019-07-30 09:34:50.224 - stderr> found org.apache.directory.api#api-asn1-api;1.0.0-M20 in user-list 2019-07-30 09:34:50.237 - stderr> found org.apache.directory.api#api-util;1.0.0-M20 in user-list 2019-07-30 09:34:50.253 - stderr> found org.apache.htrace#htrace-core;3.1.0-incubating in central 2019-07-30 09:34:50.268 - stderr> found javax.servlet.jsp#jsp-api;2.1 in user-list 2019-07-30 09:34:50.282 - stderr> found org.slf4j#slf4j-log4j12;1.7.10 in user-list 2019-07-30 09:34:50.297 - stderr> found org.apache.hadoop#hadoop-hdfs;2.7.3 in central 2019-07-30 09:34:50.324 - stderr> found io.netty#netty-all;4.0.23.Final in user-list 2019-07-30 09:34:50.349 - stderr> found xerces#xercesImpl;2.9.1 in central 2019-07-30 09:34:50.357 - stderr> found xml-apis#xml-apis;1.3.04 in user-list 2019-07-30 09:34:50.371 - stderr> found org.apache.hadoop#hadoop-mapreduce-client-app;2.7.3 in central 2019-07-30 09:34:50.38 - stderr> found org.apache.hadoop#hadoop-mapreduce-client-common;2.7.3 in central 2019-07-30 09:34:50.391 - stderr> found org.apache.hadoop#hadoop-yarn-common;2.7.3 in central 2019-07-30 09:34:50.404 - stderr> found org.apache.hadoop#hadoop-yarn-api;2.7.3 in central 2019-07-30 09:34:50.476 - stderr> found org.codehaus.jackson#jackson-jaxrs;1.9.13 in user-list 2019-07-30 09:34:50.487 - stderr> found org.codehaus.jackson#jackson-xc;1.9.13 in user-list 2019-07-30 09:34:50.545 - stderr> found org.apache.hadoop#hadoop-yarn-client;2.7.3 in central 2019-07-30 09:34:50.563 - stderr> found org.apache.hadoop#hadoop-mapreduce-client-core;2.7.3 in central 2019-07-30 09:34:50.581 - stderr> found org.apache.hadoop#hadoop-yarn-server-common;2.7.3 in central 2019-07-30 09:34:50.603 - stderr> found org.apache.hadoop#hadoop-mapreduce-client-shuffle;2.7.3 in central 2019-07-30 09:34:50.619 - stderr> found org.apache.hadoop#hadoop-mapreduce-client-jobclient;2.7.3 in central 2019-07-30 09:34:50.717 - stderr> :: resolution report :: resolve 3630ms :: artifacts dl 55ms 2019-07-30 09:34:50.718 - stderr> :: modules in use: 2019-07-30 09:34:50.719 - stderr> antlr#antlr;2.7.7 from user-list in [default] 2019-07-30 09:34:50.72 - stderr> aopalliance#aopalliance;1.0 from user-list in [default] 2019-07-30 09:34:50.72 - stderr> asm#asm;3.1 from user-list in [default] 2019-07-30 09:34:50.72 - stderr> com.google.code.findbugs#jsr305;3.0.0 from user-list in [default] 2019-07-30 09:34:50.721 - stderr> com.google.code.gson#gson;2.2.4 from user-list in [default] 2019-07-30 09:34:50.721 - stderr> com.google.guava#guava;14.0.1 from user-list in [default] 2019-07-30 09:34:50.721 - stderr> com.google.inject#guice;3.0 from user-list in [default] 2019-07-30 09:34:50.721 - stderr> com.google.inject.extensions#guice-servlet;3.0 from central in [default] 2019-07-30 09:34:50.722 - stderr> com.google.protobuf#protobuf-java;2.5.0 from user-list in [default] 2019-07-30 09:34:50.722 - stderr> com.jolbox#bonecp;0.8.0.RELEASE from user-list in [default] 2019-07-30 09:34:50.722 - stderr> com.sun.jersey#jersey-client;1.9 from user-list in [default] 2019-07-30 09:34:50.722 - stderr> com.sun.jersey#jersey-core;1.14 from user-list in [default] 2019-07-30 09:34:50.722 - stderr> com.sun.jersey#jersey-json;1.14 from user-list in [default] 2019-07-30 09:34:50.722 - stderr> com.sun.jersey#jersey-server;1.14 from user-list in [default] 2019-07-30 09:34:50.723 - stderr> com.sun.jersey.contribs#jersey-guice;1.9 from user-list in [default] 2019-07-30 09:34:50.723 - stderr> com.sun.xml.bind#jaxb-impl;2.2.3-1 from user-list in [default] 2019-07-30 09:34:50.723 - stderr> com.thoughtworks.paranamer#paranamer;2.3 from user-list in [default] 2019-07-30 09:34:50.723 - stderr> com.twitter#parquet-hadoop-bundle;1.6.0 from user-list in [default] 2019-07-30 09:34:50.723 - stderr> commons-beanutils#commons-beanutils;1.7.0 from user-list in [default] 2019-07-30 09:34:50.723 - stderr> commons-beanutils#commons-beanutils-core;1.8.0 from user-list in [default] 2019-07-30 09:34:50.724 - stderr> commons-cli#commons-cli;1.2 from user-list in [default] 2019-07-30 09:34:50.724 - stderr> commons-codec#commons-codec;1.4 from user-list in [default] 2019-07-30 09:34:50.724 - stderr> commons-collections#commons-collections;3.2.2 from central in [default] 2019-07-30 09:34:50.724 - stderr> commons-configuration#commons-configuration;1.6 from user-list in [default] 2019-07-30 09:34:50.724 - stderr> commons-dbcp#commons-dbcp;1.4 from central in [default] 2019-07-30 09:34:50.724 - stderr> commons-digester#commons-digester;1.8 from user-list in [default] 2019-07-30 09:34:50.725 - stderr> commons-httpclient#commons-httpclient;3.1 from user-list in [default] 2019-07-30 09:34:50.725 - stderr> commons-io#commons-io;2.4 from user-list in [default] 2019-07-30 09:34:50.725 - stderr> commons-lang#commons-lang;2.6 from user-list in [default] 2019-07-30 09:34:50.725 - stderr> commons-logging#commons-logging;1.1.3 from user-list in [default] 2019-07-30 09:34:50.725 - stderr> commons-net#commons-net;3.1 from user-list in [default] 2019-07-30 09:34:50.726 - stderr> commons-pool#commons-pool;1.5.4 from user-list in [default] 2019-07-30 09:34:50.726 - stderr> io.netty#netty;3.7.0.Final from central in [default] 2019-07-30 09:34:50.726 - stderr> io.netty#netty-all;4.0.23.Final from user-list in [default] 2019-07-30 09:34:50.726 - stderr> javax.activation#activation;1.1 from user-list in [default] 2019-07-30 09:34:50.727 - stderr> javax.inject#javax.inject;1 from user-list in [default] 2019-07-30 09:34:50.727 - stderr> javax.jdo#jdo-api;3.0.1 from user-list in [default] 2019-07-30 09:34:50.727 - stderr> javax.servlet.jsp#jsp-api;2.1 from user-list in [default] 2019-07-30 09:34:50.727 - stderr> javax.transaction#jta;1.1 from user-list in [default] 2019-07-30 09:34:50.727 - stderr> javax.xml.bind#jaxb-api;2.2.2 from user-list in [default] 2019-07-30 09:34:50.728 - stderr> javax.xml.stream#stax-api;1.0-2 from user-list in [default] 2019-07-30 09:34:50.728 - stderr> jline#jline;2.12 from user-list in [default] 2019-07-30 09:34:50.728 - stderr> joda-time#joda-time;2.5 from user-list in [default] 2019-07-30 09:34:50.728 - stderr> log4j#apache-log4j-extras;1.2.17 from user-list in [default] 2019-07-30 09:34:50.728 - stderr> log4j#log4j;1.2.17 from user-list in [default] 2019-07-30 09:34:50.729 - stderr> net.hydromatic#eigenbase-properties;1.1.5 from central in [default] 2019-07-30 09:34:50.729 - stderr> net.sf.opencsv#opencsv;2.3 from user-list in [default] 2019-07-30 09:34:50.729 - stderr> org.antlr#ST4;4.0.4 from user-list in [default] 2019-07-30 09:34:50.729 - stderr> org.antlr#antlr-runtime;3.4 from user-list in [default] 2019-07-30 09:34:50.729 - stderr> org.antlr#stringtemplate;3.2.1 from user-list in [default] 2019-07-30 09:34:50.73 - stderr> org.apache.ant#ant;1.9.1 from user-list in [default] 2019-07-30 09:34:50.73 - stderr> org.apache.ant#ant-launcher;1.9.1 from user-list in [default] 2019-07-30 09:34:50.73 - stderr> org.apache.avro#avro;1.7.5 from user-list in [default] 2019-07-30 09:34:50.73 - stderr> org.apache.calcite#calcite-avatica;1.2.0-incubating from user-list in [default] 2019-07-30 09:34:50.73 - stderr> org.apache.calcite#calcite-core;1.2.0-incubating from user-list in [default] 2019-07-30 09:34:50.731 - stderr> org.apache.calcite#calcite-linq4j;1.2.0-incubating from user-list in [default] 2019-07-30 09:34:50.731 - stderr> org.apache.commons#commons-compress;1.4.1 from user-list in [default] 2019-07-30 09:34:50.731 - stderr> org.apache.commons#commons-math3;3.1.1 from user-list in [default] 2019-07-30 09:34:50.731 - stderr> org.apache.derby#derby;10.10.2.0 from user-list in [default] 2019-07-30 09:34:50.731 - stderr> org.apache.directory.api#api-asn1-api;1.0.0-M20 from user-list in [default] 2019-07-30 09:34:50.732 - stderr> org.apache.directory.api#api-util;1.0.0-M20 from user-list in [default] 2019-07-30 09:34:50.732 - stderr> org.apache.directory.server#apacheds-i18n;2.0.0-M15 from user-list in [default] 2019-07-30 09:34:50.732 - stderr> org.apache.directory.server#apacheds-kerberos-codec;2.0.0-M15 from user-list in [default] 2019-07-30 09:34:50.732 - stderr> org.apache.hadoop#hadoop-annotations;2.7.3 from central in [default] 2019-07-30 09:34:50.733 - stderr> org.apache.hadoop#hadoop-auth;2.7.3 from central in [default] 2019-07-30 09:34:50.733 - stderr> org.apache.hadoop#hadoop-client;2.7.3 from central in [default] 2019-07-30 09:34:50.733 - stderr> org.apache.hadoop#hadoop-common;2.7.3 from central in [default] 2019-07-30 09:34:50.733 - stderr> org.apache.hadoop#hadoop-hdfs;2.7.3 from central in [default] 2019-07-30 09:34:50.733 - stderr> org.apache.hadoop#hadoop-mapreduce-client-app;2.7.3 from central in [default] 2019-07-30 09:34:50.734 - stderr> org.apache.hadoop#hadoop-mapreduce-client-common;2.7.3 from central in [default] 2019-07-30 09:34:50.734 - stderr> org.apache.hadoop#hadoop-mapreduce-client-core;2.7.3 from central in [default] 2019-07-30 09:34:50.734 - stderr> org.apache.hadoop#hadoop-mapreduce-client-jobclient;2.7.3 from central in [default] 2019-07-30 09:34:50.734 - stderr> org.apache.hadoop#hadoop-mapreduce-client-shuffle;2.7.3 from central in [default] 2019-07-30 09:34:50.735 - stderr> org.apache.hadoop#hadoop-yarn-api;2.7.3 from central in [default] 2019-07-30 09:34:50.735 - stderr> org.apache.hadoop#hadoop-yarn-client;2.7.3 from central in [default] 2019-07-30 09:34:50.735 - stderr> org.apache.hadoop#hadoop-yarn-common;2.7.3 from central in [default] 2019-07-30 09:34:50.735 - stderr> org.apache.hadoop#hadoop-yarn-server-applicationhistoryservice;2.6.0 from user-list in [default] 2019-07-30 09:34:50.735 - stderr> org.apache.hadoop#hadoop-yarn-server-common;2.7.3 from central in [default] 2019-07-30 09:34:50.735 - stderr> org.apache.hadoop#hadoop-yarn-server-resourcemanager;2.6.0 from user-list in [default] 2019-07-30 09:34:50.736 - stderr> org.apache.hadoop#hadoop-yarn-server-web-proxy;2.6.0 from user-list in [default] 2019-07-30 09:34:50.736 - stderr> org.apache.hive#hive-ant;1.2.2 from central in [default] 2019-07-30 09:34:50.736 - stderr> org.apache.hive#hive-common;1.2.2 from central in [default] 2019-07-30 09:34:50.736 - stderr> org.apache.hive#hive-exec;1.2.2 from central in [default] 2019-07-30 09:34:50.736 - stderr> org.apache.hive#hive-metastore;1.2.2 from central in [default] 2019-07-30 09:34:50.737 - stderr> org.apache.hive#hive-serde;1.2.2 from central in [default] 2019-07-30 09:34:50.737 - stderr> org.apache.hive#hive-shims;1.2.2 from central in [default] 2019-07-30 09:34:50.737 - stderr> org.apache.hive.shims#hive-shims-0.20S;1.2.2 from central in [default] 2019-07-30 09:34:50.737 - stderr> org.apache.hive.shims#hive-shims-0.23;1.2.2 from central in [default] 2019-07-30 09:34:50.738 - stderr> org.apache.hive.shims#hive-shims-common;1.2.2 from central in [default] 2019-07-30 09:34:50.738 - stderr> org.apache.hive.shims#hive-shims-scheduler;1.2.2 from central in [default] 2019-07-30 09:34:50.738 - stderr> org.apache.htrace#htrace-core;3.1.0-incubating from central in [default] 2019-07-30 09:34:50.738 - stderr> org.apache.httpcomponents#httpclient;4.4 from user-list in [default] 2019-07-30 09:34:50.738 - stderr> org.apache.httpcomponents#httpcore;4.4 from user-list in [default] 2019-07-30 09:34:50.739 - stderr> org.apache.ivy#ivy;2.4.0 from user-list in [default] 2019-07-30 09:34:50.739 - stderr> org.apache.thrift#libfb303;0.9.2 from user-list in [default] 2019-07-30 09:34:50.739 - stderr> org.apache.thrift#libthrift;0.9.2 from user-list in [default] 2019-07-30 09:34:50.739 - stderr> org.apache.velocity#velocity;1.5 from user-list in [default] 2019-07-30 09:34:50.739 - stderr> org.apache.zookeeper#zookeeper;3.4.6 from user-list in [default] 2019-07-30 09:34:50.739 - stderr> org.codehaus.groovy#groovy-all;2.1.6 from user-list in [default] 2019-07-30 09:34:50.739 - stderr> org.codehaus.jackson#jackson-core-asl;1.9.13 from user-list in [default] 2019-07-30 09:34:50.74 - stderr> org.codehaus.jackson#jackson-jaxrs;1.9.13 from user-list in [default] 2019-07-30 09:34:50.74 - stderr> org.codehaus.jackson#jackson-mapper-asl;1.9.13 from user-list in [default] 2019-07-30 09:34:50.74 - stderr> org.codehaus.jackson#jackson-xc;1.9.13 from user-list in [default] 2019-07-30 09:34:50.74 - stderr> org.codehaus.janino#commons-compiler;2.7.6 from central in [default] 2019-07-30 09:34:50.74 - stderr> org.codehaus.janino#janino;2.7.6 from central in [default] 2019-07-30 09:34:50.74 - stderr> org.codehaus.jettison#jettison;1.1 from user-list in [default] 2019-07-30 09:34:50.74 - stderr> org.datanucleus#datanucleus-api-jdo;3.2.6 from user-list in [default] 2019-07-30 09:34:50.74 - stderr> org.datanucleus#datanucleus-core;3.2.10 from user-list in [default] 2019-07-30 09:34:50.741 - stderr> org.datanucleus#datanucleus-rdbms;3.2.9 from user-list in [default] 2019-07-30 09:34:50.741 - stderr> org.fusesource.leveldbjni#leveldbjni-all;1.8 from user-list in [default] 2019-07-30 09:34:50.741 - stderr> org.json#json;20090211 from user-list in [default] 2019-07-30 09:34:50.741 - stderr> org.mortbay.jetty#jetty;6.1.26 from user-list in [default] 2019-07-30 09:34:50.741 - stderr> org.mortbay.jetty#jetty-util;6.1.26 from user-list in [default] 2019-07-30 09:34:50.741 - stderr> org.slf4j#slf4j-api;1.7.10 from user-list in [default] 2019-07-30 09:34:50.741 - stderr> org.slf4j#slf4j-log4j12;1.7.10 from user-list in [default] 2019-07-30 09:34:50.742 - stderr> org.sonatype.sisu.inject#cglib;2.2.1-v20090111 from user-list in [default] 2019-07-30 09:34:50.742 - stderr> org.tukaani#xz;1.0 from user-list in [default] 2019-07-30 09:34:50.742 - stderr> org.xerial.snappy#snappy-java;1.0.5 from user-list in [default] 2019-07-30 09:34:50.742 - stderr> oro#oro;2.0.8 from user-list in [default] 2019-07-30 09:34:50.742 - stderr> stax#stax-api;1.0.1 from user-list in [default] 2019-07-30 09:34:50.742 - stderr> xerces#xercesImpl;2.9.1 from central in [default] 2019-07-30 09:34:50.742 - stderr> xml-apis#xml-apis;1.3.04 from user-list in [default] 2019-07-30 09:34:50.742 - stderr> xmlenc#xmlenc;0.52 from user-list in [default] 2019-07-30 09:34:50.743 - stderr> :: evicted modules: 2019-07-30 09:34:50.743 - stderr> log4j#log4j;1.2.16 by [log4j#log4j;1.2.17] in [default] 2019-07-30 09:34:50.743 - stderr> org.slf4j#slf4j-api;1.7.5 by [org.slf4j#slf4j-api;1.7.10] in [default] 2019-07-30 09:34:50.743 - stderr> org.slf4j#slf4j-log4j12;1.7.5 by [org.slf4j#slf4j-log4j12;1.7.10] in [default] 2019-07-30 09:34:50.743 - stderr> org.apache.hadoop#hadoop-annotations;2.6.0 by [org.apache.hadoop#hadoop-annotations;2.7.3] in [default] 2019-07-30 09:34:50.743 - stderr> org.codehaus.jackson#jackson-core-asl;1.9.2 by [org.codehaus.jackson#jackson-core-asl;1.9.13] in [default] 2019-07-30 09:34:50.743 - stderr> org.codehaus.jackson#jackson-mapper-asl;1.9.2 by [org.codehaus.jackson#jackson-mapper-asl;1.9.13] in [default] 2019-07-30 09:34:50.743 - stderr> org.codehaus.jackson#jackson-jaxrs;1.9.2 by [org.codehaus.jackson#jackson-jaxrs;1.9.13] in [default] 2019-07-30 09:34:50.743 - stderr> org.codehaus.jackson#jackson-xc;1.9.2 by [org.codehaus.jackson#jackson-xc;1.9.13] in [default] 2019-07-30 09:34:50.743 - stderr> org.apache.hadoop#hadoop-yarn-common;2.6.0 by [org.apache.hadoop#hadoop-yarn-common;2.7.3] in [default] 2019-07-30 09:34:50.743 - stderr> org.apache.hadoop#hadoop-yarn-api;2.6.0 by [org.apache.hadoop#hadoop-yarn-api;2.7.3] in [default] 2019-07-30 09:34:50.743 - stderr> org.apache.hadoop#hadoop-yarn-server-common;2.6.0 by [org.apache.hadoop#hadoop-yarn-server-common;2.7.3] in [default] 2019-07-30 09:34:50.743 - stderr> commons-httpclient#commons-httpclient;3.0.1 by [commons-httpclient#commons-httpclient;3.1] in [default] 2019-07-30 09:34:50.747 - stderr> junit#junit;4.11 transitively in [default] 2019-07-30 09:34:50.747 - stderr> org.hamcrest#hamcrest-core;1.3 transitively in [default] 2019-07-30 09:34:50.747 - stderr> com.google.code.findbugs#jsr305;1.3.9 by [com.google.code.findbugs#jsr305;3.0.0] in [default] 2019-07-30 09:34:50.747 - stderr> com.google.guava#guava;11.0.2 by [com.google.guava#guava;14.0.1] in [default] 2019-07-30 09:34:50.748 - stderr> org.apache.avro#avro;1.7.4 by [org.apache.avro#avro;1.7.5] in [default] 2019-07-30 09:34:50.748 - stderr> org.apache.httpcomponents#httpclient;4.2.5 by [org.apache.httpcomponents#httpclient;4.4] in [default] 2019-07-30 09:34:50.748 - stderr> io.netty#netty;3.6.2.Final by [io.netty#netty;3.7.0.Final] in [default] 2019-07-30 09:34:50.748 - stderr> com.sun.jersey#jersey-core;1.9 by [com.sun.jersey#jersey-core;1.14] in [default] 2019-07-30 09:34:50.748 - stderr> com.sun.jersey#jersey-server;1.9 by [com.sun.jersey#jersey-server;1.14] in [default] 2019-07-30 09:34:50.748 - stderr> com.sun.jersey#jersey-json;1.9 by [com.sun.jersey#jersey-json;1.14] in [default] 2019-07-30 09:34:50.748 - stderr> --------------------------------------------------------------------- 2019-07-30 09:34:50.748 - stderr> | | modules || artifacts | 2019-07-30 09:34:50.748 - stderr> | conf | number| search|dwnlded|evicted|| number|dwnlded| 2019-07-30 09:34:50.748 - stderr> --------------------------------------------------------------------- 2019-07-30 09:34:50.749 - stderr> | default | 145 | 1 | 1 | 22 || 123 | 0 | 2019-07-30 09:34:50.749 - stderr> --------------------------------------------------------------------- 2019-07-30 09:34:50.789 - stderr> 2019-07-30 09:34:50.789 - stderr> :: problems summary :: 2019-07-30 09:34:50.789 - stderr> :::: ERRORS 2019-07-30 09:34:50.789 - stderr> unknown resolver null 2019-07-30 09:34:50.79 - stderr> 2019-07-30 09:34:50.79 - stderr> 2019-07-30 09:34:50.79 - stderr> :: USE VERBOSE OR DEBUG MESSAGE LEVEL FOR MORE DETAILS 2019-07-30 09:34:50.791 - stderr> :: retrieving :: org.apache.spark#spark-submit-parent-57f852f1-25a6-4c8a-a84a-c38d787a00b7 2019-07-30 09:34:50.791 - stderr> confs: [default] 2019-07-30 09:34:50.836 - stderr> 0 artifacts copied, 123 already retrieved (0kB/45ms) 2019-07-30 09:34:50.998 - stderr> 19/07/30 09:34:50 INFO IsolatedClientLoader: Downloaded metastore jars to /tmp/hive-v1_2-3db9434c-f1b4-4291-ab92-4b28a909c359 2019-07-30 09:34:51.579 - stderr> 19/07/30 09:34:51 INFO HiveMetaStore: 0: Opening raw store with implemenation class:org.apache.hadoop.hive.metastore.ObjectStore 2019-07-30 09:34:51.606 - stderr> 19/07/30 09:34:51 INFO ObjectStore: ObjectStore, initialize called 2019-07-30 09:34:51.719 - stderr> 19/07/30 09:34:51 INFO Persistence: Property hive.metastore.integral.jdo.pushdown unknown - will be ignored 2019-07-30 09:34:51.719 - stderr> 19/07/30 09:34:51 INFO Persistence: Property datanucleus.cache.level2 unknown - will be ignored 2019-07-30 09:35:01.404 - stderr> 19/07/30 09:35:01 INFO ObjectStore: Setting MetaStore object pin classes with hive.metastore.cache.pinobjtypes="Table,StorageDescriptor,SerDeInfo,Partition,Database,Type,FieldSchema,Order" 2019-07-30 09:35:02.862 - stderr> 19/07/30 09:35:02 INFO Datastore: The class "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as "embedded-only" so does not have its own datastore table. 2019-07-30 09:35:02.863 - stderr> 19/07/30 09:35:02 INFO Datastore: The class "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as "embedded-only" so does not have its own datastore table. 2019-07-30 09:35:03.074 - stderr> 19/07/30 09:35:03 INFO Datastore: The class "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as "embedded-only" so does not have its own datastore table. 2019-07-30 09:35:03.074 - stderr> 19/07/30 09:35:03 INFO Datastore: The class "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as "embedded-only" so does not have its own datastore table. 2019-07-30 09:35:03.154 - stderr> 19/07/30 09:35:03 INFO Query: Reading in results for query "org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is closing 2019-07-30 09:35:03.155 - stderr> 19/07/30 09:35:03 INFO MetaStoreDirectSql: Using direct SQL, underlying DB is DERBY 2019-07-30 09:35:03.158 - stderr> 19/07/30 09:35:03 INFO ObjectStore: Initialized ObjectStore 2019-07-30 09:35:03.331 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: Added admin role in metastore 2019-07-30 09:35:03.333 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: Added public role in metastore 2019-07-30 09:35:03.377 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: No user is added in admin role, since config is empty 2019-07-30 09:35:03.465 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_all_databases 2019-07-30 09:35:03.466 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins ip=unknown-ip-addr cmd=get_all_databases 2019-07-30 09:35:03.481 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_functions: db=default pat=* 2019-07-30 09:35:03.481 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins ip=unknown-ip-addr cmd=get_functions: db=default pat=* 2019-07-30 09:35:03.482 - stderr> 19/07/30 09:35:03 INFO Datastore: The class "org.apache.hadoop.hive.metastore.model.MResourceUri" is tagged as "embedded-only" so does not have its own datastore table. 2019-07-30 09:35:03.542 - stderr> 19/07/30 09:35:03 INFO SessionState: Created local directory: /tmp/1b2706a2-b4a8-44ee-a479-de703182600d_resources 2019-07-30 09:35:03.546 - stderr> 19/07/30 09:35:03 INFO SessionState: Created HDFS directory: /tmp/hive/jenkins/1b2706a2-b4a8-44ee-a479-de703182600d 2019-07-30 09:35:03.55 - stderr> 19/07/30 09:35:03 INFO SessionState: Created local directory: /tmp/jenkins/1b2706a2-b4a8-44ee-a479-de703182600d 2019-07-30 09:35:03.554 - stderr> 19/07/30 09:35:03 INFO SessionState: Created HDFS directory: /tmp/hive/jenkins/1b2706a2-b4a8-44ee-a479-de703182600d/_tmp_space.db 2019-07-30 09:35:03.557 - stderr> 19/07/30 09:35:03 INFO HiveClientImpl: Warehouse location for Hive client (version 1.2.2) is /home/jenkins/workspace/spark-master-test-sbt-hadoop-3.2/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-5ab906ca-f1e4-44d8-a45f-12c83e23e08b 2019-07-30 09:35:03.566 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_database: default 2019-07-30 09:35:03.566 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins ip=unknown-ip-addr cmd=get_database: default 2019-07-30 09:35:03.572 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_table : db=default tbl=data_source_tbl_1 2019-07-30 09:35:03.572 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins ip=unknown-ip-addr cmd=get_table : db=default tbl=data_source_tbl_1 2019-07-30 09:35:03.584 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_database: default 2019-07-30 09:35:03.584 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins ip=unknown-ip-addr cmd=get_database: default 2019-07-30 09:35:03.587 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_database: default 2019-07-30 09:35:03.587 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins ip=unknown-ip-addr cmd=get_database: default 2019-07-30 09:35:03.611 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_database: default 2019-07-30 09:35:03.612 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins ip=unknown-ip-addr cmd=get_database: default 2019-07-30 09:35:03.614 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_database: default 2019-07-30 09:35:03.614 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins ip=unknown-ip-addr cmd=get_database: default 2019-07-30 09:35:03.648 - stdout> Traceback (most recent call last): 2019-07-30 09:35:03.648 - stdout> File "/home/jenkins/workspace/spark-master-test-sbt-hadoop-3.2/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test6000080996274345592.py", line 8, in <module> 2019-07-30 09:35:03.648 - stdout> spark.sql("create table data_source_tbl_{} using json as select 1 i".format(version_index)) 2019-07-30 09:35:03.648 - stdout> File "/tmp/test-spark/spark-2.4.3/python/lib/pyspark.zip/pyspark/sql/session.py", line 767, in sql 2019-07-30 09:35:03.648 - stdout> File "/tmp/test-spark/spark-2.4.3/python/lib/py4j-0.10.7-src.zip/py4j/java_gateway.py", line 1257, in __call__ 2019-07-30 09:35:03.648 - stdout> File "/tmp/test-spark/spark-2.4.3/python/lib/pyspark.zip/pyspark/sql/utils.py", line 63, in deco 2019-07-30 09:35:03.648 - stdout> File "/tmp/test-spark/spark-2.4.3/python/lib/py4j-0.10.7-src.zip/py4j/protocol.py", line 328, in get_return_value 2019-07-30 09:35:03.65 - stdout> py4j.protocol.Py4JJavaError: An error occurred while calling o29.sql. 2019-07-30 09:35:03.65 - stdout> : java.util.ServiceConfigurationError: org.apache.spark.sql.sources.DataSourceRegister: Error reading configuration file 2019-07-30 09:35:03.65 - stdout> at java.util.ServiceLoader.fail(ServiceLoader.java:232) 2019-07-30 09:35:03.65 - stdout> at java.util.ServiceLoader.parse(ServiceLoader.java:309) 2019-07-30 09:35:03.65 - stdout> at java.util.ServiceLoader.access$200(ServiceLoader.java:185) 2019-07-30 09:35:03.65 - stdout> at java.util.ServiceLoader$LazyIterator.hasNextService(ServiceLoader.java:357) 2019-07-30 09:35:03.65 - stdout> at java.util.ServiceLoader$LazyIterator.hasNext(ServiceLoader.java:393) 2019-07-30 09:35:03.65 - stdout> at java.util.ServiceLoader$1.hasNext(ServiceLoader.java:474) 2019-07-30 09:35:03.65 - stdout> at scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:42) 2019-07-30 09:35:03.65 - stdout> at scala.collection.Iterator$class.foreach(Iterator.scala:891) 2019-07-30 09:35:03.65 - stdout> at scala.collection.AbstractIterator.foreach(Iterator.scala:1334) 2019-07-30 09:35:03.65 - stdout> at scala.collection.IterableLike$class.foreach(IterableLike.scala:72) 2019-07-30 09:35:03.65 - stdout> at scala.collection.AbstractIterable.foreach(Iterable.scala:54) 2019-07-30 09:35:03.65 - stdout> at scala.collection.TraversableLike$class.filterImpl(TraversableLike.scala:247) 2019-07-30 09:35:03.65 - stdout> at scala.collection.TraversableLike$class.filter(TraversableLike.scala:259) 2019-07-30 09:35:03.65 - stdout> at scala.collection.AbstractTraversable.filter(Traversable.scala:104) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.execution.datasources.DataSource$.lookupDataSource(DataSource.scala:630) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.execution.datasources.DataSource.providingClass$lzycompute(DataSource.scala:94) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.execution.datasources.DataSource.providingClass(DataSource.scala:93) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.execution.datasources.DataSource.writeAndRead(DataSource.scala:482) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.execution.command.CreateDataSourceTableAsSelectCommand.saveDataIntoTable(createDataSourceTables.scala:217) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.execution.command.CreateDataSourceTableAsSelectCommand.run(createDataSourceTables.scala:176) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult$lzycompute(commands.scala:104) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult(commands.scala:102) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.execution.command.DataWritingCommandExec.executeCollect(commands.scala:115) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:194) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:194) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.Dataset$$anonfun$53.apply(Dataset.scala:3364) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:78) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:125) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:73) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3363) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.Dataset.<init>(Dataset.scala:194) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:79) 2019-07-30 09:35:03.65 - stdout> at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642) 2019-07-30 09:35:03.65 - stdout> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 2019-07-30 09:35:03.65 - stdout> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 2019-07-30 09:35:03.65 - stdout> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) 2019-07-30 09:35:03.65 - stdout> at java.lang.reflect.Method.invoke(Method.java:498) 2019-07-30 09:35:03.65 - stdout> at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) 2019-07-30 09:35:03.65 - stdout> at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) 2019-07-30 09:35:03.65 - stdout> at py4j.Gateway.invoke(Gateway.java:282) 2019-07-30 09:35:03.65 - stdout> at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) 2019-07-30 09:35:03.65 - stdout> at py4j.commands.CallCommand.execute(CallCommand.java:79) 2019-07-30 09:35:03.65 - stdout> at py4j.GatewayConnection.run(GatewayConnection.java:238) 2019-07-30 09:35:03.65 - stdout> at java.lang.Thread.run(Thread.java:748) 2019-07-30 09:35:03.65 - stdout> Caused by: java.io.FileNotFoundException: /tmp/test-spark/spark-2.4.3/jars/spark-sql_2.11-2.4.3.jar (No such file or directory) 2019-07-30 09:35:03.65 - stdout> at java.util.zip.ZipFile.open(Native Method) 2019-07-30 09:35:03.65 - stdout> at java.util.zip.ZipFile.<init>(ZipFile.java:225) 2019-07-30 09:35:03.65 - stdout> at java.util.zip.ZipFile.<init>(ZipFile.java:155) 2019-07-30 09:35:03.65 - stdout> at java.util.jar.JarFile.<init>(JarFile.java:166) 2019-07-30 09:35:03.65 - stdout> at java.util.jar.JarFile.<init>(JarFile.java:103) 2019-07-30 09:35:03.65 - stdout> at sun.net.www.protocol.jar.URLJarFile.<init>(URLJarFile.java:93) 2019-07-30 09:35:03.651 - stdout> at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69) 2019-07-30 09:35:03.651 - stdout> at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:84) 2019-07-30 09:35:03.651 - stdout> at sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122) 2019-07-30 09:35:03.651 - stdout> at sun.net.www.protocol.jar.JarURLConnection.getInputStream(JarURLConnection.java:152) 2019-07-30 09:35:03.651 - stdout> at java.net.URL.openStream(URL.java:1045) 2019-07-30 09:35:03.651 - stdout> at java.util.ServiceLoader.parse(ServiceLoader.java:304) 2019-07-30 09:35:03.651 - stdout> ... 42 more 2019-07-30 09:35:03.651 - stdout> 2019-07-30 09:35:03.729 - stderr> 19/07/30 09:35:03 INFO SparkContext: Invoking stop() from shutdown hook 2019-07-30 09:35:03.743 - stderr> 19/07/30 09:35:03 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped! 2019-07-30 09:35:03.753 - stderr> 19/07/30 09:35:03 INFO MemoryStore: MemoryStore cleared 2019-07-30 09:35:03.753 - stderr> 19/07/30 09:35:03 INFO BlockManager: BlockManager stopped 2019-07-30 09:35:03.754 - stderr> 19/07/30 09:35:03 INFO BlockManagerMaster: BlockManagerMaster stopped 2019-07-30 09:35:03.76 - stderr> 19/07/30 09:35:03 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped! 2019-07-30 09:35:03.764 - stderr> 19/07/30 09:35:03 INFO SparkContext: Successfully stopped SparkContext 2019-07-30 09:35:03.765 - stderr> 19/07/30 09:35:03 INFO ShutdownHookManager: Shutdown hook called 2019-07-30 09:35:03.766 - stderr> 19/07/30 09:35:03 INFO ShutdownHookManager: Deleting directory /tmp/spark-2cfda2e6-3acb-4442-b235-604af57763b3/pyspark-2f853c67-07dc-418a-966c-6be6bad3acb8 2019-07-30 09:35:03.77 - stderr> 19/07/30 09:35:03 INFO ShutdownHookManager: Deleting directory /tmp/hive-v1_2-3db9434c-f1b4-4291-ab92-4b28a909c359 2019-07-30 09:35:03.776 - stderr> 19/07/30 09:35:03 INFO ShutdownHookManager: Deleting directory /tmp/spark-2cfda2e6-3acb-4442-b235-604af57763b3 2019-07-30 09:35:03.78 - stderr> 19/07/30 09:35:03 INFO ShutdownHookManager: Deleting directory /tmp/spark-b35b6dfe-4aaa-4ed6-bccc-a4e50b2f4671

sbt.ForkMain$ForkError: org.scalatest.exceptions.TestFailedException: spark-submit returned with exit code 1.
Command line: './bin/spark-submit' '--name' 'prepare testing tables' '--master' 'local[2]' '--conf' 'spark.ui.enabled=false' '--conf' 'spark.master.rest.enabled=false' '--conf' 'spark.sql.hive.metastore.version=1.2.1' '--conf' 'spark.sql.hive.metastore.jars=maven' '--conf' 'spark.sql.warehouse.dir=/home/jenkins/workspace/spark-master-test-sbt-hadoop-3.2/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-5ab906ca-f1e4-44d8-a45f-12c83e23e08b' '--conf' 'spark.sql.test.version.index=1' '--driver-java-options' '-Dderby.system.home=/home/jenkins/workspace/spark-master-test-sbt-hadoop-3.2/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-5ab906ca-f1e4-44d8-a45f-12c83e23e08b' '/home/jenkins/workspace/spark-master-test-sbt-hadoop-3.2/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test6000080996274345592.py'

2019-07-30 09:34:41.957 - stderr> 19/07/30 09:34:41 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
2019-07-30 09:34:42.748 - stderr> Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
2019-07-30 09:34:42.751 - stderr> 19/07/30 09:34:42 INFO SparkContext: Running Spark version 2.4.3
2019-07-30 09:34:42.781 - stderr> 19/07/30 09:34:42 INFO SparkContext: Submitted application: prepare testing tables
2019-07-30 09:34:42.84 - stderr> 19/07/30 09:34:42 INFO SecurityManager: Changing view acls to: jenkins
2019-07-30 09:34:42.84 - stderr> 19/07/30 09:34:42 INFO SecurityManager: Changing modify acls to: jenkins
2019-07-30 09:34:42.84 - stderr> 19/07/30 09:34:42 INFO SecurityManager: Changing view acls groups to: 
2019-07-30 09:34:42.841 - stderr> 19/07/30 09:34:42 INFO SecurityManager: Changing modify acls groups to: 
2019-07-30 09:34:42.841 - stderr> 19/07/30 09:34:42 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(jenkins); groups with view permissions: Set(); users  with modify permissions: Set(jenkins); groups with modify permissions: Set()
2019-07-30 09:34:43.15 - stderr> 19/07/30 09:34:43 INFO Utils: Successfully started service 'sparkDriver' on port 34599.
2019-07-30 09:34:43.183 - stderr> 19/07/30 09:34:43 INFO SparkEnv: Registering MapOutputTracker
2019-07-30 09:34:43.204 - stderr> 19/07/30 09:34:43 INFO SparkEnv: Registering BlockManagerMaster
2019-07-30 09:34:43.208 - stderr> 19/07/30 09:34:43 INFO BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
2019-07-30 09:34:43.208 - stderr> 19/07/30 09:34:43 INFO BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
2019-07-30 09:34:43.219 - stderr> 19/07/30 09:34:43 INFO DiskBlockManager: Created local directory at /tmp/blockmgr-bd6af0e6-04fc-4dfc-bb61-9f5bd92fdb3a
2019-07-30 09:34:43.241 - stderr> 19/07/30 09:34:43 INFO MemoryStore: MemoryStore started with capacity 366.3 MB
2019-07-30 09:34:43.257 - stderr> 19/07/30 09:34:43 INFO SparkEnv: Registering OutputCommitCoordinator
2019-07-30 09:34:43.353 - stderr> 19/07/30 09:34:43 INFO Executor: Starting executor ID driver on host localhost
2019-07-30 09:34:43.425 - stderr> 19/07/30 09:34:43 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 35997.
2019-07-30 09:34:43.426 - stderr> 19/07/30 09:34:43 INFO NettyBlockTransferService: Server created on amp-jenkins-worker-03.amp:35997
2019-07-30 09:34:43.427 - stderr> 19/07/30 09:34:43 INFO BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
2019-07-30 09:34:43.453 - stderr> 19/07/30 09:34:43 INFO BlockManagerMaster: Registering BlockManager BlockManagerId(driver, amp-jenkins-worker-03.amp, 35997, None)
2019-07-30 09:34:43.458 - stderr> 19/07/30 09:34:43 INFO BlockManagerMasterEndpoint: Registering block manager amp-jenkins-worker-03.amp:35997 with 366.3 MB RAM, BlockManagerId(driver, amp-jenkins-worker-03.amp, 35997, None)
2019-07-30 09:34:43.461 - stderr> 19/07/30 09:34:43 INFO BlockManagerMaster: Registered BlockManager BlockManagerId(driver, amp-jenkins-worker-03.amp, 35997, None)
2019-07-30 09:34:43.461 - stderr> 19/07/30 09:34:43 INFO BlockManager: Initialized BlockManager: BlockManagerId(driver, amp-jenkins-worker-03.amp, 35997, None)
2019-07-30 09:34:43.819 - stderr> 19/07/30 09:34:43 INFO SharedState: Setting hive.metastore.warehouse.dir ('null') to the value of spark.sql.warehouse.dir ('/home/jenkins/workspace/spark-master-test-sbt-hadoop-3.2/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-5ab906ca-f1e4-44d8-a45f-12c83e23e08b').
2019-07-30 09:34:43.82 - stderr> 19/07/30 09:34:43 INFO SharedState: Warehouse path is '/home/jenkins/workspace/spark-master-test-sbt-hadoop-3.2/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-5ab906ca-f1e4-44d8-a45f-12c83e23e08b'.
2019-07-30 09:34:44.389 - stderr> 19/07/30 09:34:44 INFO StateStoreCoordinatorRef: Registered StateStoreCoordinator endpoint
2019-07-30 09:34:46.933 - stderr> 19/07/30 09:34:46 INFO HiveUtils: Initializing HiveMetastoreConnection version 1.2.1 using maven.
2019-07-30 09:34:46.939 - stderr> http://www.datanucleus.org/downloads/maven2 added as a remote repository with the name: repo-1
2019-07-30 09:34:46.942 - stderr> Ivy Default Cache set to: /home/jenkins/.ivy2/cache
2019-07-30 09:34:46.942 - stderr> The jars for the packages stored in: /home/jenkins/.ivy2/jars
2019-07-30 09:34:46.975 - stderr> :: loading settings :: url = jar:file:/tmp/test-spark/spark-2.4.3/jars/ivy-2.4.0.jar!/org/apache/ivy/core/settings/ivysettings.xml
2019-07-30 09:34:47.028 - stderr> org.apache.hive#hive-metastore added as a dependency
2019-07-30 09:34:47.028 - stderr> org.apache.hive#hive-exec added as a dependency
2019-07-30 09:34:47.028 - stderr> org.apache.hive#hive-common added as a dependency
2019-07-30 09:34:47.029 - stderr> org.apache.hive#hive-serde added as a dependency
2019-07-30 09:34:47.029 - stderr> com.google.guava#guava added as a dependency
2019-07-30 09:34:47.029 - stderr> org.apache.hadoop#hadoop-client added as a dependency
2019-07-30 09:34:47.031 - stderr> :: resolving dependencies :: org.apache.spark#spark-submit-parent-57f852f1-25a6-4c8a-a84a-c38d787a00b7;1.0
2019-07-30 09:34:47.031 - stderr> 	confs: [default]
2019-07-30 09:34:47.43 - stderr> 	found org.apache.hive#hive-metastore;1.2.2 in central
2019-07-30 09:34:47.492 - stderr> 	found org.apache.hive#hive-serde;1.2.2 in central
2019-07-30 09:34:47.529 - stderr> 	found org.apache.hive#hive-common;1.2.2 in central
2019-07-30 09:34:47.564 - stderr> 	found org.apache.hive#hive-shims;1.2.2 in central
2019-07-30 09:34:47.605 - stderr> 	found org.apache.hive.shims#hive-shims-common;1.2.2 in central
2019-07-30 09:34:47.636 - stderr> 	found commons-logging#commons-logging;1.1.3 in user-list
2019-07-30 09:34:47.667 - stderr> 	found log4j#log4j;1.2.16 in user-list
2019-07-30 09:34:47.688 - stderr> 	found log4j#apache-log4j-extras;1.2.17 in user-list
2019-07-30 09:34:47.718 - stderr> 	found com.google.guava#guava;14.0.1 in user-list
2019-07-30 09:34:47.738 - stderr> 	found commons-lang#commons-lang;2.6 in user-list
2019-07-30 09:34:47.757 - stderr> 	found org.apache.thrift#libthrift;0.9.2 in user-list
2019-07-30 09:34:47.781 - stderr> 	found org.slf4j#slf4j-api;1.7.5 in user-list
2019-07-30 09:34:47.803 - stderr> 	found org.apache.httpcomponents#httpclient;4.4 in user-list
2019-07-30 09:34:47.826 - stderr> 	found org.apache.httpcomponents#httpcore;4.4 in user-list
2019-07-30 09:34:47.849 - stderr> 	found commons-codec#commons-codec;1.4 in user-list
2019-07-30 09:34:47.877 - stderr> 	found org.apache.zookeeper#zookeeper;3.4.6 in user-list
2019-07-30 09:34:47.904 - stderr> 	found org.slf4j#slf4j-log4j12;1.7.5 in user-list
2019-07-30 09:34:47.932 - stderr> 	found jline#jline;2.12 in user-list
2019-07-30 09:34:47.955 - stderr> 	found io.netty#netty;3.7.0.Final in central
2019-07-30 09:34:47.985 - stderr> 	found org.apache.hive.shims#hive-shims-0.20S;1.2.2 in central
2019-07-30 09:34:48.015 - stderr> 	found org.apache.hive.shims#hive-shims-0.23;1.2.2 in central
2019-07-30 09:34:48.048 - stderr> 	found org.apache.hadoop#hadoop-yarn-server-resourcemanager;2.6.0 in user-list
2019-07-30 09:34:48.077 - stderr> 	found org.apache.hadoop#hadoop-annotations;2.6.0 in user-list
2019-07-30 09:34:48.631 - stderr> 	found com.google.inject.extensions#guice-servlet;3.0 in central
2019-07-30 09:34:48.662 - stderr> 	found com.google.inject#guice;3.0 in user-list
2019-07-30 09:34:48.687 - stderr> 	found javax.inject#javax.inject;1 in user-list
2019-07-30 09:34:48.707 - stderr> 	found aopalliance#aopalliance;1.0 in user-list
2019-07-30 09:34:48.727 - stderr> 	found org.sonatype.sisu.inject#cglib;2.2.1-v20090111 in user-list
2019-07-30 09:34:48.755 - stderr> 	found asm#asm;3.1 in user-list
2019-07-30 09:34:48.78 - stderr> 	found com.google.protobuf#protobuf-java;2.5.0 in user-list
2019-07-30 09:34:48.819 - stderr> 	found commons-io#commons-io;2.4 in user-list
2019-07-30 09:34:48.846 - stderr> 	found com.sun.jersey#jersey-json;1.14 in user-list
2019-07-30 09:34:48.871 - stderr> 	found org.codehaus.jettison#jettison;1.1 in user-list
2019-07-30 09:34:48.894 - stderr> 	found com.sun.xml.bind#jaxb-impl;2.2.3-1 in user-list
2019-07-30 09:34:48.912 - stderr> 	found javax.xml.bind#jaxb-api;2.2.2 in user-list
2019-07-30 09:34:48.928 - stderr> 	found javax.xml.stream#stax-api;1.0-2 in user-list
2019-07-30 09:34:48.95 - stderr> 	found javax.activation#activation;1.1 in user-list
2019-07-30 09:34:48.966 - stderr> 	found org.codehaus.jackson#jackson-core-asl;1.9.2 in central
2019-07-30 09:34:48.981 - stderr> 	found org.codehaus.jackson#jackson-mapper-asl;1.9.2 in central
2019-07-30 09:34:48.997 - stderr> 	found org.codehaus.jackson#jackson-jaxrs;1.9.2 in user-list
2019-07-30 09:34:49.013 - stderr> 	found org.codehaus.jackson#jackson-xc;1.9.2 in user-list
2019-07-30 09:34:49.03 - stderr> 	found com.sun.jersey.contribs#jersey-guice;1.9 in user-list
2019-07-30 09:34:49.058 - stderr> 	found org.apache.hadoop#hadoop-yarn-common;2.6.0 in user-list
2019-07-30 09:34:49.088 - stderr> 	found org.apache.hadoop#hadoop-yarn-api;2.6.0 in user-list
2019-07-30 09:34:49.138 - stderr> 	found org.apache.commons#commons-compress;1.4.1 in user-list
2019-07-30 09:34:49.156 - stderr> 	found org.tukaani#xz;1.0 in user-list
2019-07-30 09:34:49.182 - stderr> 	found org.mortbay.jetty#jetty-util;6.1.26 in user-list
2019-07-30 09:34:49.201 - stderr> 	found com.sun.jersey#jersey-core;1.14 in user-list
2019-07-30 09:34:49.227 - stderr> 	found com.sun.jersey#jersey-client;1.9 in user-list
2019-07-30 09:34:49.26 - stderr> 	found commons-cli#commons-cli;1.2 in user-list
2019-07-30 09:34:49.297 - stderr> 	found com.sun.jersey#jersey-server;1.14 in user-list
2019-07-30 09:34:49.333 - stderr> 	found org.apache.hadoop#hadoop-yarn-server-common;2.6.0 in user-list
2019-07-30 09:34:49.368 - stderr> 	found org.fusesource.leveldbjni#leveldbjni-all;1.8 in user-list
2019-07-30 09:34:49.383 - stderr> 	found org.apache.hadoop#hadoop-yarn-server-applicationhistoryservice;2.6.0 in user-list
2019-07-30 09:34:49.421 - stderr> 	found commons-collections#commons-collections;3.2.2 in central
2019-07-30 09:34:49.437 - stderr> 	found org.apache.hadoop#hadoop-yarn-server-web-proxy;2.6.0 in user-list
2019-07-30 09:34:49.453 - stderr> 	found commons-httpclient#commons-httpclient;3.0.1 in central
2019-07-30 09:34:49.466 - stderr> 	found junit#junit;4.11 in user-list
2019-07-30 09:34:49.481 - stderr> 	found org.hamcrest#hamcrest-core;1.3 in user-list
2019-07-30 09:34:49.496 - stderr> 	found org.mortbay.jetty#jetty;6.1.26 in user-list
2019-07-30 09:34:49.53 - stderr> 	found org.apache.hive.shims#hive-shims-scheduler;1.2.2 in central
2019-07-30 09:34:49.548 - stderr> 	found joda-time#joda-time;2.5 in user-list
2019-07-30 09:34:49.561 - stderr> 	found org.apache.ant#ant;1.9.1 in user-list
2019-07-30 09:34:49.57 - stderr> 	found org.apache.ant#ant-launcher;1.9.1 in user-list
2019-07-30 09:34:49.579 - stderr> 	found org.json#json;20090211 in user-list
2019-07-30 09:34:49.589 - stderr> 	found com.google.code.findbugs#jsr305;3.0.0 in user-list
2019-07-30 09:34:49.599 - stderr> 	found org.apache.avro#avro;1.7.5 in user-list
2019-07-30 09:34:49.614 - stderr> 	found com.thoughtworks.paranamer#paranamer;2.3 in user-list
2019-07-30 09:34:49.623 - stderr> 	found org.xerial.snappy#snappy-java;1.0.5 in user-list
2019-07-30 09:34:49.632 - stderr> 	found net.sf.opencsv#opencsv;2.3 in user-list
2019-07-30 09:34:49.642 - stderr> 	found com.twitter#parquet-hadoop-bundle;1.6.0 in user-list
2019-07-30 09:34:49.652 - stderr> 	found com.jolbox#bonecp;0.8.0.RELEASE in user-list
2019-07-30 09:34:49.66 - stderr> 	found org.apache.derby#derby;10.10.2.0 in user-list
2019-07-30 09:34:49.669 - stderr> 	found org.datanucleus#datanucleus-api-jdo;3.2.6 in user-list
2019-07-30 09:34:49.678 - stderr> 	found org.datanucleus#datanucleus-core;3.2.10 in user-list
2019-07-30 09:34:49.689 - stderr> 	found org.datanucleus#datanucleus-rdbms;3.2.9 in user-list
2019-07-30 09:34:49.699 - stderr> 	found commons-pool#commons-pool;1.5.4 in user-list
2019-07-30 09:34:49.707 - stderr> 	found commons-dbcp#commons-dbcp;1.4 in central
2019-07-30 09:34:49.716 - stderr> 	found javax.jdo#jdo-api;3.0.1 in user-list
2019-07-30 09:34:49.725 - stderr> 	found javax.transaction#jta;1.1 in user-list
2019-07-30 09:34:49.733 - stderr> 	found org.antlr#antlr-runtime;3.4 in user-list
2019-07-30 09:34:49.74 - stderr> 	found org.antlr#stringtemplate;3.2.1 in user-list
2019-07-30 09:34:49.748 - stderr> 	found antlr#antlr;2.7.7 in user-list
2019-07-30 09:34:49.758 - stderr> 	found org.apache.thrift#libfb303;0.9.2 in user-list
2019-07-30 09:34:49.771 - stderr> 	found org.apache.hive#hive-exec;1.2.2 in central
2019-07-30 09:34:49.784 - stderr> 	found org.apache.hive#hive-ant;1.2.2 in central
2019-07-30 09:34:49.804 - stderr> 	found org.apache.velocity#velocity;1.5 in user-list
2019-07-30 09:34:49.823 - stderr> 	found oro#oro;2.0.8 in user-list
2019-07-30 09:34:49.876 - stderr> 	found org.antlr#ST4;4.0.4 in user-list
2019-07-30 09:34:49.89 - stderr> 	found org.apache.ivy#ivy;2.4.0 in user-list
2019-07-30 09:34:49.901 - stderr> 	found org.codehaus.groovy#groovy-all;2.1.6 in user-list
2019-07-30 09:34:49.912 - stderr> 	found org.apache.calcite#calcite-core;1.2.0-incubating in user-list
2019-07-30 09:34:49.922 - stderr> 	found org.apache.calcite#calcite-avatica;1.2.0-incubating in user-list
2019-07-30 09:34:49.931 - stderr> 	found org.apache.calcite#calcite-linq4j;1.2.0-incubating in user-list
2019-07-30 09:34:49.95 - stderr> 	found net.hydromatic#eigenbase-properties;1.1.5 in central
2019-07-30 09:34:49.959 - stderr> 	found org.codehaus.janino#janino;2.7.6 in central
2019-07-30 09:34:49.969 - stderr> 	found org.codehaus.janino#commons-compiler;2.7.6 in central
2019-07-30 09:34:49.98 - stderr> 	found stax#stax-api;1.0.1 in user-list
2019-07-30 09:34:49.993 - stderr> 	found org.apache.hadoop#hadoop-client;2.7.3 in central
2019-07-30 09:34:50.004 - stderr> 	found org.apache.hadoop#hadoop-common;2.7.3 in central
2019-07-30 09:34:50.022 - stderr> 	found org.apache.hadoop#hadoop-annotations;2.7.3 in central
2019-07-30 09:34:50.042 - stderr> 	found org.apache.commons#commons-math3;3.1.1 in user-list
2019-07-30 09:34:50.051 - stderr> 	found xmlenc#xmlenc;0.52 in user-list
2019-07-30 09:34:50.059 - stderr> 	found commons-httpclient#commons-httpclient;3.1 in user-list
2019-07-30 09:34:50.078 - stderr> 	found commons-net#commons-net;3.1 in user-list
2019-07-30 09:34:50.089 - stderr> 	found log4j#log4j;1.2.17 in user-list
2019-07-30 09:34:50.101 - stderr> 	found commons-configuration#commons-configuration;1.6 in user-list
2019-07-30 09:34:50.113 - stderr> 	found commons-digester#commons-digester;1.8 in user-list
2019-07-30 09:34:50.122 - stderr> 	found commons-beanutils#commons-beanutils;1.7.0 in user-list
2019-07-30 09:34:50.131 - stderr> 	found commons-beanutils#commons-beanutils-core;1.8.0 in user-list
2019-07-30 09:34:50.14 - stderr> 	found org.slf4j#slf4j-api;1.7.10 in user-list
2019-07-30 09:34:50.149 - stderr> 	found org.codehaus.jackson#jackson-core-asl;1.9.13 in user-list
2019-07-30 09:34:50.157 - stderr> 	found org.codehaus.jackson#jackson-mapper-asl;1.9.13 in user-list
2019-07-30 09:34:50.173 - stderr> 	found com.google.code.gson#gson;2.2.4 in user-list
2019-07-30 09:34:50.186 - stderr> 	found org.apache.hadoop#hadoop-auth;2.7.3 in central
2019-07-30 09:34:50.203 - stderr> 	found org.apache.directory.server#apacheds-kerberos-codec;2.0.0-M15 in user-list
2019-07-30 09:34:50.214 - stderr> 	found org.apache.directory.server#apacheds-i18n;2.0.0-M15 in user-list
2019-07-30 09:34:50.224 - stderr> 	found org.apache.directory.api#api-asn1-api;1.0.0-M20 in user-list
2019-07-30 09:34:50.237 - stderr> 	found org.apache.directory.api#api-util;1.0.0-M20 in user-list
2019-07-30 09:34:50.253 - stderr> 	found org.apache.htrace#htrace-core;3.1.0-incubating in central
2019-07-30 09:34:50.268 - stderr> 	found javax.servlet.jsp#jsp-api;2.1 in user-list
2019-07-30 09:34:50.282 - stderr> 	found org.slf4j#slf4j-log4j12;1.7.10 in user-list
2019-07-30 09:34:50.297 - stderr> 	found org.apache.hadoop#hadoop-hdfs;2.7.3 in central
2019-07-30 09:34:50.324 - stderr> 	found io.netty#netty-all;4.0.23.Final in user-list
2019-07-30 09:34:50.349 - stderr> 	found xerces#xercesImpl;2.9.1 in central
2019-07-30 09:34:50.357 - stderr> 	found xml-apis#xml-apis;1.3.04 in user-list
2019-07-30 09:34:50.371 - stderr> 	found org.apache.hadoop#hadoop-mapreduce-client-app;2.7.3 in central
2019-07-30 09:34:50.38 - stderr> 	found org.apache.hadoop#hadoop-mapreduce-client-common;2.7.3 in central
2019-07-30 09:34:50.391 - stderr> 	found org.apache.hadoop#hadoop-yarn-common;2.7.3 in central
2019-07-30 09:34:50.404 - stderr> 	found org.apache.hadoop#hadoop-yarn-api;2.7.3 in central
2019-07-30 09:34:50.476 - stderr> 	found org.codehaus.jackson#jackson-jaxrs;1.9.13 in user-list
2019-07-30 09:34:50.487 - stderr> 	found org.codehaus.jackson#jackson-xc;1.9.13 in user-list
2019-07-30 09:34:50.545 - stderr> 	found org.apache.hadoop#hadoop-yarn-client;2.7.3 in central
2019-07-30 09:34:50.563 - stderr> 	found org.apache.hadoop#hadoop-mapreduce-client-core;2.7.3 in central
2019-07-30 09:34:50.581 - stderr> 	found org.apache.hadoop#hadoop-yarn-server-common;2.7.3 in central
2019-07-30 09:34:50.603 - stderr> 	found org.apache.hadoop#hadoop-mapreduce-client-shuffle;2.7.3 in central
2019-07-30 09:34:50.619 - stderr> 	found org.apache.hadoop#hadoop-mapreduce-client-jobclient;2.7.3 in central
2019-07-30 09:34:50.717 - stderr> :: resolution report :: resolve 3630ms :: artifacts dl 55ms
2019-07-30 09:34:50.718 - stderr> 	:: modules in use:
2019-07-30 09:34:50.719 - stderr> 	antlr#antlr;2.7.7 from user-list in [default]
2019-07-30 09:34:50.72 - stderr> 	aopalliance#aopalliance;1.0 from user-list in [default]
2019-07-30 09:34:50.72 - stderr> 	asm#asm;3.1 from user-list in [default]
2019-07-30 09:34:50.72 - stderr> 	com.google.code.findbugs#jsr305;3.0.0 from user-list in [default]
2019-07-30 09:34:50.721 - stderr> 	com.google.code.gson#gson;2.2.4 from user-list in [default]
2019-07-30 09:34:50.721 - stderr> 	com.google.guava#guava;14.0.1 from user-list in [default]
2019-07-30 09:34:50.721 - stderr> 	com.google.inject#guice;3.0 from user-list in [default]
2019-07-30 09:34:50.721 - stderr> 	com.google.inject.extensions#guice-servlet;3.0 from central in [default]
2019-07-30 09:34:50.722 - stderr> 	com.google.protobuf#protobuf-java;2.5.0 from user-list in [default]
2019-07-30 09:34:50.722 - stderr> 	com.jolbox#bonecp;0.8.0.RELEASE from user-list in [default]
2019-07-30 09:34:50.722 - stderr> 	com.sun.jersey#jersey-client;1.9 from user-list in [default]
2019-07-30 09:34:50.722 - stderr> 	com.sun.jersey#jersey-core;1.14 from user-list in [default]
2019-07-30 09:34:50.722 - stderr> 	com.sun.jersey#jersey-json;1.14 from user-list in [default]
2019-07-30 09:34:50.722 - stderr> 	com.sun.jersey#jersey-server;1.14 from user-list in [default]
2019-07-30 09:34:50.723 - stderr> 	com.sun.jersey.contribs#jersey-guice;1.9 from user-list in [default]
2019-07-30 09:34:50.723 - stderr> 	com.sun.xml.bind#jaxb-impl;2.2.3-1 from user-list in [default]
2019-07-30 09:34:50.723 - stderr> 	com.thoughtworks.paranamer#paranamer;2.3 from user-list in [default]
2019-07-30 09:34:50.723 - stderr> 	com.twitter#parquet-hadoop-bundle;1.6.0 from user-list in [default]
2019-07-30 09:34:50.723 - stderr> 	commons-beanutils#commons-beanutils;1.7.0 from user-list in [default]
2019-07-30 09:34:50.723 - stderr> 	commons-beanutils#commons-beanutils-core;1.8.0 from user-list in [default]
2019-07-30 09:34:50.724 - stderr> 	commons-cli#commons-cli;1.2 from user-list in [default]
2019-07-30 09:34:50.724 - stderr> 	commons-codec#commons-codec;1.4 from user-list in [default]
2019-07-30 09:34:50.724 - stderr> 	commons-collections#commons-collections;3.2.2 from central in [default]
2019-07-30 09:34:50.724 - stderr> 	commons-configuration#commons-configuration;1.6 from user-list in [default]
2019-07-30 09:34:50.724 - stderr> 	commons-dbcp#commons-dbcp;1.4 from central in [default]
2019-07-30 09:34:50.724 - stderr> 	commons-digester#commons-digester;1.8 from user-list in [default]
2019-07-30 09:34:50.725 - stderr> 	commons-httpclient#commons-httpclient;3.1 from user-list in [default]
2019-07-30 09:34:50.725 - stderr> 	commons-io#commons-io;2.4 from user-list in [default]
2019-07-30 09:34:50.725 - stderr> 	commons-lang#commons-lang;2.6 from user-list in [default]
2019-07-30 09:34:50.725 - stderr> 	commons-logging#commons-logging;1.1.3 from user-list in [default]
2019-07-30 09:34:50.725 - stderr> 	commons-net#commons-net;3.1 from user-list in [default]
2019-07-30 09:34:50.726 - stderr> 	commons-pool#commons-pool;1.5.4 from user-list in [default]
2019-07-30 09:34:50.726 - stderr> 	io.netty#netty;3.7.0.Final from central in [default]
2019-07-30 09:34:50.726 - stderr> 	io.netty#netty-all;4.0.23.Final from user-list in [default]
2019-07-30 09:34:50.726 - stderr> 	javax.activation#activation;1.1 from user-list in [default]
2019-07-30 09:34:50.727 - stderr> 	javax.inject#javax.inject;1 from user-list in [default]
2019-07-30 09:34:50.727 - stderr> 	javax.jdo#jdo-api;3.0.1 from user-list in [default]
2019-07-30 09:34:50.727 - stderr> 	javax.servlet.jsp#jsp-api;2.1 from user-list in [default]
2019-07-30 09:34:50.727 - stderr> 	javax.transaction#jta;1.1 from user-list in [default]
2019-07-30 09:34:50.727 - stderr> 	javax.xml.bind#jaxb-api;2.2.2 from user-list in [default]
2019-07-30 09:34:50.728 - stderr> 	javax.xml.stream#stax-api;1.0-2 from user-list in [default]
2019-07-30 09:34:50.728 - stderr> 	jline#jline;2.12 from user-list in [default]
2019-07-30 09:34:50.728 - stderr> 	joda-time#joda-time;2.5 from user-list in [default]
2019-07-30 09:34:50.728 - stderr> 	log4j#apache-log4j-extras;1.2.17 from user-list in [default]
2019-07-30 09:34:50.728 - stderr> 	log4j#log4j;1.2.17 from user-list in [default]
2019-07-30 09:34:50.729 - stderr> 	net.hydromatic#eigenbase-properties;1.1.5 from central in [default]
2019-07-30 09:34:50.729 - stderr> 	net.sf.opencsv#opencsv;2.3 from user-list in [default]
2019-07-30 09:34:50.729 - stderr> 	org.antlr#ST4;4.0.4 from user-list in [default]
2019-07-30 09:34:50.729 - stderr> 	org.antlr#antlr-runtime;3.4 from user-list in [default]
2019-07-30 09:34:50.729 - stderr> 	org.antlr#stringtemplate;3.2.1 from user-list in [default]
2019-07-30 09:34:50.73 - stderr> 	org.apache.ant#ant;1.9.1 from user-list in [default]
2019-07-30 09:34:50.73 - stderr> 	org.apache.ant#ant-launcher;1.9.1 from user-list in [default]
2019-07-30 09:34:50.73 - stderr> 	org.apache.avro#avro;1.7.5 from user-list in [default]
2019-07-30 09:34:50.73 - stderr> 	org.apache.calcite#calcite-avatica;1.2.0-incubating from user-list in [default]
2019-07-30 09:34:50.73 - stderr> 	org.apache.calcite#calcite-core;1.2.0-incubating from user-list in [default]
2019-07-30 09:34:50.731 - stderr> 	org.apache.calcite#calcite-linq4j;1.2.0-incubating from user-list in [default]
2019-07-30 09:34:50.731 - stderr> 	org.apache.commons#commons-compress;1.4.1 from user-list in [default]
2019-07-30 09:34:50.731 - stderr> 	org.apache.commons#commons-math3;3.1.1 from user-list in [default]
2019-07-30 09:34:50.731 - stderr> 	org.apache.derby#derby;10.10.2.0 from user-list in [default]
2019-07-30 09:34:50.731 - stderr> 	org.apache.directory.api#api-asn1-api;1.0.0-M20 from user-list in [default]
2019-07-30 09:34:50.732 - stderr> 	org.apache.directory.api#api-util;1.0.0-M20 from user-list in [default]
2019-07-30 09:34:50.732 - stderr> 	org.apache.directory.server#apacheds-i18n;2.0.0-M15 from user-list in [default]
2019-07-30 09:34:50.732 - stderr> 	org.apache.directory.server#apacheds-kerberos-codec;2.0.0-M15 from user-list in [default]
2019-07-30 09:34:50.732 - stderr> 	org.apache.hadoop#hadoop-annotations;2.7.3 from central in [default]
2019-07-30 09:34:50.733 - stderr> 	org.apache.hadoop#hadoop-auth;2.7.3 from central in [default]
2019-07-30 09:34:50.733 - stderr> 	org.apache.hadoop#hadoop-client;2.7.3 from central in [default]
2019-07-30 09:34:50.733 - stderr> 	org.apache.hadoop#hadoop-common;2.7.3 from central in [default]
2019-07-30 09:34:50.733 - stderr> 	org.apache.hadoop#hadoop-hdfs;2.7.3 from central in [default]
2019-07-30 09:34:50.733 - stderr> 	org.apache.hadoop#hadoop-mapreduce-client-app;2.7.3 from central in [default]
2019-07-30 09:34:50.734 - stderr> 	org.apache.hadoop#hadoop-mapreduce-client-common;2.7.3 from central in [default]
2019-07-30 09:34:50.734 - stderr> 	org.apache.hadoop#hadoop-mapreduce-client-core;2.7.3 from central in [default]
2019-07-30 09:34:50.734 - stderr> 	org.apache.hadoop#hadoop-mapreduce-client-jobclient;2.7.3 from central in [default]
2019-07-30 09:34:50.734 - stderr> 	org.apache.hadoop#hadoop-mapreduce-client-shuffle;2.7.3 from central in [default]
2019-07-30 09:34:50.735 - stderr> 	org.apache.hadoop#hadoop-yarn-api;2.7.3 from central in [default]
2019-07-30 09:34:50.735 - stderr> 	org.apache.hadoop#hadoop-yarn-client;2.7.3 from central in [default]
2019-07-30 09:34:50.735 - stderr> 	org.apache.hadoop#hadoop-yarn-common;2.7.3 from central in [default]
2019-07-30 09:34:50.735 - stderr> 	org.apache.hadoop#hadoop-yarn-server-applicationhistoryservice;2.6.0 from user-list in [default]
2019-07-30 09:34:50.735 - stderr> 	org.apache.hadoop#hadoop-yarn-server-common;2.7.3 from central in [default]
2019-07-30 09:34:50.735 - stderr> 	org.apache.hadoop#hadoop-yarn-server-resourcemanager;2.6.0 from user-list in [default]
2019-07-30 09:34:50.736 - stderr> 	org.apache.hadoop#hadoop-yarn-server-web-proxy;2.6.0 from user-list in [default]
2019-07-30 09:34:50.736 - stderr> 	org.apache.hive#hive-ant;1.2.2 from central in [default]
2019-07-30 09:34:50.736 - stderr> 	org.apache.hive#hive-common;1.2.2 from central in [default]
2019-07-30 09:34:50.736 - stderr> 	org.apache.hive#hive-exec;1.2.2 from central in [default]
2019-07-30 09:34:50.736 - stderr> 	org.apache.hive#hive-metastore;1.2.2 from central in [default]
2019-07-30 09:34:50.737 - stderr> 	org.apache.hive#hive-serde;1.2.2 from central in [default]
2019-07-30 09:34:50.737 - stderr> 	org.apache.hive#hive-shims;1.2.2 from central in [default]
2019-07-30 09:34:50.737 - stderr> 	org.apache.hive.shims#hive-shims-0.20S;1.2.2 from central in [default]
2019-07-30 09:34:50.737 - stderr> 	org.apache.hive.shims#hive-shims-0.23;1.2.2 from central in [default]
2019-07-30 09:34:50.738 - stderr> 	org.apache.hive.shims#hive-shims-common;1.2.2 from central in [default]
2019-07-30 09:34:50.738 - stderr> 	org.apache.hive.shims#hive-shims-scheduler;1.2.2 from central in [default]
2019-07-30 09:34:50.738 - stderr> 	org.apache.htrace#htrace-core;3.1.0-incubating from central in [default]
2019-07-30 09:34:50.738 - stderr> 	org.apache.httpcomponents#httpclient;4.4 from user-list in [default]
2019-07-30 09:34:50.738 - stderr> 	org.apache.httpcomponents#httpcore;4.4 from user-list in [default]
2019-07-30 09:34:50.739 - stderr> 	org.apache.ivy#ivy;2.4.0 from user-list in [default]
2019-07-30 09:34:50.739 - stderr> 	org.apache.thrift#libfb303;0.9.2 from user-list in [default]
2019-07-30 09:34:50.739 - stderr> 	org.apache.thrift#libthrift;0.9.2 from user-list in [default]
2019-07-30 09:34:50.739 - stderr> 	org.apache.velocity#velocity;1.5 from user-list in [default]
2019-07-30 09:34:50.739 - stderr> 	org.apache.zookeeper#zookeeper;3.4.6 from user-list in [default]
2019-07-30 09:34:50.739 - stderr> 	org.codehaus.groovy#groovy-all;2.1.6 from user-list in [default]
2019-07-30 09:34:50.739 - stderr> 	org.codehaus.jackson#jackson-core-asl;1.9.13 from user-list in [default]
2019-07-30 09:34:50.74 - stderr> 	org.codehaus.jackson#jackson-jaxrs;1.9.13 from user-list in [default]
2019-07-30 09:34:50.74 - stderr> 	org.codehaus.jackson#jackson-mapper-asl;1.9.13 from user-list in [default]
2019-07-30 09:34:50.74 - stderr> 	org.codehaus.jackson#jackson-xc;1.9.13 from user-list in [default]
2019-07-30 09:34:50.74 - stderr> 	org.codehaus.janino#commons-compiler;2.7.6 from central in [default]
2019-07-30 09:34:50.74 - stderr> 	org.codehaus.janino#janino;2.7.6 from central in [default]
2019-07-30 09:34:50.74 - stderr> 	org.codehaus.jettison#jettison;1.1 from user-list in [default]
2019-07-30 09:34:50.74 - stderr> 	org.datanucleus#datanucleus-api-jdo;3.2.6 from user-list in [default]
2019-07-30 09:34:50.74 - stderr> 	org.datanucleus#datanucleus-core;3.2.10 from user-list in [default]
2019-07-30 09:34:50.741 - stderr> 	org.datanucleus#datanucleus-rdbms;3.2.9 from user-list in [default]
2019-07-30 09:34:50.741 - stderr> 	org.fusesource.leveldbjni#leveldbjni-all;1.8 from user-list in [default]
2019-07-30 09:34:50.741 - stderr> 	org.json#json;20090211 from user-list in [default]
2019-07-30 09:34:50.741 - stderr> 	org.mortbay.jetty#jetty;6.1.26 from user-list in [default]
2019-07-30 09:34:50.741 - stderr> 	org.mortbay.jetty#jetty-util;6.1.26 from user-list in [default]
2019-07-30 09:34:50.741 - stderr> 	org.slf4j#slf4j-api;1.7.10 from user-list in [default]
2019-07-30 09:34:50.741 - stderr> 	org.slf4j#slf4j-log4j12;1.7.10 from user-list in [default]
2019-07-30 09:34:50.742 - stderr> 	org.sonatype.sisu.inject#cglib;2.2.1-v20090111 from user-list in [default]
2019-07-30 09:34:50.742 - stderr> 	org.tukaani#xz;1.0 from user-list in [default]
2019-07-30 09:34:50.742 - stderr> 	org.xerial.snappy#snappy-java;1.0.5 from user-list in [default]
2019-07-30 09:34:50.742 - stderr> 	oro#oro;2.0.8 from user-list in [default]
2019-07-30 09:34:50.742 - stderr> 	stax#stax-api;1.0.1 from user-list in [default]
2019-07-30 09:34:50.742 - stderr> 	xerces#xercesImpl;2.9.1 from central in [default]
2019-07-30 09:34:50.742 - stderr> 	xml-apis#xml-apis;1.3.04 from user-list in [default]
2019-07-30 09:34:50.742 - stderr> 	xmlenc#xmlenc;0.52 from user-list in [default]
2019-07-30 09:34:50.743 - stderr> 	:: evicted modules:
2019-07-30 09:34:50.743 - stderr> 	log4j#log4j;1.2.16 by [log4j#log4j;1.2.17] in [default]
2019-07-30 09:34:50.743 - stderr> 	org.slf4j#slf4j-api;1.7.5 by [org.slf4j#slf4j-api;1.7.10] in [default]
2019-07-30 09:34:50.743 - stderr> 	org.slf4j#slf4j-log4j12;1.7.5 by [org.slf4j#slf4j-log4j12;1.7.10] in [default]
2019-07-30 09:34:50.743 - stderr> 	org.apache.hadoop#hadoop-annotations;2.6.0 by [org.apache.hadoop#hadoop-annotations;2.7.3] in [default]
2019-07-30 09:34:50.743 - stderr> 	org.codehaus.jackson#jackson-core-asl;1.9.2 by [org.codehaus.jackson#jackson-core-asl;1.9.13] in [default]
2019-07-30 09:34:50.743 - stderr> 	org.codehaus.jackson#jackson-mapper-asl;1.9.2 by [org.codehaus.jackson#jackson-mapper-asl;1.9.13] in [default]
2019-07-30 09:34:50.743 - stderr> 	org.codehaus.jackson#jackson-jaxrs;1.9.2 by [org.codehaus.jackson#jackson-jaxrs;1.9.13] in [default]
2019-07-30 09:34:50.743 - stderr> 	org.codehaus.jackson#jackson-xc;1.9.2 by [org.codehaus.jackson#jackson-xc;1.9.13] in [default]
2019-07-30 09:34:50.743 - stderr> 	org.apache.hadoop#hadoop-yarn-common;2.6.0 by [org.apache.hadoop#hadoop-yarn-common;2.7.3] in [default]
2019-07-30 09:34:50.743 - stderr> 	org.apache.hadoop#hadoop-yarn-api;2.6.0 by [org.apache.hadoop#hadoop-yarn-api;2.7.3] in [default]
2019-07-30 09:34:50.743 - stderr> 	org.apache.hadoop#hadoop-yarn-server-common;2.6.0 by [org.apache.hadoop#hadoop-yarn-server-common;2.7.3] in [default]
2019-07-30 09:34:50.743 - stderr> 	commons-httpclient#commons-httpclient;3.0.1 by [commons-httpclient#commons-httpclient;3.1] in [default]
2019-07-30 09:34:50.747 - stderr> 	junit#junit;4.11 transitively in [default]
2019-07-30 09:34:50.747 - stderr> 	org.hamcrest#hamcrest-core;1.3 transitively in [default]
2019-07-30 09:34:50.747 - stderr> 	com.google.code.findbugs#jsr305;1.3.9 by [com.google.code.findbugs#jsr305;3.0.0] in [default]
2019-07-30 09:34:50.747 - stderr> 	com.google.guava#guava;11.0.2 by [com.google.guava#guava;14.0.1] in [default]
2019-07-30 09:34:50.748 - stderr> 	org.apache.avro#avro;1.7.4 by [org.apache.avro#avro;1.7.5] in [default]
2019-07-30 09:34:50.748 - stderr> 	org.apache.httpcomponents#httpclient;4.2.5 by [org.apache.httpcomponents#httpclient;4.4] in [default]
2019-07-30 09:34:50.748 - stderr> 	io.netty#netty;3.6.2.Final by [io.netty#netty;3.7.0.Final] in [default]
2019-07-30 09:34:50.748 - stderr> 	com.sun.jersey#jersey-core;1.9 by [com.sun.jersey#jersey-core;1.14] in [default]
2019-07-30 09:34:50.748 - stderr> 	com.sun.jersey#jersey-server;1.9 by [com.sun.jersey#jersey-server;1.14] in [default]
2019-07-30 09:34:50.748 - stderr> 	com.sun.jersey#jersey-json;1.9 by [com.sun.jersey#jersey-json;1.14] in [default]
2019-07-30 09:34:50.748 - stderr> 	---------------------------------------------------------------------
2019-07-30 09:34:50.748 - stderr> 	|                  |            modules            ||   artifacts   |
2019-07-30 09:34:50.748 - stderr> 	|       conf       | number| search|dwnlded|evicted|| number|dwnlded|
2019-07-30 09:34:50.748 - stderr> 	---------------------------------------------------------------------
2019-07-30 09:34:50.749 - stderr> 	|      default     |  145  |   1   |   1   |   22  ||  123  |   0   |
2019-07-30 09:34:50.749 - stderr> 	---------------------------------------------------------------------
2019-07-30 09:34:50.789 - stderr> 
2019-07-30 09:34:50.789 - stderr> :: problems summary ::
2019-07-30 09:34:50.789 - stderr> :::: ERRORS
2019-07-30 09:34:50.789 - stderr> 	unknown resolver null
2019-07-30 09:34:50.79 - stderr> 
2019-07-30 09:34:50.79 - stderr> 
2019-07-30 09:34:50.79 - stderr> :: USE VERBOSE OR DEBUG MESSAGE LEVEL FOR MORE DETAILS
2019-07-30 09:34:50.791 - stderr> :: retrieving :: org.apache.spark#spark-submit-parent-57f852f1-25a6-4c8a-a84a-c38d787a00b7
2019-07-30 09:34:50.791 - stderr> 	confs: [default]
2019-07-30 09:34:50.836 - stderr> 	0 artifacts copied, 123 already retrieved (0kB/45ms)
2019-07-30 09:34:50.998 - stderr> 19/07/30 09:34:50 INFO IsolatedClientLoader: Downloaded metastore jars to /tmp/hive-v1_2-3db9434c-f1b4-4291-ab92-4b28a909c359
2019-07-30 09:34:51.579 - stderr> 19/07/30 09:34:51 INFO HiveMetaStore: 0: Opening raw store with implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
2019-07-30 09:34:51.606 - stderr> 19/07/30 09:34:51 INFO ObjectStore: ObjectStore, initialize called
2019-07-30 09:34:51.719 - stderr> 19/07/30 09:34:51 INFO Persistence: Property hive.metastore.integral.jdo.pushdown unknown - will be ignored
2019-07-30 09:34:51.719 - stderr> 19/07/30 09:34:51 INFO Persistence: Property datanucleus.cache.level2 unknown - will be ignored
2019-07-30 09:35:01.404 - stderr> 19/07/30 09:35:01 INFO ObjectStore: Setting MetaStore object pin classes with hive.metastore.cache.pinobjtypes="Table,StorageDescriptor,SerDeInfo,Partition,Database,Type,FieldSchema,Order"
2019-07-30 09:35:02.862 - stderr> 19/07/30 09:35:02 INFO Datastore: The class "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as "embedded-only" so does not have its own datastore table.
2019-07-30 09:35:02.863 - stderr> 19/07/30 09:35:02 INFO Datastore: The class "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as "embedded-only" so does not have its own datastore table.
2019-07-30 09:35:03.074 - stderr> 19/07/30 09:35:03 INFO Datastore: The class "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as "embedded-only" so does not have its own datastore table.
2019-07-30 09:35:03.074 - stderr> 19/07/30 09:35:03 INFO Datastore: The class "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as "embedded-only" so does not have its own datastore table.
2019-07-30 09:35:03.154 - stderr> 19/07/30 09:35:03 INFO Query: Reading in results for query "org.datanucleus.store.rdbms.query.SQLQuery@0" since the connection used is closing
2019-07-30 09:35:03.155 - stderr> 19/07/30 09:35:03 INFO MetaStoreDirectSql: Using direct SQL, underlying DB is DERBY
2019-07-30 09:35:03.158 - stderr> 19/07/30 09:35:03 INFO ObjectStore: Initialized ObjectStore
2019-07-30 09:35:03.331 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: Added admin role in metastore
2019-07-30 09:35:03.333 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: Added public role in metastore
2019-07-30 09:35:03.377 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: No user is added in admin role, since config is empty
2019-07-30 09:35:03.465 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_all_databases
2019-07-30 09:35:03.466 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_all_databases	
2019-07-30 09:35:03.481 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_functions: db=default pat=*
2019-07-30 09:35:03.481 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_functions: db=default pat=*	
2019-07-30 09:35:03.482 - stderr> 19/07/30 09:35:03 INFO Datastore: The class "org.apache.hadoop.hive.metastore.model.MResourceUri" is tagged as "embedded-only" so does not have its own datastore table.
2019-07-30 09:35:03.542 - stderr> 19/07/30 09:35:03 INFO SessionState: Created local directory: /tmp/1b2706a2-b4a8-44ee-a479-de703182600d_resources
2019-07-30 09:35:03.546 - stderr> 19/07/30 09:35:03 INFO SessionState: Created HDFS directory: /tmp/hive/jenkins/1b2706a2-b4a8-44ee-a479-de703182600d
2019-07-30 09:35:03.55 - stderr> 19/07/30 09:35:03 INFO SessionState: Created local directory: /tmp/jenkins/1b2706a2-b4a8-44ee-a479-de703182600d
2019-07-30 09:35:03.554 - stderr> 19/07/30 09:35:03 INFO SessionState: Created HDFS directory: /tmp/hive/jenkins/1b2706a2-b4a8-44ee-a479-de703182600d/_tmp_space.db
2019-07-30 09:35:03.557 - stderr> 19/07/30 09:35:03 INFO HiveClientImpl: Warehouse location for Hive client (version 1.2.2) is /home/jenkins/workspace/spark-master-test-sbt-hadoop-3.2/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-5ab906ca-f1e4-44d8-a45f-12c83e23e08b
2019-07-30 09:35:03.566 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_database: default
2019-07-30 09:35:03.566 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_database: default	
2019-07-30 09:35:03.572 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_table : db=default tbl=data_source_tbl_1
2019-07-30 09:35:03.572 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_table : db=default tbl=data_source_tbl_1	
2019-07-30 09:35:03.584 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_database: default
2019-07-30 09:35:03.584 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_database: default	
2019-07-30 09:35:03.587 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_database: default
2019-07-30 09:35:03.587 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_database: default	
2019-07-30 09:35:03.611 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_database: default
2019-07-30 09:35:03.612 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_database: default	
2019-07-30 09:35:03.614 - stderr> 19/07/30 09:35:03 INFO HiveMetaStore: 0: get_database: default
2019-07-30 09:35:03.614 - stderr> 19/07/30 09:35:03 INFO audit: ugi=jenkins	ip=unknown-ip-addr	cmd=get_database: default	
2019-07-30 09:35:03.648 - stdout> Traceback (most recent call last):
2019-07-30 09:35:03.648 - stdout>   File "/home/jenkins/workspace/spark-master-test-sbt-hadoop-3.2/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test6000080996274345592.py", line 8, in <module>
2019-07-30 09:35:03.648 - stdout>     spark.sql("create table data_source_tbl_{} using json as select 1 i".format(version_index))
2019-07-30 09:35:03.648 - stdout>   File "/tmp/test-spark/spark-2.4.3/python/lib/pyspark.zip/pyspark/sql/session.py", line 767, in sql
2019-07-30 09:35:03.648 - stdout>   File "/tmp/test-spark/spark-2.4.3/python/lib/py4j-0.10.7-src.zip/py4j/java_gateway.py", line 1257, in __call__
2019-07-30 09:35:03.648 - stdout>   File "/tmp/test-spark/spark-2.4.3/python/lib/pyspark.zip/pyspark/sql/utils.py", line 63, in deco
2019-07-30 09:35:03.648 - stdout>   File "/tmp/test-spark/spark-2.4.3/python/lib/py4j-0.10.7-src.zip/py4j/protocol.py", line 328, in get_return_value
2019-07-30 09:35:03.65 - stdout> py4j.protocol.Py4JJavaError: An error occurred while calling o29.sql.
2019-07-30 09:35:03.65 - stdout> : java.util.ServiceConfigurationError: org.apache.spark.sql.sources.DataSourceRegister: Error reading configuration file
2019-07-30 09:35:03.65 - stdout> 	at java.util.ServiceLoader.fail(ServiceLoader.java:232)
2019-07-30 09:35:03.65 - stdout> 	at java.util.ServiceLoader.parse(ServiceLoader.java:309)
2019-07-30 09:35:03.65 - stdout> 	at java.util.ServiceLoader.access$200(ServiceLoader.java:185)
2019-07-30 09:35:03.65 - stdout> 	at java.util.ServiceLoader$LazyIterator.hasNextService(ServiceLoader.java:357)
2019-07-30 09:35:03.65 - stdout> 	at java.util.ServiceLoader$LazyIterator.hasNext(ServiceLoader.java:393)
2019-07-30 09:35:03.65 - stdout> 	at java.util.ServiceLoader$1.hasNext(ServiceLoader.java:474)
2019-07-30 09:35:03.65 - stdout> 	at scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:42)
2019-07-30 09:35:03.65 - stdout> 	at scala.collection.Iterator$class.foreach(Iterator.scala:891)
2019-07-30 09:35:03.65 - stdout> 	at scala.collection.AbstractIterator.foreach(Iterator.scala:1334)
2019-07-30 09:35:03.65 - stdout> 	at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
2019-07-30 09:35:03.65 - stdout> 	at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
2019-07-30 09:35:03.65 - stdout> 	at scala.collection.TraversableLike$class.filterImpl(TraversableLike.scala:247)
2019-07-30 09:35:03.65 - stdout> 	at scala.collection.TraversableLike$class.filter(TraversableLike.scala:259)
2019-07-30 09:35:03.65 - stdout> 	at scala.collection.AbstractTraversable.filter(Traversable.scala:104)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.execution.datasources.DataSource$.lookupDataSource(DataSource.scala:630)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.execution.datasources.DataSource.providingClass$lzycompute(DataSource.scala:94)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.execution.datasources.DataSource.providingClass(DataSource.scala:93)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.execution.datasources.DataSource.writeAndRead(DataSource.scala:482)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.execution.command.CreateDataSourceTableAsSelectCommand.saveDataIntoTable(createDataSourceTables.scala:217)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.execution.command.CreateDataSourceTableAsSelectCommand.run(createDataSourceTables.scala:176)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult$lzycompute(commands.scala:104)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.execution.command.DataWritingCommandExec.sideEffectResult(commands.scala:102)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.execution.command.DataWritingCommandExec.executeCollect(commands.scala:115)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:194)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:194)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.Dataset$$anonfun$53.apply(Dataset.scala:3364)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:78)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:125)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:73)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3363)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.Dataset.<init>(Dataset.scala:194)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:79)
2019-07-30 09:35:03.65 - stdout> 	at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
2019-07-30 09:35:03.65 - stdout> 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
2019-07-30 09:35:03.65 - stdout> 	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
2019-07-30 09:35:03.65 - stdout> 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
2019-07-30 09:35:03.65 - stdout> 	at java.lang.reflect.Method.invoke(Method.java:498)
2019-07-30 09:35:03.65 - stdout> 	at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
2019-07-30 09:35:03.65 - stdout> 	at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
2019-07-30 09:35:03.65 - stdout> 	at py4j.Gateway.invoke(Gateway.java:282)
2019-07-30 09:35:03.65 - stdout> 	at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
2019-07-30 09:35:03.65 - stdout> 	at py4j.commands.CallCommand.execute(CallCommand.java:79)
2019-07-30 09:35:03.65 - stdout> 	at py4j.GatewayConnection.run(GatewayConnection.java:238)
2019-07-30 09:35:03.65 - stdout> 	at java.lang.Thread.run(Thread.java:748)
2019-07-30 09:35:03.65 - stdout> Caused by: java.io.FileNotFoundException: /tmp/test-spark/spark-2.4.3/jars/spark-sql_2.11-2.4.3.jar (No such file or directory)
2019-07-30 09:35:03.65 - stdout> 	at java.util.zip.ZipFile.open(Native Method)
2019-07-30 09:35:03.65 - stdout> 	at java.util.zip.ZipFile.<init>(ZipFile.java:225)
2019-07-30 09:35:03.65 - stdout> 	at java.util.zip.ZipFile.<init>(ZipFile.java:155)
2019-07-30 09:35:03.65 - stdout> 	at java.util.jar.JarFile.<init>(JarFile.java:166)
2019-07-30 09:35:03.65 - stdout> 	at java.util.jar.JarFile.<init>(JarFile.java:103)
2019-07-30 09:35:03.65 - stdout> 	at sun.net.www.protocol.jar.URLJarFile.<init>(URLJarFile.java:93)
2019-07-30 09:35:03.651 - stdout> 	at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69)
2019-07-30 09:35:03.651 - stdout> 	at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:84)
2019-07-30 09:35:03.651 - stdout> 	at sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122)
2019-07-30 09:35:03.651 - stdout> 	at sun.net.www.protocol.jar.JarURLConnection.getInputStream(JarURLConnection.java:152)
2019-07-30 09:35:03.651 - stdout> 	at java.net.URL.openStream(URL.java:1045)
2019-07-30 09:35:03.651 - stdout> 	at java.util.ServiceLoader.parse(ServiceLoader.java:304)
2019-07-30 09:35:03.651 - stdout> 	... 42 more
2019-07-30 09:35:03.651 - stdout> 
2019-07-30 09:35:03.729 - stderr> 19/07/30 09:35:03 INFO SparkContext: Invoking stop() from shutdown hook
2019-07-30 09:35:03.743 - stderr> 19/07/30 09:35:03 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
2019-07-30 09:35:03.753 - stderr> 19/07/30 09:35:03 INFO MemoryStore: MemoryStore cleared
2019-07-30 09:35:03.753 - stderr> 19/07/30 09:35:03 INFO BlockManager: BlockManager stopped
2019-07-30 09:35:03.754 - stderr> 19/07/30 09:35:03 INFO BlockManagerMaster: BlockManagerMaster stopped
2019-07-30 09:35:03.76 - stderr> 19/07/30 09:35:03 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
2019-07-30 09:35:03.764 - stderr> 19/07/30 09:35:03 INFO SparkContext: Successfully stopped SparkContext
2019-07-30 09:35:03.765 - stderr> 19/07/30 09:35:03 INFO ShutdownHookManager: Shutdown hook called
2019-07-30 09:35:03.766 - stderr> 19/07/30 09:35:03 INFO ShutdownHookManager: Deleting directory /tmp/spark-2cfda2e6-3acb-4442-b235-604af57763b3/pyspark-2f853c67-07dc-418a-966c-6be6bad3acb8
2019-07-30 09:35:03.77 - stderr> 19/07/30 09:35:03 INFO ShutdownHookManager: Deleting directory /tmp/hive-v1_2-3db9434c-f1b4-4291-ab92-4b28a909c359
2019-07-30 09:35:03.776 - stderr> 19/07/30 09:35:03 INFO ShutdownHookManager: Deleting directory /tmp/spark-2cfda2e6-3acb-4442-b235-604af57763b3
2019-07-30 09:35:03.78 - stderr> 19/07/30 09:35:03 INFO ShutdownHookManager: Deleting directory /tmp/spark-b35b6dfe-4aaa-4ed6-bccc-a4e50b2f4671
           
	at org.scalatest.Assertions.newAssertionFailedException(Assertions.scala:528)
	at org.scalatest.Assertions.newAssertionFailedException$(Assertions.scala:527)
	at org.scalatest.FunSuite.newAssertionFailedException(FunSuite.scala:1560)
	at org.scalatest.Assertions.fail(Assertions.scala:1089)
	at org.scalatest.Assertions.fail$(Assertions.scala:1085)
	at org.scalatest.FunSuite.fail(FunSuite.scala:1560)
	at org.apache.spark.sql.hive.SparkSubmitTestUtils.runSparkSubmit(SparkSubmitTestUtils.scala:94)
	at org.apache.spark.sql.hive.SparkSubmitTestUtils.runSparkSubmit$(SparkSubmitTestUtils.scala:41)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.runSparkSubmit(HiveExternalCatalogVersionsSuite.scala:47)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.$anonfun$beforeAll$1(HiveExternalCatalogVersionsSuite.scala:198)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.$anonfun$beforeAll$1$adapted(HiveExternalCatalogVersionsSuite.scala:181)
	at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
	at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
	at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.beforeAll(HiveExternalCatalogVersionsSuite.scala:181)
	at org.scalatest.BeforeAndAfterAll.liftedTree1$1(BeforeAndAfterAll.scala:212)
	at org.scalatest.BeforeAndAfterAll.run(BeforeAndAfterAll.scala:210)
	at org.scalatest.BeforeAndAfterAll.run$(BeforeAndAfterAll.scala:208)
	at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:56)
	at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:314)
	at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:507)
	at sbt.ForkMain$Run$2.call(ForkMain.java:296)
	at sbt.ForkMain$Run$2.call(ForkMain.java:286)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)