org.scalatest.exceptions.TestFailedException: spark-submit returned with exit code 1. Command line: './bin/spark-submit' '--name' 'prepare testing tables' '--master' 'local[2]' '--conf' 'spark.ui.enabled=false' '--conf' 'spark.master.rest.enabled=false' '--conf' 'spark.sql.hive.metastore.version=1.2.1' '--conf' 'spark.sql.hive.metastore.jars=maven' '--conf' 'spark.sql.warehouse.dir=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-67f09022-7724-465b-90ac-4a1a2ad4f64d' '--conf' 'spark.sql.test.version.index=0' '--driver-java-options' '-Dderby.system.home=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-67f09022-7724-465b-90ac-4a1a2ad4f64d' '/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test5652630777716270402.py' 2019-09-09 08:16:29.845 - stdout> 2019-09-09 08:16:29 WARN NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2019-09-09 08:16:30.512 - stdout> 2019-09-09 08:16:30 INFO SparkContext:54 - Running Spark version 2.3.4 2019-09-09 08:16:30.548 - stdout> 2019-09-09 08:16:30 INFO SparkContext:54 - Submitted application: prepare testing tables 2019-09-09 08:16:30.621 - stdout> 2019-09-09 08:16:30 INFO SecurityManager:54 - Changing view acls to: jenkins 2019-09-09 08:16:30.622 - stdout> 2019-09-09 08:16:30 INFO SecurityManager:54 - Changing modify acls to: jenkins 2019-09-09 08:16:30.622 - stdout> 2019-09-09 08:16:30 INFO SecurityManager:54 - Changing view acls groups to: 2019-09-09 08:16:30.622 - stdout> 2019-09-09 08:16:30 INFO SecurityManager:54 - Changing modify acls groups to: 2019-09-09 08:16:30.622 - stdout> 2019-09-09 08:16:30 INFO SecurityManager:54 - SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(jenkins); groups with view permissions: Set(); users with modify permissions: Set(jenkins); groups with modify permissions: Set() 2019-09-09 08:16:30.922 - stdout> 2019-09-09 08:16:30 INFO Utils:54 - Successfully started service 'sparkDriver' on port 37011. 2019-09-09 08:16:30.951 - stdout> 2019-09-09 08:16:30 INFO SparkEnv:54 - Registering MapOutputTracker 2019-09-09 08:16:30.979 - stdout> 2019-09-09 08:16:30 INFO SparkEnv:54 - Registering BlockManagerMaster 2019-09-09 08:16:30.984 - stdout> 2019-09-09 08:16:30 INFO BlockManagerMasterEndpoint:54 - Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information 2019-09-09 08:16:30.985 - stdout> 2019-09-09 08:16:30 INFO BlockManagerMasterEndpoint:54 - BlockManagerMasterEndpoint up 2019-09-09 08:16:30.998 - stdout> 2019-09-09 08:16:30 INFO DiskBlockManager:54 - Created local directory at /tmp/blockmgr-729e5747-230f-48dd-bef9-a8df7de10bcd 2019-09-09 08:16:31.02 - stdout> 2019-09-09 08:16:31 INFO MemoryStore:54 - MemoryStore started with capacity 366.3 MB 2019-09-09 08:16:31.036 - stdout> 2019-09-09 08:16:31 INFO SparkEnv:54 - Registering OutputCommitCoordinator 2019-09-09 08:16:31.359 - stdout> 2019-09-09 08:16:31 INFO SparkContext:54 - Added file file:/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test5652630777716270402.py at file:/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test5652630777716270402.py with timestamp 1568042191358 2019-09-09 08:16:31.361 - stdout> 2019-09-09 08:16:31 INFO Utils:54 - Copying /home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test5652630777716270402.py to /tmp/spark-eae4b1da-f853-4e88-aba1-716c25407600/userFiles-621cca74-9d46-479c-9ba3-49884137f876/test5652630777716270402.py 2019-09-09 08:16:31.447 - stdout> 2019-09-09 08:16:31 INFO Executor:54 - Starting executor ID driver on host localhost 2019-09-09 08:16:31.471 - stdout> 2019-09-09 08:16:31 INFO Utils:54 - Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 46637. 2019-09-09 08:16:31.472 - stdout> 2019-09-09 08:16:31 INFO NettyBlockTransferService:54 - Server created on amp-jenkins-worker-05.amp:46637 2019-09-09 08:16:31.473 - stdout> 2019-09-09 08:16:31 INFO BlockManager:54 - Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy 2019-09-09 08:16:31.52 - stdout> 2019-09-09 08:16:31 INFO BlockManagerMaster:54 - Registering BlockManager BlockManagerId(driver, amp-jenkins-worker-05.amp, 46637, None) 2019-09-09 08:16:31.525 - stdout> 2019-09-09 08:16:31 INFO BlockManagerMasterEndpoint:54 - Registering block manager amp-jenkins-worker-05.amp:46637 with 366.3 MB RAM, BlockManagerId(driver, amp-jenkins-worker-05.amp, 46637, None) 2019-09-09 08:16:31.528 - stdout> 2019-09-09 08:16:31 INFO BlockManagerMaster:54 - Registered BlockManager BlockManagerId(driver, amp-jenkins-worker-05.amp, 46637, None) 2019-09-09 08:16:31.549 - stdout> 2019-09-09 08:16:31 INFO BlockManager:54 - Initialized BlockManager: BlockManagerId(driver, amp-jenkins-worker-05.amp, 46637, None) 2019-09-09 08:16:31.815 - stdout> 2019-09-09 08:16:31 INFO log:192 - Logging initialized @3085ms 2019-09-09 08:16:32.032 - stdout> 2019-09-09 08:16:32 INFO SharedState:54 - Setting hive.metastore.warehouse.dir ('null') to the value of spark.sql.warehouse.dir ('/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-67f09022-7724-465b-90ac-4a1a2ad4f64d'). 2019-09-09 08:16:32.033 - stdout> 2019-09-09 08:16:32 INFO SharedState:54 - Warehouse path is '/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-67f09022-7724-465b-90ac-4a1a2ad4f64d'. 2019-09-09 08:16:32.503 - stdout> 2019-09-09 08:16:32 INFO StateStoreCoordinatorRef:54 - Registered StateStoreCoordinator endpoint 2019-09-09 08:16:32.859 - stdout> 2019-09-09 08:16:32 INFO HiveUtils:54 - Initializing HiveMetastoreConnection version 1.2.1 using maven. 2019-09-09 08:16:32.864 - stderr> http://www.datanucleus.org/downloads/maven2 added as a remote repository with the name: repo-1 2019-09-09 08:16:32.867 - stderr> Ivy Default Cache set to: /home/jenkins/.ivy2/cache 2019-09-09 08:16:32.868 - stderr> The jars for the packages stored in: /home/jenkins/.ivy2/jars 2019-09-09 08:16:32.896 - stderr> :: loading settings :: url = jar:file:/tmp/test-spark/spark-2.3.4/jars/ivy-2.4.0.jar!/org/apache/ivy/core/settings/ivysettings.xml 2019-09-09 08:16:32.949 - stderr> org.apache.hive#hive-metastore added as a dependency 2019-09-09 08:16:32.949 - stderr> org.apache.hive#hive-exec added as a dependency 2019-09-09 08:16:32.949 - stderr> org.apache.hive#hive-common added as a dependency 2019-09-09 08:16:32.949 - stderr> org.apache.hive#hive-serde added as a dependency 2019-09-09 08:16:32.949 - stderr> com.google.guava#guava added as a dependency 2019-09-09 08:16:32.949 - stderr> org.apache.hadoop#hadoop-client added as a dependency 2019-09-09 08:16:32.951 - stderr> :: resolving dependencies :: org.apache.spark#spark-submit-parent-7932b6ca-48dd-4850-83e5-bef1daf10e13;1.0 2019-09-09 08:16:32.952 - stderr> confs: [default] 2019-09-09 08:16:33.161 - stderr> found org.apache.hive#hive-metastore;1.2.2 in central 2019-09-09 08:16:33.211 - stderr> found org.apache.hive#hive-serde;1.2.2 in central 2019-09-09 08:16:33.247 - stderr> found org.apache.hive#hive-common;1.2.2 in central 2019-09-09 08:16:33.274 - stderr> found org.apache.hive#hive-shims;1.2.2 in central 2019-09-09 08:16:33.302 - stderr> found org.apache.hive.shims#hive-shims-common;1.2.2 in central 2019-09-09 08:16:33.326 - stderr> found commons-logging#commons-logging;1.1.3 in central 2019-09-09 08:16:33.35 - stderr> found log4j#log4j;1.2.16 in central 2019-09-09 08:16:33.372 - stderr> found log4j#apache-log4j-extras;1.2.17 in central 2019-09-09 08:16:33.396 - stderr> found com.google.guava#guava;14.0.1 in central 2019-09-09 08:16:33.415 - stderr> found commons-lang#commons-lang;2.6 in central 2019-09-09 08:16:33.433 - stderr> found org.apache.thrift#libthrift;0.9.2 in central 2019-09-09 08:16:33.45 - stderr> found org.slf4j#slf4j-api;1.7.5 in central 2019-09-09 08:16:33.47 - stderr> found org.apache.httpcomponents#httpclient;4.4 in central 2019-09-09 08:16:33.494 - stderr> found org.apache.httpcomponents#httpcore;4.4 in central 2019-09-09 08:16:33.514 - stderr> found commons-codec#commons-codec;1.4 in central 2019-09-09 08:16:33.535 - stderr> found org.apache.zookeeper#zookeeper;3.4.6 in central 2019-09-09 08:16:33.554 - stderr> found org.slf4j#slf4j-log4j12;1.7.5 in central 2019-09-09 08:16:33.576 - stderr> found jline#jline;2.12 in central 2019-09-09 08:16:33.594 - stderr> found io.netty#netty;3.7.0.Final in central 2019-09-09 08:16:33.619 - stderr> found org.apache.hive.shims#hive-shims-0.20S;1.2.2 in central 2019-09-09 08:16:33.658 - stderr> found org.apache.hive.shims#hive-shims-0.23;1.2.2 in central 2019-09-09 08:16:33.72 - stderr> found org.apache.hadoop#hadoop-yarn-server-resourcemanager;2.6.0 in central 2019-09-09 08:16:33.759 - stderr> found org.apache.hadoop#hadoop-annotations;2.6.0 in central 2019-09-09 08:16:33.786 - stderr> found com.google.inject.extensions#guice-servlet;3.0 in central 2019-09-09 08:16:33.807 - stderr> found com.google.inject#guice;3.0 in central 2019-09-09 08:16:33.826 - stderr> found javax.inject#javax.inject;1 in central 2019-09-09 08:16:33.845 - stderr> found aopalliance#aopalliance;1.0 in central 2019-09-09 08:16:33.864 - stderr> found org.sonatype.sisu.inject#cglib;2.2.1-v20090111 in central 2019-09-09 08:16:33.883 - stderr> found asm#asm;3.2 in central 2019-09-09 08:16:33.903 - stderr> found com.google.protobuf#protobuf-java;2.5.0 in central 2019-09-09 08:16:33.926 - stderr> found commons-io#commons-io;2.4 in central 2019-09-09 08:16:33.951 - stderr> found com.sun.jersey#jersey-json;1.14 in central 2019-09-09 08:16:33.976 - stderr> found org.codehaus.jettison#jettison;1.1 in central 2019-09-09 08:16:33.999 - stderr> found com.sun.xml.bind#jaxb-impl;2.2.3-1 in central 2019-09-09 08:16:34.019 - stderr> found javax.xml.bind#jaxb-api;2.2.2 in central 2019-09-09 08:16:34.034 - stderr> found javax.xml.stream#stax-api;1.0-2 in central 2019-09-09 08:16:34.05 - stderr> found javax.activation#activation;1.1 in central 2019-09-09 08:16:34.066 - stderr> found org.codehaus.jackson#jackson-core-asl;1.9.2 in central 2019-09-09 08:16:34.082 - stderr> found org.codehaus.jackson#jackson-mapper-asl;1.9.2 in central 2019-09-09 08:16:34.097 - stderr> found org.codehaus.jackson#jackson-jaxrs;1.9.2 in central 2019-09-09 08:16:34.115 - stderr> found org.codehaus.jackson#jackson-xc;1.9.2 in central 2019-09-09 08:16:34.134 - stderr> found com.sun.jersey.contribs#jersey-guice;1.9 in central 2019-09-09 08:16:34.165 - stderr> found org.apache.hadoop#hadoop-yarn-common;2.6.0 in central 2019-09-09 08:16:34.199 - stderr> found org.apache.hadoop#hadoop-yarn-api;2.6.0 in central 2019-09-09 08:16:34.269 - stderr> found org.apache.commons#commons-compress;1.4.1 in central 2019-09-09 08:16:34.295 - stderr> found org.tukaani#xz;1.0 in central 2019-09-09 08:16:34.348 - stderr> found org.mortbay.jetty#jetty-util;6.1.26 in central 2019-09-09 08:16:34.372 - stderr> found com.sun.jersey#jersey-core;1.14 in central 2019-09-09 08:16:34.389 - stderr> found com.sun.jersey#jersey-client;1.9 in central 2019-09-09 08:16:34.418 - stderr> found commons-cli#commons-cli;1.2 in central 2019-09-09 08:16:34.457 - stderr> found com.sun.jersey#jersey-server;1.14 in central 2019-09-09 08:16:34.508 - stderr> found org.apache.hadoop#hadoop-yarn-server-common;2.6.0 in central 2019-09-09 08:16:34.536 - stderr> found org.fusesource.leveldbjni#leveldbjni-all;1.8 in central 2019-09-09 08:16:34.564 - stderr> found org.apache.hadoop#hadoop-yarn-server-applicationhistoryservice;2.6.0 in central 2019-09-09 08:16:34.612 - stderr> found commons-collections#commons-collections;3.2.2 in central 2019-09-09 08:16:34.63 - stderr> found org.apache.hadoop#hadoop-yarn-server-web-proxy;2.6.0 in central 2019-09-09 08:16:34.646 - stderr> found commons-httpclient#commons-httpclient;3.0.1 in central 2019-09-09 08:16:34.66 - stderr> found junit#junit;4.11 in central 2019-09-09 08:16:34.675 - stderr> found org.hamcrest#hamcrest-core;1.3 in central 2019-09-09 08:16:34.691 - stderr> found org.mortbay.jetty#jetty;6.1.26 in central 2019-09-09 08:16:34.722 - stderr> found org.apache.hive.shims#hive-shims-scheduler;1.2.2 in central 2019-09-09 08:16:34.736 - stderr> found joda-time#joda-time;2.5 in central 2019-09-09 08:16:34.745 - stderr> found org.apache.ant#ant;1.9.1 in central 2019-09-09 08:16:34.756 - stderr> found org.apache.ant#ant-launcher;1.9.1 in central 2019-09-09 08:16:34.766 - stderr> found org.json#json;20090211 in central 2019-09-09 08:16:34.776 - stderr> found com.google.code.findbugs#jsr305;3.0.0 in central 2019-09-09 08:16:34.787 - stderr> found org.apache.avro#avro;1.7.5 in central 2019-09-09 08:16:34.8 - stderr> found com.thoughtworks.paranamer#paranamer;2.3 in central 2019-09-09 08:16:34.811 - stderr> found org.xerial.snappy#snappy-java;1.0.5 in central 2019-09-09 08:16:34.828 - stderr> found net.sf.opencsv#opencsv;2.3 in central 2019-09-09 08:16:34.839 - stderr> found com.twitter#parquet-hadoop-bundle;1.6.0 in central 2019-09-09 08:16:34.851 - stderr> found com.jolbox#bonecp;0.8.0.RELEASE in central 2019-09-09 08:16:34.861 - stderr> found org.apache.derby#derby;10.10.2.0 in central 2019-09-09 08:16:34.87 - stderr> found org.datanucleus#datanucleus-api-jdo;3.2.6 in central 2019-09-09 08:16:34.879 - stderr> found org.datanucleus#datanucleus-core;3.2.10 in central 2019-09-09 08:16:34.889 - stderr> found org.datanucleus#datanucleus-rdbms;3.2.9 in central 2019-09-09 08:16:34.897 - stderr> found commons-pool#commons-pool;1.5.4 in central 2019-09-09 08:16:34.905 - stderr> found commons-dbcp#commons-dbcp;1.4 in central 2019-09-09 08:16:34.914 - stderr> found javax.jdo#jdo-api;3.0.1 in central 2019-09-09 08:16:34.922 - stderr> found javax.transaction#jta;1.1 in central 2019-09-09 08:16:34.929 - stderr> found org.antlr#antlr-runtime;3.4 in central 2019-09-09 08:16:34.939 - stderr> found org.antlr#stringtemplate;3.2.1 in central 2019-09-09 08:16:34.95 - stderr> found antlr#antlr;2.7.7 in central 2019-09-09 08:16:34.961 - stderr> found org.apache.thrift#libfb303;0.9.2 in central 2019-09-09 08:16:34.976 - stderr> found org.apache.hive#hive-exec;1.2.2 in central 2019-09-09 08:16:34.991 - stderr> found org.apache.hive#hive-ant;1.2.2 in central 2019-09-09 08:16:35.01 - stderr> found org.apache.velocity#velocity;1.5 in central 2019-09-09 08:16:35.024 - stderr> found oro#oro;2.0.8 in central 2019-09-09 08:16:35.065 - stderr> found org.antlr#ST4;4.0.4 in central 2019-09-09 08:16:35.09 - stderr> found org.apache.ivy#ivy;2.4.0 in central 2019-09-09 08:16:35.109 - stderr> found org.codehaus.groovy#groovy-all;2.1.6 in central 2019-09-09 08:16:35.123 - stderr> found org.apache.calcite#calcite-core;1.2.0-incubating in central 2019-09-09 08:16:35.137 - stderr> found org.apache.calcite#calcite-avatica;1.2.0-incubating in central 2019-09-09 08:16:35.15 - stderr> found org.apache.calcite#calcite-linq4j;1.2.0-incubating in central 2019-09-09 08:16:35.171 - stderr> found net.hydromatic#eigenbase-properties;1.1.5 in central 2019-09-09 08:16:35.18 - stderr> found org.codehaus.janino#janino;2.7.6 in central 2019-09-09 08:16:35.19 - stderr> found org.codehaus.janino#commons-compiler;2.7.6 in central 2019-09-09 08:16:35.203 - stderr> found stax#stax-api;1.0.1 in central 2019-09-09 08:16:35.232 - stderr> found org.apache.hadoop#hadoop-client;2.7.3 in central 2019-09-09 08:16:35.257 - stderr> found org.apache.hadoop#hadoop-common;2.7.3 in central 2019-09-09 08:16:35.276 - stderr> found org.apache.hadoop#hadoop-annotations;2.7.3 in central 2019-09-09 08:16:35.297 - stderr> found org.apache.commons#commons-math3;3.1.1 in central 2019-09-09 08:16:35.311 - stderr> found xmlenc#xmlenc;0.52 in central 2019-09-09 08:16:35.322 - stderr> found commons-httpclient#commons-httpclient;3.1 in central 2019-09-09 08:16:35.349 - stderr> found commons-net#commons-net;3.1 in central 2019-09-09 08:16:35.377 - stderr> found log4j#log4j;1.2.17 in central 2019-09-09 08:16:35.395 - stderr> found commons-configuration#commons-configuration;1.6 in central 2019-09-09 08:16:35.412 - stderr> found commons-digester#commons-digester;1.8 in central 2019-09-09 08:16:35.428 - stderr> found commons-beanutils#commons-beanutils;1.7.0 in central 2019-09-09 08:16:35.442 - stderr> found commons-beanutils#commons-beanutils-core;1.8.0 in central 2019-09-09 08:16:35.471 - stderr> found org.slf4j#slf4j-api;1.7.10 in central 2019-09-09 08:16:35.479 - stderr> found org.codehaus.jackson#jackson-core-asl;1.9.13 in central 2019-09-09 08:16:35.489 - stderr> found org.codehaus.jackson#jackson-mapper-asl;1.9.13 in central 2019-09-09 08:16:35.511 - stderr> found com.google.code.gson#gson;2.2.4 in central 2019-09-09 08:16:35.522 - stderr> found org.apache.hadoop#hadoop-auth;2.7.3 in central 2019-09-09 08:16:35.54 - stderr> found org.apache.directory.server#apacheds-kerberos-codec;2.0.0-M15 in central 2019-09-09 08:16:35.552 - stderr> found org.apache.directory.server#apacheds-i18n;2.0.0-M15 in central 2019-09-09 08:16:35.563 - stderr> found org.apache.directory.api#api-asn1-api;1.0.0-M20 in central 2019-09-09 08:16:35.574 - stderr> found org.apache.directory.api#api-util;1.0.0-M20 in central 2019-09-09 08:16:35.59 - stderr> found org.apache.htrace#htrace-core;3.1.0-incubating in central 2019-09-09 08:16:35.611 - stderr> found javax.servlet.jsp#jsp-api;2.1 in central 2019-09-09 08:16:35.631 - stderr> found org.slf4j#slf4j-log4j12;1.7.10 in central 2019-09-09 08:16:35.658 - stderr> found org.apache.hadoop#hadoop-hdfs;2.7.3 in central 2019-09-09 08:16:35.686 - stderr> found io.netty#netty-all;4.0.23.Final in central 2019-09-09 08:16:35.693 - stderr> found xerces#xercesImpl;2.9.1 in central 2019-09-09 08:16:35.701 - stderr> found xml-apis#xml-apis;1.3.04 in central 2019-09-09 08:16:35.715 - stderr> found org.apache.hadoop#hadoop-mapreduce-client-app;2.7.3 in central 2019-09-09 08:16:35.724 - stderr> found org.apache.hadoop#hadoop-mapreduce-client-common;2.7.3 in central 2019-09-09 08:16:35.734 - stderr> found org.apache.hadoop#hadoop-yarn-common;2.7.3 in central 2019-09-09 08:16:35.749 - stderr> found org.apache.hadoop#hadoop-yarn-api;2.7.3 in central 2019-09-09 08:16:35.815 - stderr> found org.codehaus.jackson#jackson-jaxrs;1.9.13 in central 2019-09-09 08:16:35.827 - stderr> found org.codehaus.jackson#jackson-xc;1.9.13 in central 2019-09-09 08:16:35.886 - stderr> found org.apache.hadoop#hadoop-yarn-client;2.7.3 in central 2019-09-09 08:16:35.9 - stderr> found org.apache.hadoop#hadoop-mapreduce-client-core;2.7.3 in central 2019-09-09 08:16:35.915 - stderr> found org.apache.hadoop#hadoop-yarn-server-common;2.7.3 in central 2019-09-09 08:16:35.935 - stderr> found org.apache.hadoop#hadoop-mapreduce-client-shuffle;2.7.3 in central 2019-09-09 08:16:35.95 - stderr> found org.apache.hadoop#hadoop-mapreduce-client-jobclient;2.7.3 in central 2019-09-09 08:16:36.06 - stderr> :: resolution report :: resolve 3044ms :: artifacts dl 65ms 2019-09-09 08:16:36.06 - stderr> :: modules in use: 2019-09-09 08:16:36.062 - stderr> antlr#antlr;2.7.7 from central in [default] 2019-09-09 08:16:36.063 - stderr> aopalliance#aopalliance;1.0 from central in [default] 2019-09-09 08:16:36.063 - stderr> asm#asm;3.2 from central in [default] 2019-09-09 08:16:36.063 - stderr> com.google.code.findbugs#jsr305;3.0.0 from central in [default] 2019-09-09 08:16:36.063 - stderr> com.google.code.gson#gson;2.2.4 from central in [default] 2019-09-09 08:16:36.064 - stderr> com.google.guava#guava;14.0.1 from central in [default] 2019-09-09 08:16:36.064 - stderr> com.google.inject#guice;3.0 from central in [default] 2019-09-09 08:16:36.064 - stderr> com.google.inject.extensions#guice-servlet;3.0 from central in [default] 2019-09-09 08:16:36.064 - stderr> com.google.protobuf#protobuf-java;2.5.0 from central in [default] 2019-09-09 08:16:36.065 - stderr> com.jolbox#bonecp;0.8.0.RELEASE from central in [default] 2019-09-09 08:16:36.065 - stderr> com.sun.jersey#jersey-client;1.9 from central in [default] 2019-09-09 08:16:36.065 - stderr> com.sun.jersey#jersey-core;1.14 from central in [default] 2019-09-09 08:16:36.065 - stderr> com.sun.jersey#jersey-json;1.14 from central in [default] 2019-09-09 08:16:36.066 - stderr> com.sun.jersey#jersey-server;1.14 from central in [default] 2019-09-09 08:16:36.066 - stderr> com.sun.jersey.contribs#jersey-guice;1.9 from central in [default] 2019-09-09 08:16:36.066 - stderr> com.sun.xml.bind#jaxb-impl;2.2.3-1 from central in [default] 2019-09-09 08:16:36.067 - stderr> com.thoughtworks.paranamer#paranamer;2.3 from central in [default] 2019-09-09 08:16:36.067 - stderr> com.twitter#parquet-hadoop-bundle;1.6.0 from central in [default] 2019-09-09 08:16:36.067 - stderr> commons-beanutils#commons-beanutils;1.7.0 from central in [default] 2019-09-09 08:16:36.067 - stderr> commons-beanutils#commons-beanutils-core;1.8.0 from central in [default] 2019-09-09 08:16:36.068 - stderr> commons-cli#commons-cli;1.2 from central in [default] 2019-09-09 08:16:36.068 - stderr> commons-codec#commons-codec;1.4 from central in [default] 2019-09-09 08:16:36.068 - stderr> commons-collections#commons-collections;3.2.2 from central in [default] 2019-09-09 08:16:36.068 - stderr> commons-configuration#commons-configuration;1.6 from central in [default] 2019-09-09 08:16:36.069 - stderr> commons-dbcp#commons-dbcp;1.4 from central in [default] 2019-09-09 08:16:36.069 - stderr> commons-digester#commons-digester;1.8 from central in [default] 2019-09-09 08:16:36.069 - stderr> commons-httpclient#commons-httpclient;3.1 from central in [default] 2019-09-09 08:16:36.069 - stderr> commons-io#commons-io;2.4 from central in [default] 2019-09-09 08:16:36.07 - stderr> commons-lang#commons-lang;2.6 from central in [default] 2019-09-09 08:16:36.07 - stderr> commons-logging#commons-logging;1.1.3 from central in [default] 2019-09-09 08:16:36.07 - stderr> commons-net#commons-net;3.1 from central in [default] 2019-09-09 08:16:36.07 - stderr> commons-pool#commons-pool;1.5.4 from central in [default] 2019-09-09 08:16:36.071 - stderr> io.netty#netty;3.7.0.Final from central in [default] 2019-09-09 08:16:36.071 - stderr> io.netty#netty-all;4.0.23.Final from central in [default] 2019-09-09 08:16:36.071 - stderr> javax.activation#activation;1.1 from central in [default] 2019-09-09 08:16:36.071 - stderr> javax.inject#javax.inject;1 from central in [default] 2019-09-09 08:16:36.072 - stderr> javax.jdo#jdo-api;3.0.1 from central in [default] 2019-09-09 08:16:36.072 - stderr> javax.servlet.jsp#jsp-api;2.1 from central in [default] 2019-09-09 08:16:36.072 - stderr> javax.transaction#jta;1.1 from central in [default] 2019-09-09 08:16:36.072 - stderr> javax.xml.bind#jaxb-api;2.2.2 from central in [default] 2019-09-09 08:16:36.072 - stderr> javax.xml.stream#stax-api;1.0-2 from central in [default] 2019-09-09 08:16:36.073 - stderr> jline#jline;2.12 from central in [default] 2019-09-09 08:16:36.073 - stderr> joda-time#joda-time;2.5 from central in [default] 2019-09-09 08:16:36.073 - stderr> log4j#apache-log4j-extras;1.2.17 from central in [default] 2019-09-09 08:16:36.073 - stderr> log4j#log4j;1.2.17 from central in [default] 2019-09-09 08:16:36.074 - stderr> net.hydromatic#eigenbase-properties;1.1.5 from central in [default] 2019-09-09 08:16:36.074 - stderr> net.sf.opencsv#opencsv;2.3 from central in [default] 2019-09-09 08:16:36.074 - stderr> org.antlr#ST4;4.0.4 from central in [default] 2019-09-09 08:16:36.074 - stderr> org.antlr#antlr-runtime;3.4 from central in [default] 2019-09-09 08:16:36.075 - stderr> org.antlr#stringtemplate;3.2.1 from central in [default] 2019-09-09 08:16:36.075 - stderr> org.apache.ant#ant;1.9.1 from central in [default] 2019-09-09 08:16:36.075 - stderr> org.apache.ant#ant-launcher;1.9.1 from central in [default] 2019-09-09 08:16:36.075 - stderr> org.apache.avro#avro;1.7.5 from central in [default] 2019-09-09 08:16:36.076 - stderr> org.apache.calcite#calcite-avatica;1.2.0-incubating from central in [default] 2019-09-09 08:16:36.076 - stderr> org.apache.calcite#calcite-core;1.2.0-incubating from central in [default] 2019-09-09 08:16:36.076 - stderr> org.apache.calcite#calcite-linq4j;1.2.0-incubating from central in [default] 2019-09-09 08:16:36.076 - stderr> org.apache.commons#commons-compress;1.4.1 from central in [default] 2019-09-09 08:16:36.077 - stderr> org.apache.commons#commons-math3;3.1.1 from central in [default] 2019-09-09 08:16:36.077 - stderr> org.apache.derby#derby;10.10.2.0 from central in [default] 2019-09-09 08:16:36.077 - stderr> org.apache.directory.api#api-asn1-api;1.0.0-M20 from central in [default] 2019-09-09 08:16:36.077 - stderr> org.apache.directory.api#api-util;1.0.0-M20 from central in [default] 2019-09-09 08:16:36.078 - stderr> org.apache.directory.server#apacheds-i18n;2.0.0-M15 from central in [default] 2019-09-09 08:16:36.078 - stderr> org.apache.directory.server#apacheds-kerberos-codec;2.0.0-M15 from central in [default] 2019-09-09 08:16:36.078 - stderr> org.apache.hadoop#hadoop-annotations;2.7.3 from central in [default] 2019-09-09 08:16:36.078 - stderr> org.apache.hadoop#hadoop-auth;2.7.3 from central in [default] 2019-09-09 08:16:36.078 - stderr> org.apache.hadoop#hadoop-client;2.7.3 from central in [default] 2019-09-09 08:16:36.079 - stderr> org.apache.hadoop#hadoop-common;2.7.3 from central in [default] 2019-09-09 08:16:36.079 - stderr> org.apache.hadoop#hadoop-hdfs;2.7.3 from central in [default] 2019-09-09 08:16:36.079 - stderr> org.apache.hadoop#hadoop-mapreduce-client-app;2.7.3 from central in [default] 2019-09-09 08:16:36.079 - stderr> org.apache.hadoop#hadoop-mapreduce-client-common;2.7.3 from central in [default] 2019-09-09 08:16:36.08 - stderr> org.apache.hadoop#hadoop-mapreduce-client-core;2.7.3 from central in [default] 2019-09-09 08:16:36.08 - stderr> org.apache.hadoop#hadoop-mapreduce-client-jobclient;2.7.3 from central in [default] 2019-09-09 08:16:36.08 - stderr> org.apache.hadoop#hadoop-mapreduce-client-shuffle;2.7.3 from central in [default] 2019-09-09 08:16:36.08 - stderr> org.apache.hadoop#hadoop-yarn-api;2.7.3 from central in [default] 2019-09-09 08:16:36.081 - stderr> org.apache.hadoop#hadoop-yarn-client;2.7.3 from central in [default] 2019-09-09 08:16:36.081 - stderr> org.apache.hadoop#hadoop-yarn-common;2.7.3 from central in [default] 2019-09-09 08:16:36.081 - stderr> org.apache.hadoop#hadoop-yarn-server-applicationhistoryservice;2.6.0 from central in [default] 2019-09-09 08:16:36.082 - stderr> org.apache.hadoop#hadoop-yarn-server-common;2.7.3 from central in [default] 2019-09-09 08:16:36.082 - stderr> org.apache.hadoop#hadoop-yarn-server-resourcemanager;2.6.0 from central in [default] 2019-09-09 08:16:36.082 - stderr> org.apache.hadoop#hadoop-yarn-server-web-proxy;2.6.0 from central in [default] 2019-09-09 08:16:36.082 - stderr> org.apache.hive#hive-ant;1.2.2 from central in [default] 2019-09-09 08:16:36.083 - stderr> org.apache.hive#hive-common;1.2.2 from central in [default] 2019-09-09 08:16:36.083 - stderr> org.apache.hive#hive-exec;1.2.2 from central in [default] 2019-09-09 08:16:36.083 - stderr> org.apache.hive#hive-metastore;1.2.2 from central in [default] 2019-09-09 08:16:36.083 - stderr> org.apache.hive#hive-serde;1.2.2 from central in [default] 2019-09-09 08:16:36.083 - stderr> org.apache.hive#hive-shims;1.2.2 from central in [default] 2019-09-09 08:16:36.084 - stderr> org.apache.hive.shims#hive-shims-0.20S;1.2.2 from central in [default] 2019-09-09 08:16:36.084 - stderr> org.apache.hive.shims#hive-shims-0.23;1.2.2 from central in [default] 2019-09-09 08:16:36.084 - stderr> org.apache.hive.shims#hive-shims-common;1.2.2 from central in [default] 2019-09-09 08:16:36.084 - stderr> org.apache.hive.shims#hive-shims-scheduler;1.2.2 from central in [default] 2019-09-09 08:16:36.085 - stderr> org.apache.htrace#htrace-core;3.1.0-incubating from central in [default] 2019-09-09 08:16:36.085 - stderr> org.apache.httpcomponents#httpclient;4.4 from central in [default] 2019-09-09 08:16:36.085 - stderr> org.apache.httpcomponents#httpcore;4.4 from central in [default] 2019-09-09 08:16:36.085 - stderr> org.apache.ivy#ivy;2.4.0 from central in [default] 2019-09-09 08:16:36.086 - stderr> org.apache.thrift#libfb303;0.9.2 from central in [default] 2019-09-09 08:16:36.086 - stderr> org.apache.thrift#libthrift;0.9.2 from central in [default] 2019-09-09 08:16:36.086 - stderr> org.apache.velocity#velocity;1.5 from central in [default] 2019-09-09 08:16:36.086 - stderr> org.apache.zookeeper#zookeeper;3.4.6 from central in [default] 2019-09-09 08:16:36.087 - stderr> org.codehaus.groovy#groovy-all;2.1.6 from central in [default] 2019-09-09 08:16:36.087 - stderr> org.codehaus.jackson#jackson-core-asl;1.9.13 from central in [default] 2019-09-09 08:16:36.087 - stderr> org.codehaus.jackson#jackson-jaxrs;1.9.13 from central in [default] 2019-09-09 08:16:36.087 - stderr> org.codehaus.jackson#jackson-mapper-asl;1.9.13 from central in [default] 2019-09-09 08:16:36.088 - stderr> org.codehaus.jackson#jackson-xc;1.9.13 from central in [default] 2019-09-09 08:16:36.088 - stderr> org.codehaus.janino#commons-compiler;2.7.6 from central in [default] 2019-09-09 08:16:36.088 - stderr> org.codehaus.janino#janino;2.7.6 from central in [default] 2019-09-09 08:16:36.088 - stderr> org.codehaus.jettison#jettison;1.1 from central in [default] 2019-09-09 08:16:36.089 - stderr> org.datanucleus#datanucleus-api-jdo;3.2.6 from central in [default] 2019-09-09 08:16:36.089 - stderr> org.datanucleus#datanucleus-core;3.2.10 from central in [default] 2019-09-09 08:16:36.089 - stderr> org.datanucleus#datanucleus-rdbms;3.2.9 from central in [default] 2019-09-09 08:16:36.089 - stderr> org.fusesource.leveldbjni#leveldbjni-all;1.8 from central in [default] 2019-09-09 08:16:36.09 - stderr> org.json#json;20090211 from central in [default] 2019-09-09 08:16:36.09 - stderr> org.mortbay.jetty#jetty;6.1.26 from central in [default] 2019-09-09 08:16:36.09 - stderr> org.mortbay.jetty#jetty-util;6.1.26 from central in [default] 2019-09-09 08:16:36.09 - stderr> org.slf4j#slf4j-api;1.7.10 from central in [default] 2019-09-09 08:16:36.091 - stderr> org.slf4j#slf4j-log4j12;1.7.10 from central in [default] 2019-09-09 08:16:36.091 - stderr> org.sonatype.sisu.inject#cglib;2.2.1-v20090111 from central in [default] 2019-09-09 08:16:36.091 - stderr> org.tukaani#xz;1.0 from central in [default] 2019-09-09 08:16:36.091 - stderr> org.xerial.snappy#snappy-java;1.0.5 from central in [default] 2019-09-09 08:16:36.091 - stderr> oro#oro;2.0.8 from central in [default] 2019-09-09 08:16:36.092 - stderr> stax#stax-api;1.0.1 from central in [default] 2019-09-09 08:16:36.092 - stderr> xerces#xercesImpl;2.9.1 from central in [default] 2019-09-09 08:16:36.092 - stderr> xml-apis#xml-apis;1.3.04 from central in [default] 2019-09-09 08:16:36.092 - stderr> xmlenc#xmlenc;0.52 from central in [default] 2019-09-09 08:16:36.092 - stderr> :: evicted modules: 2019-09-09 08:16:36.093 - stderr> log4j#log4j;1.2.16 by [log4j#log4j;1.2.17] in [default] 2019-09-09 08:16:36.093 - stderr> org.slf4j#slf4j-api;1.7.5 by [org.slf4j#slf4j-api;1.7.10] in [default] 2019-09-09 08:16:36.093 - stderr> org.slf4j#slf4j-log4j12;1.7.5 by [org.slf4j#slf4j-log4j12;1.7.10] in [default] 2019-09-09 08:16:36.093 - stderr> org.apache.hadoop#hadoop-annotations;2.6.0 by [org.apache.hadoop#hadoop-annotations;2.7.3] in [default] 2019-09-09 08:16:36.093 - stderr> org.codehaus.jackson#jackson-core-asl;1.9.2 by [org.codehaus.jackson#jackson-core-asl;1.9.13] in [default] 2019-09-09 08:16:36.093 - stderr> org.codehaus.jackson#jackson-mapper-asl;1.9.2 by [org.codehaus.jackson#jackson-mapper-asl;1.9.13] in [default] 2019-09-09 08:16:36.094 - stderr> org.codehaus.jackson#jackson-jaxrs;1.9.2 by [org.codehaus.jackson#jackson-jaxrs;1.9.13] in [default] 2019-09-09 08:16:36.094 - stderr> org.codehaus.jackson#jackson-xc;1.9.2 by [org.codehaus.jackson#jackson-xc;1.9.13] in [default] 2019-09-09 08:16:36.094 - stderr> org.apache.hadoop#hadoop-yarn-common;2.6.0 by [org.apache.hadoop#hadoop-yarn-common;2.7.3] in [default] 2019-09-09 08:16:36.094 - stderr> org.apache.hadoop#hadoop-yarn-api;2.6.0 by [org.apache.hadoop#hadoop-yarn-api;2.7.3] in [default] 2019-09-09 08:16:36.094 - stderr> org.apache.hadoop#hadoop-yarn-server-common;2.6.0 by [org.apache.hadoop#hadoop-yarn-server-common;2.7.3] in [default] 2019-09-09 08:16:36.094 - stderr> commons-httpclient#commons-httpclient;3.0.1 by [commons-httpclient#commons-httpclient;3.1] in [default] 2019-09-09 08:16:36.094 - stderr> junit#junit;4.11 transitively in [default] 2019-09-09 08:16:36.094 - stderr> org.hamcrest#hamcrest-core;1.3 transitively in [default] 2019-09-09 08:16:36.094 - stderr> com.google.code.findbugs#jsr305;1.3.9 by [com.google.code.findbugs#jsr305;3.0.0] in [default] 2019-09-09 08:16:36.094 - stderr> com.google.guava#guava;11.0.2 by [com.google.guava#guava;14.0.1] in [default] 2019-09-09 08:16:36.094 - stderr> org.apache.avro#avro;1.7.4 by [org.apache.avro#avro;1.7.5] in [default] 2019-09-09 08:16:36.095 - stderr> org.apache.httpcomponents#httpclient;4.2.5 by [org.apache.httpcomponents#httpclient;4.4] in [default] 2019-09-09 08:16:36.095 - stderr> io.netty#netty;3.6.2.Final by [io.netty#netty;3.7.0.Final] in [default] 2019-09-09 08:16:36.095 - stderr> com.sun.jersey#jersey-core;1.9 by [com.sun.jersey#jersey-core;1.14] in [default] 2019-09-09 08:16:36.095 - stderr> com.sun.jersey#jersey-server;1.9 by [com.sun.jersey#jersey-server;1.14] in [default] 2019-09-09 08:16:36.095 - stderr> com.sun.jersey#jersey-json;1.9 by [com.sun.jersey#jersey-json;1.14] in [default] 2019-09-09 08:16:36.095 - stderr> --------------------------------------------------------------------- 2019-09-09 08:16:36.095 - stderr> | | modules || artifacts | 2019-09-09 08:16:36.096 - stderr> | conf | number| search|dwnlded|evicted|| number|dwnlded| 2019-09-09 08:16:36.096 - stderr> --------------------------------------------------------------------- 2019-09-09 08:16:36.097 - stderr> | default | 145 | 0 | 0 | 22 || 123 | 0 | 2019-09-09 08:16:36.097 - stderr> --------------------------------------------------------------------- 2019-09-09 08:16:36.141 - stderr> :: retrieving :: org.apache.spark#spark-submit-parent-7932b6ca-48dd-4850-83e5-bef1daf10e13 2019-09-09 08:16:36.142 - stderr> confs: [default] 2019-09-09 08:16:36.197 - stderr> 0 artifacts copied, 123 already retrieved (0kB/55ms) 2019-09-09 08:16:36.385 - stdout> 2019-09-09 08:16:36 INFO IsolatedClientLoader:54 - Downloaded metastore jars to /tmp/hive-v1_2-3f0c8c30-eb51-4b33-8fc9-21b57b3cb06c 2019-09-09 08:16:36.959 - stdout> 2019-09-09 08:16:36 INFO HiveMetaStore:589 - 0: Opening raw store with implemenation class:org.apache.hadoop.hive.metastore.ObjectStore 2019-09-09 08:16:36.985 - stdout> 2019-09-09 08:16:36 INFO ObjectStore:289 - ObjectStore, initialize called 2019-09-09 08:16:37.108 - stdout> 2019-09-09 08:16:37 INFO Persistence:77 - Property hive.metastore.integral.jdo.pushdown unknown - will be ignored 2019-09-09 08:16:37.108 - stdout> 2019-09-09 08:16:37 INFO Persistence:77 - Property datanucleus.cache.level2 unknown - will be ignored 2019-09-09 08:16:46.805 - stdout> 2019-09-09 08:16:46 INFO ObjectStore:370 - Setting MetaStore object pin classes with hive.metastore.cache.pinobjtypes="Table,StorageDescriptor,SerDeInfo,Partition,Database,Type,FieldSchema,Order" 2019-09-09 08:16:48.245 - stdout> 2019-09-09 08:16:48 INFO Datastore:77 - The class "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as "embedded-only" so does not have its own datastore table. 2019-09-09 08:16:48.246 - stdout> 2019-09-09 08:16:48 INFO Datastore:77 - The class "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as "embedded-only" so does not have its own datastore table. 2019-09-09 08:16:56.252 - stdout> 2019-09-09 08:16:56 INFO Datastore:77 - The class "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as "embedded-only" so does not have its own datastore table. 2019-09-09 08:16:56.253 - stdout> 2019-09-09 08:16:56 INFO Datastore:77 - The class "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as "embedded-only" so does not have its own datastore table. 2019-09-09 08:17:02.114 - stdout> 2019-09-09 08:17:02 INFO MetaStoreDirectSql:139 - Using direct SQL, underlying DB is DERBY 2019-09-09 08:17:02.116 - stdout> 2019-09-09 08:17:02 INFO ObjectStore:272 - Initialized ObjectStore 2019-09-09 08:17:02.387 - stdout> 2019-09-09 08:17:02 WARN ObjectStore:6666 - Version information not found in metastore. hive.metastore.schema.verification is not enabled so recording the schema version 1.2.0 2019-09-09 08:17:02.702 - stdout> 2019-09-09 08:17:02 WARN ObjectStore:568 - Failed to get database default, returning NoSuchObjectException 2019-09-09 08:17:03.121 - stdout> 2019-09-09 08:17:03 INFO HiveMetaStore:663 - Added admin role in metastore 2019-09-09 08:17:03.138 - stdout> 2019-09-09 08:17:03 INFO HiveMetaStore:672 - Added public role in metastore 2019-09-09 08:17:03.596 - stdout> 2019-09-09 08:17:03 INFO HiveMetaStore:712 - No user is added in admin role, since config is empty 2019-09-09 08:17:03.686 - stdout> 2019-09-09 08:17:03 INFO HiveMetaStore:746 - 0: get_all_databases 2019-09-09 08:17:03.687 - stdout> 2019-09-09 08:17:03 INFO audit:371 - ugi=jenkins ip=unknown-ip-addr cmd=get_all_databases 2019-09-09 08:17:03.702 - stdout> 2019-09-09 08:17:03 INFO HiveMetaStore:746 - 0: get_functions: db=default pat=* 2019-09-09 08:17:03.702 - stdout> 2019-09-09 08:17:03 INFO audit:371 - ugi=jenkins ip=unknown-ip-addr cmd=get_functions: db=default pat=* 2019-09-09 08:17:03.704 - stdout> 2019-09-09 08:17:03 INFO Datastore:77 - The class "org.apache.hadoop.hive.metastore.model.MResourceUri" is tagged as "embedded-only" so does not have its own datastore table. 2019-09-09 08:17:05.211 - stdout> 2019-09-09 08:17:05 INFO SessionState:641 - Created local directory: /tmp/0d0c2532-9023-4376-9990-d239a122eb81_resources 2019-09-09 08:17:05.214 - stdout> 2019-09-09 08:17:05 INFO SessionState:641 - Created HDFS directory: /tmp/hive/jenkins/0d0c2532-9023-4376-9990-d239a122eb81 2019-09-09 08:17:05.218 - stdout> 2019-09-09 08:17:05 INFO SessionState:641 - Created local directory: /tmp/jenkins/0d0c2532-9023-4376-9990-d239a122eb81 2019-09-09 08:17:05.222 - stdout> 2019-09-09 08:17:05 INFO SessionState:641 - Created HDFS directory: /tmp/hive/jenkins/0d0c2532-9023-4376-9990-d239a122eb81/_tmp_space.db 2019-09-09 08:17:05.225 - stdout> 2019-09-09 08:17:05 INFO HiveClientImpl:54 - Warehouse location for Hive client (version 1.2.2) is /home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-67f09022-7724-465b-90ac-4a1a2ad4f64d 2019-09-09 08:17:05.235 - stdout> 2019-09-09 08:17:05 INFO HiveMetaStore:746 - 0: get_database: default 2019-09-09 08:17:05.235 - stdout> 2019-09-09 08:17:05 INFO audit:371 - ugi=jenkins ip=unknown-ip-addr cmd=get_database: default 2019-09-09 08:17:05.256 - stdout> 2019-09-09 08:17:05 INFO HiveMetaStore:746 - 0: get_database: global_temp 2019-09-09 08:17:05.256 - stdout> 2019-09-09 08:17:05 INFO audit:371 - ugi=jenkins ip=unknown-ip-addr cmd=get_database: global_temp 2019-09-09 08:17:05.257 - stdout> 2019-09-09 08:17:05 WARN ObjectStore:568 - Failed to get database global_temp, returning NoSuchObjectException 2019-09-09 08:17:06.749 - stdout> Traceback (most recent call last): 2019-09-09 08:17:06.749 - stdout> File "/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test5652630777716270402.py", line 8, in <module> 2019-09-09 08:17:06.749 - stdout> spark.sql("create table data_source_tbl_{} using json as select 1 i".format(version_index)) 2019-09-09 08:17:06.749 - stdout> File "/tmp/test-spark/spark-2.3.4/python/lib/pyspark.zip/pyspark/sql/session.py", line 710, in sql 2019-09-09 08:17:06.75 - stdout> File "/tmp/test-spark/spark-2.3.4/python/lib/py4j-0.10.7-src.zip/py4j/java_gateway.py", line 1257, in __call__ 2019-09-09 08:17:06.75 - stdout> File "/tmp/test-spark/spark-2.3.4/python/lib/pyspark.zip/pyspark/sql/utils.py", line 63, in deco 2019-09-09 08:17:06.75 - stdout> File "/tmp/test-spark/spark-2.3.4/python/lib/py4j-0.10.7-src.zip/py4j/protocol.py", line 328, in get_return_value 2019-09-09 08:17:06.751 - stdout> py4j.protocol.Py4JJavaError: An error occurred while calling o29.sql. 2019-09-09 08:17:06.751 - stdout> : java.lang.ExceptionInInitializerError 2019-09-09 08:17:06.751 - stdout> at org.apache.spark.SparkContext.withScope(SparkContext.scala:693) 2019-09-09 08:17:06.751 - stdout> at org.apache.spark.SparkContext.parallelize(SparkContext.scala:710) 2019-09-09 08:17:06.751 - stdout> at org.apache.spark.sql.execution.SparkStrategies.singleRowRdd$lzycompute(SparkStrategies.scala:399) 2019-09-09 08:17:06.751 - stdout> at org.apache.spark.sql.execution.SparkStrategies.singleRowRdd(SparkStrategies.scala:399) 2019-09-09 08:17:06.752 - stdout> at org.apache.spark.sql.execution.SparkStrategies$BasicOperators$.apply(SparkStrategies.scala:536) 2019-09-09 08:17:06.752 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:63) 2019-09-09 08:17:06.752 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:63) 2019-09-09 08:17:06.752 - stdout> at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434) 2019-09-09 08:17:06.752 - stdout> at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440) 2019-09-09 08:17:06.752 - stdout> at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439) 2019-09-09 08:17:06.752 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:93) 2019-09-09 08:17:06.752 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:78) 2019-09-09 08:17:06.752 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:75) 2019-09-09 08:17:06.752 - stdout> at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157) 2019-09-09 08:17:06.752 - stdout> at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157) 2019-09-09 08:17:06.752 - stdout> at scala.collection.Iterator$class.foreach(Iterator.scala:893) 2019-09-09 08:17:06.752 - stdout> at scala.collection.AbstractIterator.foreach(Iterator.scala:1336) 2019-09-09 08:17:06.752 - stdout> at scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157) 2019-09-09 08:17:06.752 - stdout> at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336) 2019-09-09 08:17:06.752 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:75) 2019-09-09 08:17:06.752 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:67) 2019-09-09 08:17:06.752 - stdout> at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434) 2019-09-09 08:17:06.752 - stdout> at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440) 2019-09-09 08:17:06.752 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:93) 2019-09-09 08:17:06.752 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:78) 2019-09-09 08:17:06.752 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:75) 2019-09-09 08:17:06.752 - stdout> at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157) 2019-09-09 08:17:06.752 - stdout> at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157) 2019-09-09 08:17:06.752 - stdout> at scala.collection.Iterator$class.foreach(Iterator.scala:893) 2019-09-09 08:17:06.752 - stdout> at scala.collection.AbstractIterator.foreach(Iterator.scala:1336) 2019-09-09 08:17:06.752 - stdout> at scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157) 2019-09-09 08:17:06.753 - stdout> at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:75) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:67) 2019-09-09 08:17:06.753 - stdout> at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434) 2019-09-09 08:17:06.753 - stdout> at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:93) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:78) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:75) 2019-09-09 08:17:06.753 - stdout> at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157) 2019-09-09 08:17:06.753 - stdout> at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157) 2019-09-09 08:17:06.753 - stdout> at scala.collection.Iterator$class.foreach(Iterator.scala:893) 2019-09-09 08:17:06.753 - stdout> at scala.collection.AbstractIterator.foreach(Iterator.scala:1336) 2019-09-09 08:17:06.753 - stdout> at scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157) 2019-09-09 08:17:06.753 - stdout> at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:75) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:67) 2019-09-09 08:17:06.753 - stdout> at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434) 2019-09-09 08:17:06.753 - stdout> at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:93) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.execution.QueryExecution.sparkPlan$lzycompute(QueryExecution.scala:72) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.execution.QueryExecution.sparkPlan(QueryExecution.scala:68) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.execution.QueryExecution.executedPlan$lzycompute(QueryExecution.scala:77) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.execution.QueryExecution.executedPlan(QueryExecution.scala:77) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3260) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.Dataset.<init>(Dataset.scala:190) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:75) 2019-09-09 08:17:06.753 - stdout> at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642) 2019-09-09 08:17:06.753 - stdout> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 2019-09-09 08:17:06.754 - stdout> at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 2019-09-09 08:17:06.754 - stdout> at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) 2019-09-09 08:17:06.754 - stdout> at java.lang.reflect.Method.invoke(Method.java:498) 2019-09-09 08:17:06.754 - stdout> at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) 2019-09-09 08:17:06.754 - stdout> at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) 2019-09-09 08:17:06.754 - stdout> at py4j.Gateway.invoke(Gateway.java:282) 2019-09-09 08:17:06.754 - stdout> at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) 2019-09-09 08:17:06.754 - stdout> at py4j.commands.CallCommand.execute(CallCommand.java:79) 2019-09-09 08:17:06.754 - stdout> at py4j.GatewayConnection.run(GatewayConnection.java:238) 2019-09-09 08:17:06.754 - stdout> at java.lang.Thread.run(Thread.java:748) 2019-09-09 08:17:06.754 - stdout> Caused by: java.util.NoSuchElementException: key not found: groupId 2019-09-09 08:17:06.754 - stdout> at scala.collection.MapLike$class.default(MapLike.scala:228) 2019-09-09 08:17:06.754 - stdout> at scala.collection.AbstractMap.default(Map.scala:59) 2019-09-09 08:17:06.754 - stdout> at scala.collection.MapLike$class.apply(MapLike.scala:141) 2019-09-09 08:17:06.754 - stdout> at scala.collection.AbstractMap.apply(Map.scala:59) 2019-09-09 08:17:06.754 - stdout> at com.fasterxml.jackson.module.scala.JacksonModule$.version$lzycompute(JacksonModule.scala:27) 2019-09-09 08:17:06.754 - stdout> at com.fasterxml.jackson.module.scala.JacksonModule$.version(JacksonModule.scala:26) 2019-09-09 08:17:06.754 - stdout> at com.fasterxml.jackson.module.scala.JacksonModule$class.version(JacksonModule.scala:49) 2019-09-09 08:17:06.754 - stdout> at com.fasterxml.jackson.module.scala.DefaultScalaModule.version(DefaultScalaModule.scala:19) 2019-09-09 08:17:06.754 - stdout> at com.fasterxml.jackson.databind.ObjectMapper.registerModule(ObjectMapper.java:710) 2019-09-09 08:17:06.754 - stdout> at org.apache.spark.rdd.RDDOperationScope$.<init>(RDDOperationScope.scala:82) 2019-09-09 08:17:06.754 - stdout> at org.apache.spark.rdd.RDDOperationScope$.<clinit>(RDDOperationScope.scala) 2019-09-09 08:17:06.754 - stdout> ... 69 more 2019-09-09 08:17:06.754 - stdout> 2019-09-09 08:17:06.821 - stdout> 2019-09-09 08:17:06 INFO SparkContext:54 - Invoking stop() from shutdown hook 2019-09-09 08:17:06.835 - stdout> 2019-09-09 08:17:06 INFO MapOutputTrackerMasterEndpoint:54 - MapOutputTrackerMasterEndpoint stopped! 2019-09-09 08:17:06.846 - stdout> 2019-09-09 08:17:06 INFO MemoryStore:54 - MemoryStore cleared 2019-09-09 08:17:06.846 - stdout> 2019-09-09 08:17:06 INFO BlockManager:54 - BlockManager stopped 2019-09-09 08:17:06.848 - stdout> 2019-09-09 08:17:06 INFO BlockManagerMaster:54 - BlockManagerMaster stopped 2019-09-09 08:17:06.853 - stdout> 2019-09-09 08:17:06 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint:54 - OutputCommitCoordinator stopped! 2019-09-09 08:17:06.856 - stdout> 2019-09-09 08:17:06 INFO SparkContext:54 - Successfully stopped SparkContext 2019-09-09 08:17:06.857 - stdout> 2019-09-09 08:17:06 INFO ShutdownHookManager:54 - Shutdown hook called 2019-09-09 08:17:06.858 - stdout> 2019-09-09 08:17:06 INFO ShutdownHookManager:54 - Deleting directory /tmp/spark-eae4b1da-f853-4e88-aba1-716c25407600/pyspark-d447c291-a72d-4e8d-817a-0f78591683cc 2019-09-09 08:17:06.859 - stdout> 2019-09-09 08:17:06 INFO ShutdownHookManager:54 - Deleting directory /tmp/hive-v1_2-3f0c8c30-eb51-4b33-8fc9-21b57b3cb06c 2019-09-09 08:17:06.861 - stdout> 2019-09-09 08:17:06 INFO ShutdownHookManager:54 - Deleting directory /tmp/spark-c02044a6-1cf1-42c1-a617-3e5739b0362e 2019-09-09 08:17:06.862 - stdout> 2019-09-09 08:17:06 INFO ShutdownHookManager:54 - Deleting directory /tmp/spark-eae4b1da-f853-4e88-aba1-716c25407600

sbt.ForkMain$ForkError: org.scalatest.exceptions.TestFailedException: spark-submit returned with exit code 1.
Command line: './bin/spark-submit' '--name' 'prepare testing tables' '--master' 'local[2]' '--conf' 'spark.ui.enabled=false' '--conf' 'spark.master.rest.enabled=false' '--conf' 'spark.sql.hive.metastore.version=1.2.1' '--conf' 'spark.sql.hive.metastore.jars=maven' '--conf' 'spark.sql.warehouse.dir=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-67f09022-7724-465b-90ac-4a1a2ad4f64d' '--conf' 'spark.sql.test.version.index=0' '--driver-java-options' '-Dderby.system.home=/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-67f09022-7724-465b-90ac-4a1a2ad4f64d' '/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test5652630777716270402.py'

2019-09-09 08:16:29.845 - stdout> 2019-09-09 08:16:29 WARN  NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
2019-09-09 08:16:30.512 - stdout> 2019-09-09 08:16:30 INFO  SparkContext:54 - Running Spark version 2.3.4
2019-09-09 08:16:30.548 - stdout> 2019-09-09 08:16:30 INFO  SparkContext:54 - Submitted application: prepare testing tables
2019-09-09 08:16:30.621 - stdout> 2019-09-09 08:16:30 INFO  SecurityManager:54 - Changing view acls to: jenkins
2019-09-09 08:16:30.622 - stdout> 2019-09-09 08:16:30 INFO  SecurityManager:54 - Changing modify acls to: jenkins
2019-09-09 08:16:30.622 - stdout> 2019-09-09 08:16:30 INFO  SecurityManager:54 - Changing view acls groups to: 
2019-09-09 08:16:30.622 - stdout> 2019-09-09 08:16:30 INFO  SecurityManager:54 - Changing modify acls groups to: 
2019-09-09 08:16:30.622 - stdout> 2019-09-09 08:16:30 INFO  SecurityManager:54 - SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(jenkins); groups with view permissions: Set(); users  with modify permissions: Set(jenkins); groups with modify permissions: Set()
2019-09-09 08:16:30.922 - stdout> 2019-09-09 08:16:30 INFO  Utils:54 - Successfully started service 'sparkDriver' on port 37011.
2019-09-09 08:16:30.951 - stdout> 2019-09-09 08:16:30 INFO  SparkEnv:54 - Registering MapOutputTracker
2019-09-09 08:16:30.979 - stdout> 2019-09-09 08:16:30 INFO  SparkEnv:54 - Registering BlockManagerMaster
2019-09-09 08:16:30.984 - stdout> 2019-09-09 08:16:30 INFO  BlockManagerMasterEndpoint:54 - Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
2019-09-09 08:16:30.985 - stdout> 2019-09-09 08:16:30 INFO  BlockManagerMasterEndpoint:54 - BlockManagerMasterEndpoint up
2019-09-09 08:16:30.998 - stdout> 2019-09-09 08:16:30 INFO  DiskBlockManager:54 - Created local directory at /tmp/blockmgr-729e5747-230f-48dd-bef9-a8df7de10bcd
2019-09-09 08:16:31.02 - stdout> 2019-09-09 08:16:31 INFO  MemoryStore:54 - MemoryStore started with capacity 366.3 MB
2019-09-09 08:16:31.036 - stdout> 2019-09-09 08:16:31 INFO  SparkEnv:54 - Registering OutputCommitCoordinator
2019-09-09 08:16:31.359 - stdout> 2019-09-09 08:16:31 INFO  SparkContext:54 - Added file file:/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test5652630777716270402.py at file:/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test5652630777716270402.py with timestamp 1568042191358
2019-09-09 08:16:31.361 - stdout> 2019-09-09 08:16:31 INFO  Utils:54 - Copying /home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test5652630777716270402.py to /tmp/spark-eae4b1da-f853-4e88-aba1-716c25407600/userFiles-621cca74-9d46-479c-9ba3-49884137f876/test5652630777716270402.py
2019-09-09 08:16:31.447 - stdout> 2019-09-09 08:16:31 INFO  Executor:54 - Starting executor ID driver on host localhost
2019-09-09 08:16:31.471 - stdout> 2019-09-09 08:16:31 INFO  Utils:54 - Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 46637.
2019-09-09 08:16:31.472 - stdout> 2019-09-09 08:16:31 INFO  NettyBlockTransferService:54 - Server created on amp-jenkins-worker-05.amp:46637
2019-09-09 08:16:31.473 - stdout> 2019-09-09 08:16:31 INFO  BlockManager:54 - Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
2019-09-09 08:16:31.52 - stdout> 2019-09-09 08:16:31 INFO  BlockManagerMaster:54 - Registering BlockManager BlockManagerId(driver, amp-jenkins-worker-05.amp, 46637, None)
2019-09-09 08:16:31.525 - stdout> 2019-09-09 08:16:31 INFO  BlockManagerMasterEndpoint:54 - Registering block manager amp-jenkins-worker-05.amp:46637 with 366.3 MB RAM, BlockManagerId(driver, amp-jenkins-worker-05.amp, 46637, None)
2019-09-09 08:16:31.528 - stdout> 2019-09-09 08:16:31 INFO  BlockManagerMaster:54 - Registered BlockManager BlockManagerId(driver, amp-jenkins-worker-05.amp, 46637, None)
2019-09-09 08:16:31.549 - stdout> 2019-09-09 08:16:31 INFO  BlockManager:54 - Initialized BlockManager: BlockManagerId(driver, amp-jenkins-worker-05.amp, 46637, None)
2019-09-09 08:16:31.815 - stdout> 2019-09-09 08:16:31 INFO  log:192 - Logging initialized @3085ms
2019-09-09 08:16:32.032 - stdout> 2019-09-09 08:16:32 INFO  SharedState:54 - Setting hive.metastore.warehouse.dir ('null') to the value of spark.sql.warehouse.dir ('/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-67f09022-7724-465b-90ac-4a1a2ad4f64d').
2019-09-09 08:16:32.033 - stdout> 2019-09-09 08:16:32 INFO  SharedState:54 - Warehouse path is '/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-67f09022-7724-465b-90ac-4a1a2ad4f64d'.
2019-09-09 08:16:32.503 - stdout> 2019-09-09 08:16:32 INFO  StateStoreCoordinatorRef:54 - Registered StateStoreCoordinator endpoint
2019-09-09 08:16:32.859 - stdout> 2019-09-09 08:16:32 INFO  HiveUtils:54 - Initializing HiveMetastoreConnection version 1.2.1 using maven.
2019-09-09 08:16:32.864 - stderr> http://www.datanucleus.org/downloads/maven2 added as a remote repository with the name: repo-1
2019-09-09 08:16:32.867 - stderr> Ivy Default Cache set to: /home/jenkins/.ivy2/cache
2019-09-09 08:16:32.868 - stderr> The jars for the packages stored in: /home/jenkins/.ivy2/jars
2019-09-09 08:16:32.896 - stderr> :: loading settings :: url = jar:file:/tmp/test-spark/spark-2.3.4/jars/ivy-2.4.0.jar!/org/apache/ivy/core/settings/ivysettings.xml
2019-09-09 08:16:32.949 - stderr> org.apache.hive#hive-metastore added as a dependency
2019-09-09 08:16:32.949 - stderr> org.apache.hive#hive-exec added as a dependency
2019-09-09 08:16:32.949 - stderr> org.apache.hive#hive-common added as a dependency
2019-09-09 08:16:32.949 - stderr> org.apache.hive#hive-serde added as a dependency
2019-09-09 08:16:32.949 - stderr> com.google.guava#guava added as a dependency
2019-09-09 08:16:32.949 - stderr> org.apache.hadoop#hadoop-client added as a dependency
2019-09-09 08:16:32.951 - stderr> :: resolving dependencies :: org.apache.spark#spark-submit-parent-7932b6ca-48dd-4850-83e5-bef1daf10e13;1.0
2019-09-09 08:16:32.952 - stderr> 	confs: [default]
2019-09-09 08:16:33.161 - stderr> 	found org.apache.hive#hive-metastore;1.2.2 in central
2019-09-09 08:16:33.211 - stderr> 	found org.apache.hive#hive-serde;1.2.2 in central
2019-09-09 08:16:33.247 - stderr> 	found org.apache.hive#hive-common;1.2.2 in central
2019-09-09 08:16:33.274 - stderr> 	found org.apache.hive#hive-shims;1.2.2 in central
2019-09-09 08:16:33.302 - stderr> 	found org.apache.hive.shims#hive-shims-common;1.2.2 in central
2019-09-09 08:16:33.326 - stderr> 	found commons-logging#commons-logging;1.1.3 in central
2019-09-09 08:16:33.35 - stderr> 	found log4j#log4j;1.2.16 in central
2019-09-09 08:16:33.372 - stderr> 	found log4j#apache-log4j-extras;1.2.17 in central
2019-09-09 08:16:33.396 - stderr> 	found com.google.guava#guava;14.0.1 in central
2019-09-09 08:16:33.415 - stderr> 	found commons-lang#commons-lang;2.6 in central
2019-09-09 08:16:33.433 - stderr> 	found org.apache.thrift#libthrift;0.9.2 in central
2019-09-09 08:16:33.45 - stderr> 	found org.slf4j#slf4j-api;1.7.5 in central
2019-09-09 08:16:33.47 - stderr> 	found org.apache.httpcomponents#httpclient;4.4 in central
2019-09-09 08:16:33.494 - stderr> 	found org.apache.httpcomponents#httpcore;4.4 in central
2019-09-09 08:16:33.514 - stderr> 	found commons-codec#commons-codec;1.4 in central
2019-09-09 08:16:33.535 - stderr> 	found org.apache.zookeeper#zookeeper;3.4.6 in central
2019-09-09 08:16:33.554 - stderr> 	found org.slf4j#slf4j-log4j12;1.7.5 in central
2019-09-09 08:16:33.576 - stderr> 	found jline#jline;2.12 in central
2019-09-09 08:16:33.594 - stderr> 	found io.netty#netty;3.7.0.Final in central
2019-09-09 08:16:33.619 - stderr> 	found org.apache.hive.shims#hive-shims-0.20S;1.2.2 in central
2019-09-09 08:16:33.658 - stderr> 	found org.apache.hive.shims#hive-shims-0.23;1.2.2 in central
2019-09-09 08:16:33.72 - stderr> 	found org.apache.hadoop#hadoop-yarn-server-resourcemanager;2.6.0 in central
2019-09-09 08:16:33.759 - stderr> 	found org.apache.hadoop#hadoop-annotations;2.6.0 in central
2019-09-09 08:16:33.786 - stderr> 	found com.google.inject.extensions#guice-servlet;3.0 in central
2019-09-09 08:16:33.807 - stderr> 	found com.google.inject#guice;3.0 in central
2019-09-09 08:16:33.826 - stderr> 	found javax.inject#javax.inject;1 in central
2019-09-09 08:16:33.845 - stderr> 	found aopalliance#aopalliance;1.0 in central
2019-09-09 08:16:33.864 - stderr> 	found org.sonatype.sisu.inject#cglib;2.2.1-v20090111 in central
2019-09-09 08:16:33.883 - stderr> 	found asm#asm;3.2 in central
2019-09-09 08:16:33.903 - stderr> 	found com.google.protobuf#protobuf-java;2.5.0 in central
2019-09-09 08:16:33.926 - stderr> 	found commons-io#commons-io;2.4 in central
2019-09-09 08:16:33.951 - stderr> 	found com.sun.jersey#jersey-json;1.14 in central
2019-09-09 08:16:33.976 - stderr> 	found org.codehaus.jettison#jettison;1.1 in central
2019-09-09 08:16:33.999 - stderr> 	found com.sun.xml.bind#jaxb-impl;2.2.3-1 in central
2019-09-09 08:16:34.019 - stderr> 	found javax.xml.bind#jaxb-api;2.2.2 in central
2019-09-09 08:16:34.034 - stderr> 	found javax.xml.stream#stax-api;1.0-2 in central
2019-09-09 08:16:34.05 - stderr> 	found javax.activation#activation;1.1 in central
2019-09-09 08:16:34.066 - stderr> 	found org.codehaus.jackson#jackson-core-asl;1.9.2 in central
2019-09-09 08:16:34.082 - stderr> 	found org.codehaus.jackson#jackson-mapper-asl;1.9.2 in central
2019-09-09 08:16:34.097 - stderr> 	found org.codehaus.jackson#jackson-jaxrs;1.9.2 in central
2019-09-09 08:16:34.115 - stderr> 	found org.codehaus.jackson#jackson-xc;1.9.2 in central
2019-09-09 08:16:34.134 - stderr> 	found com.sun.jersey.contribs#jersey-guice;1.9 in central
2019-09-09 08:16:34.165 - stderr> 	found org.apache.hadoop#hadoop-yarn-common;2.6.0 in central
2019-09-09 08:16:34.199 - stderr> 	found org.apache.hadoop#hadoop-yarn-api;2.6.0 in central
2019-09-09 08:16:34.269 - stderr> 	found org.apache.commons#commons-compress;1.4.1 in central
2019-09-09 08:16:34.295 - stderr> 	found org.tukaani#xz;1.0 in central
2019-09-09 08:16:34.348 - stderr> 	found org.mortbay.jetty#jetty-util;6.1.26 in central
2019-09-09 08:16:34.372 - stderr> 	found com.sun.jersey#jersey-core;1.14 in central
2019-09-09 08:16:34.389 - stderr> 	found com.sun.jersey#jersey-client;1.9 in central
2019-09-09 08:16:34.418 - stderr> 	found commons-cli#commons-cli;1.2 in central
2019-09-09 08:16:34.457 - stderr> 	found com.sun.jersey#jersey-server;1.14 in central
2019-09-09 08:16:34.508 - stderr> 	found org.apache.hadoop#hadoop-yarn-server-common;2.6.0 in central
2019-09-09 08:16:34.536 - stderr> 	found org.fusesource.leveldbjni#leveldbjni-all;1.8 in central
2019-09-09 08:16:34.564 - stderr> 	found org.apache.hadoop#hadoop-yarn-server-applicationhistoryservice;2.6.0 in central
2019-09-09 08:16:34.612 - stderr> 	found commons-collections#commons-collections;3.2.2 in central
2019-09-09 08:16:34.63 - stderr> 	found org.apache.hadoop#hadoop-yarn-server-web-proxy;2.6.0 in central
2019-09-09 08:16:34.646 - stderr> 	found commons-httpclient#commons-httpclient;3.0.1 in central
2019-09-09 08:16:34.66 - stderr> 	found junit#junit;4.11 in central
2019-09-09 08:16:34.675 - stderr> 	found org.hamcrest#hamcrest-core;1.3 in central
2019-09-09 08:16:34.691 - stderr> 	found org.mortbay.jetty#jetty;6.1.26 in central
2019-09-09 08:16:34.722 - stderr> 	found org.apache.hive.shims#hive-shims-scheduler;1.2.2 in central
2019-09-09 08:16:34.736 - stderr> 	found joda-time#joda-time;2.5 in central
2019-09-09 08:16:34.745 - stderr> 	found org.apache.ant#ant;1.9.1 in central
2019-09-09 08:16:34.756 - stderr> 	found org.apache.ant#ant-launcher;1.9.1 in central
2019-09-09 08:16:34.766 - stderr> 	found org.json#json;20090211 in central
2019-09-09 08:16:34.776 - stderr> 	found com.google.code.findbugs#jsr305;3.0.0 in central
2019-09-09 08:16:34.787 - stderr> 	found org.apache.avro#avro;1.7.5 in central
2019-09-09 08:16:34.8 - stderr> 	found com.thoughtworks.paranamer#paranamer;2.3 in central
2019-09-09 08:16:34.811 - stderr> 	found org.xerial.snappy#snappy-java;1.0.5 in central
2019-09-09 08:16:34.828 - stderr> 	found net.sf.opencsv#opencsv;2.3 in central
2019-09-09 08:16:34.839 - stderr> 	found com.twitter#parquet-hadoop-bundle;1.6.0 in central
2019-09-09 08:16:34.851 - stderr> 	found com.jolbox#bonecp;0.8.0.RELEASE in central
2019-09-09 08:16:34.861 - stderr> 	found org.apache.derby#derby;10.10.2.0 in central
2019-09-09 08:16:34.87 - stderr> 	found org.datanucleus#datanucleus-api-jdo;3.2.6 in central
2019-09-09 08:16:34.879 - stderr> 	found org.datanucleus#datanucleus-core;3.2.10 in central
2019-09-09 08:16:34.889 - stderr> 	found org.datanucleus#datanucleus-rdbms;3.2.9 in central
2019-09-09 08:16:34.897 - stderr> 	found commons-pool#commons-pool;1.5.4 in central
2019-09-09 08:16:34.905 - stderr> 	found commons-dbcp#commons-dbcp;1.4 in central
2019-09-09 08:16:34.914 - stderr> 	found javax.jdo#jdo-api;3.0.1 in central
2019-09-09 08:16:34.922 - stderr> 	found javax.transaction#jta;1.1 in central
2019-09-09 08:16:34.929 - stderr> 	found org.antlr#antlr-runtime;3.4 in central
2019-09-09 08:16:34.939 - stderr> 	found org.antlr#stringtemplate;3.2.1 in central
2019-09-09 08:16:34.95 - stderr> 	found antlr#antlr;2.7.7 in central
2019-09-09 08:16:34.961 - stderr> 	found org.apache.thrift#libfb303;0.9.2 in central
2019-09-09 08:16:34.976 - stderr> 	found org.apache.hive#hive-exec;1.2.2 in central
2019-09-09 08:16:34.991 - stderr> 	found org.apache.hive#hive-ant;1.2.2 in central
2019-09-09 08:16:35.01 - stderr> 	found org.apache.velocity#velocity;1.5 in central
2019-09-09 08:16:35.024 - stderr> 	found oro#oro;2.0.8 in central
2019-09-09 08:16:35.065 - stderr> 	found org.antlr#ST4;4.0.4 in central
2019-09-09 08:16:35.09 - stderr> 	found org.apache.ivy#ivy;2.4.0 in central
2019-09-09 08:16:35.109 - stderr> 	found org.codehaus.groovy#groovy-all;2.1.6 in central
2019-09-09 08:16:35.123 - stderr> 	found org.apache.calcite#calcite-core;1.2.0-incubating in central
2019-09-09 08:16:35.137 - stderr> 	found org.apache.calcite#calcite-avatica;1.2.0-incubating in central
2019-09-09 08:16:35.15 - stderr> 	found org.apache.calcite#calcite-linq4j;1.2.0-incubating in central
2019-09-09 08:16:35.171 - stderr> 	found net.hydromatic#eigenbase-properties;1.1.5 in central
2019-09-09 08:16:35.18 - stderr> 	found org.codehaus.janino#janino;2.7.6 in central
2019-09-09 08:16:35.19 - stderr> 	found org.codehaus.janino#commons-compiler;2.7.6 in central
2019-09-09 08:16:35.203 - stderr> 	found stax#stax-api;1.0.1 in central
2019-09-09 08:16:35.232 - stderr> 	found org.apache.hadoop#hadoop-client;2.7.3 in central
2019-09-09 08:16:35.257 - stderr> 	found org.apache.hadoop#hadoop-common;2.7.3 in central
2019-09-09 08:16:35.276 - stderr> 	found org.apache.hadoop#hadoop-annotations;2.7.3 in central
2019-09-09 08:16:35.297 - stderr> 	found org.apache.commons#commons-math3;3.1.1 in central
2019-09-09 08:16:35.311 - stderr> 	found xmlenc#xmlenc;0.52 in central
2019-09-09 08:16:35.322 - stderr> 	found commons-httpclient#commons-httpclient;3.1 in central
2019-09-09 08:16:35.349 - stderr> 	found commons-net#commons-net;3.1 in central
2019-09-09 08:16:35.377 - stderr> 	found log4j#log4j;1.2.17 in central
2019-09-09 08:16:35.395 - stderr> 	found commons-configuration#commons-configuration;1.6 in central
2019-09-09 08:16:35.412 - stderr> 	found commons-digester#commons-digester;1.8 in central
2019-09-09 08:16:35.428 - stderr> 	found commons-beanutils#commons-beanutils;1.7.0 in central
2019-09-09 08:16:35.442 - stderr> 	found commons-beanutils#commons-beanutils-core;1.8.0 in central
2019-09-09 08:16:35.471 - stderr> 	found org.slf4j#slf4j-api;1.7.10 in central
2019-09-09 08:16:35.479 - stderr> 	found org.codehaus.jackson#jackson-core-asl;1.9.13 in central
2019-09-09 08:16:35.489 - stderr> 	found org.codehaus.jackson#jackson-mapper-asl;1.9.13 in central
2019-09-09 08:16:35.511 - stderr> 	found com.google.code.gson#gson;2.2.4 in central
2019-09-09 08:16:35.522 - stderr> 	found org.apache.hadoop#hadoop-auth;2.7.3 in central
2019-09-09 08:16:35.54 - stderr> 	found org.apache.directory.server#apacheds-kerberos-codec;2.0.0-M15 in central
2019-09-09 08:16:35.552 - stderr> 	found org.apache.directory.server#apacheds-i18n;2.0.0-M15 in central
2019-09-09 08:16:35.563 - stderr> 	found org.apache.directory.api#api-asn1-api;1.0.0-M20 in central
2019-09-09 08:16:35.574 - stderr> 	found org.apache.directory.api#api-util;1.0.0-M20 in central
2019-09-09 08:16:35.59 - stderr> 	found org.apache.htrace#htrace-core;3.1.0-incubating in central
2019-09-09 08:16:35.611 - stderr> 	found javax.servlet.jsp#jsp-api;2.1 in central
2019-09-09 08:16:35.631 - stderr> 	found org.slf4j#slf4j-log4j12;1.7.10 in central
2019-09-09 08:16:35.658 - stderr> 	found org.apache.hadoop#hadoop-hdfs;2.7.3 in central
2019-09-09 08:16:35.686 - stderr> 	found io.netty#netty-all;4.0.23.Final in central
2019-09-09 08:16:35.693 - stderr> 	found xerces#xercesImpl;2.9.1 in central
2019-09-09 08:16:35.701 - stderr> 	found xml-apis#xml-apis;1.3.04 in central
2019-09-09 08:16:35.715 - stderr> 	found org.apache.hadoop#hadoop-mapreduce-client-app;2.7.3 in central
2019-09-09 08:16:35.724 - stderr> 	found org.apache.hadoop#hadoop-mapreduce-client-common;2.7.3 in central
2019-09-09 08:16:35.734 - stderr> 	found org.apache.hadoop#hadoop-yarn-common;2.7.3 in central
2019-09-09 08:16:35.749 - stderr> 	found org.apache.hadoop#hadoop-yarn-api;2.7.3 in central
2019-09-09 08:16:35.815 - stderr> 	found org.codehaus.jackson#jackson-jaxrs;1.9.13 in central
2019-09-09 08:16:35.827 - stderr> 	found org.codehaus.jackson#jackson-xc;1.9.13 in central
2019-09-09 08:16:35.886 - stderr> 	found org.apache.hadoop#hadoop-yarn-client;2.7.3 in central
2019-09-09 08:16:35.9 - stderr> 	found org.apache.hadoop#hadoop-mapreduce-client-core;2.7.3 in central
2019-09-09 08:16:35.915 - stderr> 	found org.apache.hadoop#hadoop-yarn-server-common;2.7.3 in central
2019-09-09 08:16:35.935 - stderr> 	found org.apache.hadoop#hadoop-mapreduce-client-shuffle;2.7.3 in central
2019-09-09 08:16:35.95 - stderr> 	found org.apache.hadoop#hadoop-mapreduce-client-jobclient;2.7.3 in central
2019-09-09 08:16:36.06 - stderr> :: resolution report :: resolve 3044ms :: artifacts dl 65ms
2019-09-09 08:16:36.06 - stderr> 	:: modules in use:
2019-09-09 08:16:36.062 - stderr> 	antlr#antlr;2.7.7 from central in [default]
2019-09-09 08:16:36.063 - stderr> 	aopalliance#aopalliance;1.0 from central in [default]
2019-09-09 08:16:36.063 - stderr> 	asm#asm;3.2 from central in [default]
2019-09-09 08:16:36.063 - stderr> 	com.google.code.findbugs#jsr305;3.0.0 from central in [default]
2019-09-09 08:16:36.063 - stderr> 	com.google.code.gson#gson;2.2.4 from central in [default]
2019-09-09 08:16:36.064 - stderr> 	com.google.guava#guava;14.0.1 from central in [default]
2019-09-09 08:16:36.064 - stderr> 	com.google.inject#guice;3.0 from central in [default]
2019-09-09 08:16:36.064 - stderr> 	com.google.inject.extensions#guice-servlet;3.0 from central in [default]
2019-09-09 08:16:36.064 - stderr> 	com.google.protobuf#protobuf-java;2.5.0 from central in [default]
2019-09-09 08:16:36.065 - stderr> 	com.jolbox#bonecp;0.8.0.RELEASE from central in [default]
2019-09-09 08:16:36.065 - stderr> 	com.sun.jersey#jersey-client;1.9 from central in [default]
2019-09-09 08:16:36.065 - stderr> 	com.sun.jersey#jersey-core;1.14 from central in [default]
2019-09-09 08:16:36.065 - stderr> 	com.sun.jersey#jersey-json;1.14 from central in [default]
2019-09-09 08:16:36.066 - stderr> 	com.sun.jersey#jersey-server;1.14 from central in [default]
2019-09-09 08:16:36.066 - stderr> 	com.sun.jersey.contribs#jersey-guice;1.9 from central in [default]
2019-09-09 08:16:36.066 - stderr> 	com.sun.xml.bind#jaxb-impl;2.2.3-1 from central in [default]
2019-09-09 08:16:36.067 - stderr> 	com.thoughtworks.paranamer#paranamer;2.3 from central in [default]
2019-09-09 08:16:36.067 - stderr> 	com.twitter#parquet-hadoop-bundle;1.6.0 from central in [default]
2019-09-09 08:16:36.067 - stderr> 	commons-beanutils#commons-beanutils;1.7.0 from central in [default]
2019-09-09 08:16:36.067 - stderr> 	commons-beanutils#commons-beanutils-core;1.8.0 from central in [default]
2019-09-09 08:16:36.068 - stderr> 	commons-cli#commons-cli;1.2 from central in [default]
2019-09-09 08:16:36.068 - stderr> 	commons-codec#commons-codec;1.4 from central in [default]
2019-09-09 08:16:36.068 - stderr> 	commons-collections#commons-collections;3.2.2 from central in [default]
2019-09-09 08:16:36.068 - stderr> 	commons-configuration#commons-configuration;1.6 from central in [default]
2019-09-09 08:16:36.069 - stderr> 	commons-dbcp#commons-dbcp;1.4 from central in [default]
2019-09-09 08:16:36.069 - stderr> 	commons-digester#commons-digester;1.8 from central in [default]
2019-09-09 08:16:36.069 - stderr> 	commons-httpclient#commons-httpclient;3.1 from central in [default]
2019-09-09 08:16:36.069 - stderr> 	commons-io#commons-io;2.4 from central in [default]
2019-09-09 08:16:36.07 - stderr> 	commons-lang#commons-lang;2.6 from central in [default]
2019-09-09 08:16:36.07 - stderr> 	commons-logging#commons-logging;1.1.3 from central in [default]
2019-09-09 08:16:36.07 - stderr> 	commons-net#commons-net;3.1 from central in [default]
2019-09-09 08:16:36.07 - stderr> 	commons-pool#commons-pool;1.5.4 from central in [default]
2019-09-09 08:16:36.071 - stderr> 	io.netty#netty;3.7.0.Final from central in [default]
2019-09-09 08:16:36.071 - stderr> 	io.netty#netty-all;4.0.23.Final from central in [default]
2019-09-09 08:16:36.071 - stderr> 	javax.activation#activation;1.1 from central in [default]
2019-09-09 08:16:36.071 - stderr> 	javax.inject#javax.inject;1 from central in [default]
2019-09-09 08:16:36.072 - stderr> 	javax.jdo#jdo-api;3.0.1 from central in [default]
2019-09-09 08:16:36.072 - stderr> 	javax.servlet.jsp#jsp-api;2.1 from central in [default]
2019-09-09 08:16:36.072 - stderr> 	javax.transaction#jta;1.1 from central in [default]
2019-09-09 08:16:36.072 - stderr> 	javax.xml.bind#jaxb-api;2.2.2 from central in [default]
2019-09-09 08:16:36.072 - stderr> 	javax.xml.stream#stax-api;1.0-2 from central in [default]
2019-09-09 08:16:36.073 - stderr> 	jline#jline;2.12 from central in [default]
2019-09-09 08:16:36.073 - stderr> 	joda-time#joda-time;2.5 from central in [default]
2019-09-09 08:16:36.073 - stderr> 	log4j#apache-log4j-extras;1.2.17 from central in [default]
2019-09-09 08:16:36.073 - stderr> 	log4j#log4j;1.2.17 from central in [default]
2019-09-09 08:16:36.074 - stderr> 	net.hydromatic#eigenbase-properties;1.1.5 from central in [default]
2019-09-09 08:16:36.074 - stderr> 	net.sf.opencsv#opencsv;2.3 from central in [default]
2019-09-09 08:16:36.074 - stderr> 	org.antlr#ST4;4.0.4 from central in [default]
2019-09-09 08:16:36.074 - stderr> 	org.antlr#antlr-runtime;3.4 from central in [default]
2019-09-09 08:16:36.075 - stderr> 	org.antlr#stringtemplate;3.2.1 from central in [default]
2019-09-09 08:16:36.075 - stderr> 	org.apache.ant#ant;1.9.1 from central in [default]
2019-09-09 08:16:36.075 - stderr> 	org.apache.ant#ant-launcher;1.9.1 from central in [default]
2019-09-09 08:16:36.075 - stderr> 	org.apache.avro#avro;1.7.5 from central in [default]
2019-09-09 08:16:36.076 - stderr> 	org.apache.calcite#calcite-avatica;1.2.0-incubating from central in [default]
2019-09-09 08:16:36.076 - stderr> 	org.apache.calcite#calcite-core;1.2.0-incubating from central in [default]
2019-09-09 08:16:36.076 - stderr> 	org.apache.calcite#calcite-linq4j;1.2.0-incubating from central in [default]
2019-09-09 08:16:36.076 - stderr> 	org.apache.commons#commons-compress;1.4.1 from central in [default]
2019-09-09 08:16:36.077 - stderr> 	org.apache.commons#commons-math3;3.1.1 from central in [default]
2019-09-09 08:16:36.077 - stderr> 	org.apache.derby#derby;10.10.2.0 from central in [default]
2019-09-09 08:16:36.077 - stderr> 	org.apache.directory.api#api-asn1-api;1.0.0-M20 from central in [default]
2019-09-09 08:16:36.077 - stderr> 	org.apache.directory.api#api-util;1.0.0-M20 from central in [default]
2019-09-09 08:16:36.078 - stderr> 	org.apache.directory.server#apacheds-i18n;2.0.0-M15 from central in [default]
2019-09-09 08:16:36.078 - stderr> 	org.apache.directory.server#apacheds-kerberos-codec;2.0.0-M15 from central in [default]
2019-09-09 08:16:36.078 - stderr> 	org.apache.hadoop#hadoop-annotations;2.7.3 from central in [default]
2019-09-09 08:16:36.078 - stderr> 	org.apache.hadoop#hadoop-auth;2.7.3 from central in [default]
2019-09-09 08:16:36.078 - stderr> 	org.apache.hadoop#hadoop-client;2.7.3 from central in [default]
2019-09-09 08:16:36.079 - stderr> 	org.apache.hadoop#hadoop-common;2.7.3 from central in [default]
2019-09-09 08:16:36.079 - stderr> 	org.apache.hadoop#hadoop-hdfs;2.7.3 from central in [default]
2019-09-09 08:16:36.079 - stderr> 	org.apache.hadoop#hadoop-mapreduce-client-app;2.7.3 from central in [default]
2019-09-09 08:16:36.079 - stderr> 	org.apache.hadoop#hadoop-mapreduce-client-common;2.7.3 from central in [default]
2019-09-09 08:16:36.08 - stderr> 	org.apache.hadoop#hadoop-mapreduce-client-core;2.7.3 from central in [default]
2019-09-09 08:16:36.08 - stderr> 	org.apache.hadoop#hadoop-mapreduce-client-jobclient;2.7.3 from central in [default]
2019-09-09 08:16:36.08 - stderr> 	org.apache.hadoop#hadoop-mapreduce-client-shuffle;2.7.3 from central in [default]
2019-09-09 08:16:36.08 - stderr> 	org.apache.hadoop#hadoop-yarn-api;2.7.3 from central in [default]
2019-09-09 08:16:36.081 - stderr> 	org.apache.hadoop#hadoop-yarn-client;2.7.3 from central in [default]
2019-09-09 08:16:36.081 - stderr> 	org.apache.hadoop#hadoop-yarn-common;2.7.3 from central in [default]
2019-09-09 08:16:36.081 - stderr> 	org.apache.hadoop#hadoop-yarn-server-applicationhistoryservice;2.6.0 from central in [default]
2019-09-09 08:16:36.082 - stderr> 	org.apache.hadoop#hadoop-yarn-server-common;2.7.3 from central in [default]
2019-09-09 08:16:36.082 - stderr> 	org.apache.hadoop#hadoop-yarn-server-resourcemanager;2.6.0 from central in [default]
2019-09-09 08:16:36.082 - stderr> 	org.apache.hadoop#hadoop-yarn-server-web-proxy;2.6.0 from central in [default]
2019-09-09 08:16:36.082 - stderr> 	org.apache.hive#hive-ant;1.2.2 from central in [default]
2019-09-09 08:16:36.083 - stderr> 	org.apache.hive#hive-common;1.2.2 from central in [default]
2019-09-09 08:16:36.083 - stderr> 	org.apache.hive#hive-exec;1.2.2 from central in [default]
2019-09-09 08:16:36.083 - stderr> 	org.apache.hive#hive-metastore;1.2.2 from central in [default]
2019-09-09 08:16:36.083 - stderr> 	org.apache.hive#hive-serde;1.2.2 from central in [default]
2019-09-09 08:16:36.083 - stderr> 	org.apache.hive#hive-shims;1.2.2 from central in [default]
2019-09-09 08:16:36.084 - stderr> 	org.apache.hive.shims#hive-shims-0.20S;1.2.2 from central in [default]
2019-09-09 08:16:36.084 - stderr> 	org.apache.hive.shims#hive-shims-0.23;1.2.2 from central in [default]
2019-09-09 08:16:36.084 - stderr> 	org.apache.hive.shims#hive-shims-common;1.2.2 from central in [default]
2019-09-09 08:16:36.084 - stderr> 	org.apache.hive.shims#hive-shims-scheduler;1.2.2 from central in [default]
2019-09-09 08:16:36.085 - stderr> 	org.apache.htrace#htrace-core;3.1.0-incubating from central in [default]
2019-09-09 08:16:36.085 - stderr> 	org.apache.httpcomponents#httpclient;4.4 from central in [default]
2019-09-09 08:16:36.085 - stderr> 	org.apache.httpcomponents#httpcore;4.4 from central in [default]
2019-09-09 08:16:36.085 - stderr> 	org.apache.ivy#ivy;2.4.0 from central in [default]
2019-09-09 08:16:36.086 - stderr> 	org.apache.thrift#libfb303;0.9.2 from central in [default]
2019-09-09 08:16:36.086 - stderr> 	org.apache.thrift#libthrift;0.9.2 from central in [default]
2019-09-09 08:16:36.086 - stderr> 	org.apache.velocity#velocity;1.5 from central in [default]
2019-09-09 08:16:36.086 - stderr> 	org.apache.zookeeper#zookeeper;3.4.6 from central in [default]
2019-09-09 08:16:36.087 - stderr> 	org.codehaus.groovy#groovy-all;2.1.6 from central in [default]
2019-09-09 08:16:36.087 - stderr> 	org.codehaus.jackson#jackson-core-asl;1.9.13 from central in [default]
2019-09-09 08:16:36.087 - stderr> 	org.codehaus.jackson#jackson-jaxrs;1.9.13 from central in [default]
2019-09-09 08:16:36.087 - stderr> 	org.codehaus.jackson#jackson-mapper-asl;1.9.13 from central in [default]
2019-09-09 08:16:36.088 - stderr> 	org.codehaus.jackson#jackson-xc;1.9.13 from central in [default]
2019-09-09 08:16:36.088 - stderr> 	org.codehaus.janino#commons-compiler;2.7.6 from central in [default]
2019-09-09 08:16:36.088 - stderr> 	org.codehaus.janino#janino;2.7.6 from central in [default]
2019-09-09 08:16:36.088 - stderr> 	org.codehaus.jettison#jettison;1.1 from central in [default]
2019-09-09 08:16:36.089 - stderr> 	org.datanucleus#datanucleus-api-jdo;3.2.6 from central in [default]
2019-09-09 08:16:36.089 - stderr> 	org.datanucleus#datanucleus-core;3.2.10 from central in [default]
2019-09-09 08:16:36.089 - stderr> 	org.datanucleus#datanucleus-rdbms;3.2.9 from central in [default]
2019-09-09 08:16:36.089 - stderr> 	org.fusesource.leveldbjni#leveldbjni-all;1.8 from central in [default]
2019-09-09 08:16:36.09 - stderr> 	org.json#json;20090211 from central in [default]
2019-09-09 08:16:36.09 - stderr> 	org.mortbay.jetty#jetty;6.1.26 from central in [default]
2019-09-09 08:16:36.09 - stderr> 	org.mortbay.jetty#jetty-util;6.1.26 from central in [default]
2019-09-09 08:16:36.09 - stderr> 	org.slf4j#slf4j-api;1.7.10 from central in [default]
2019-09-09 08:16:36.091 - stderr> 	org.slf4j#slf4j-log4j12;1.7.10 from central in [default]
2019-09-09 08:16:36.091 - stderr> 	org.sonatype.sisu.inject#cglib;2.2.1-v20090111 from central in [default]
2019-09-09 08:16:36.091 - stderr> 	org.tukaani#xz;1.0 from central in [default]
2019-09-09 08:16:36.091 - stderr> 	org.xerial.snappy#snappy-java;1.0.5 from central in [default]
2019-09-09 08:16:36.091 - stderr> 	oro#oro;2.0.8 from central in [default]
2019-09-09 08:16:36.092 - stderr> 	stax#stax-api;1.0.1 from central in [default]
2019-09-09 08:16:36.092 - stderr> 	xerces#xercesImpl;2.9.1 from central in [default]
2019-09-09 08:16:36.092 - stderr> 	xml-apis#xml-apis;1.3.04 from central in [default]
2019-09-09 08:16:36.092 - stderr> 	xmlenc#xmlenc;0.52 from central in [default]
2019-09-09 08:16:36.092 - stderr> 	:: evicted modules:
2019-09-09 08:16:36.093 - stderr> 	log4j#log4j;1.2.16 by [log4j#log4j;1.2.17] in [default]
2019-09-09 08:16:36.093 - stderr> 	org.slf4j#slf4j-api;1.7.5 by [org.slf4j#slf4j-api;1.7.10] in [default]
2019-09-09 08:16:36.093 - stderr> 	org.slf4j#slf4j-log4j12;1.7.5 by [org.slf4j#slf4j-log4j12;1.7.10] in [default]
2019-09-09 08:16:36.093 - stderr> 	org.apache.hadoop#hadoop-annotations;2.6.0 by [org.apache.hadoop#hadoop-annotations;2.7.3] in [default]
2019-09-09 08:16:36.093 - stderr> 	org.codehaus.jackson#jackson-core-asl;1.9.2 by [org.codehaus.jackson#jackson-core-asl;1.9.13] in [default]
2019-09-09 08:16:36.093 - stderr> 	org.codehaus.jackson#jackson-mapper-asl;1.9.2 by [org.codehaus.jackson#jackson-mapper-asl;1.9.13] in [default]
2019-09-09 08:16:36.094 - stderr> 	org.codehaus.jackson#jackson-jaxrs;1.9.2 by [org.codehaus.jackson#jackson-jaxrs;1.9.13] in [default]
2019-09-09 08:16:36.094 - stderr> 	org.codehaus.jackson#jackson-xc;1.9.2 by [org.codehaus.jackson#jackson-xc;1.9.13] in [default]
2019-09-09 08:16:36.094 - stderr> 	org.apache.hadoop#hadoop-yarn-common;2.6.0 by [org.apache.hadoop#hadoop-yarn-common;2.7.3] in [default]
2019-09-09 08:16:36.094 - stderr> 	org.apache.hadoop#hadoop-yarn-api;2.6.0 by [org.apache.hadoop#hadoop-yarn-api;2.7.3] in [default]
2019-09-09 08:16:36.094 - stderr> 	org.apache.hadoop#hadoop-yarn-server-common;2.6.0 by [org.apache.hadoop#hadoop-yarn-server-common;2.7.3] in [default]
2019-09-09 08:16:36.094 - stderr> 	commons-httpclient#commons-httpclient;3.0.1 by [commons-httpclient#commons-httpclient;3.1] in [default]
2019-09-09 08:16:36.094 - stderr> 	junit#junit;4.11 transitively in [default]
2019-09-09 08:16:36.094 - stderr> 	org.hamcrest#hamcrest-core;1.3 transitively in [default]
2019-09-09 08:16:36.094 - stderr> 	com.google.code.findbugs#jsr305;1.3.9 by [com.google.code.findbugs#jsr305;3.0.0] in [default]
2019-09-09 08:16:36.094 - stderr> 	com.google.guava#guava;11.0.2 by [com.google.guava#guava;14.0.1] in [default]
2019-09-09 08:16:36.094 - stderr> 	org.apache.avro#avro;1.7.4 by [org.apache.avro#avro;1.7.5] in [default]
2019-09-09 08:16:36.095 - stderr> 	org.apache.httpcomponents#httpclient;4.2.5 by [org.apache.httpcomponents#httpclient;4.4] in [default]
2019-09-09 08:16:36.095 - stderr> 	io.netty#netty;3.6.2.Final by [io.netty#netty;3.7.0.Final] in [default]
2019-09-09 08:16:36.095 - stderr> 	com.sun.jersey#jersey-core;1.9 by [com.sun.jersey#jersey-core;1.14] in [default]
2019-09-09 08:16:36.095 - stderr> 	com.sun.jersey#jersey-server;1.9 by [com.sun.jersey#jersey-server;1.14] in [default]
2019-09-09 08:16:36.095 - stderr> 	com.sun.jersey#jersey-json;1.9 by [com.sun.jersey#jersey-json;1.14] in [default]
2019-09-09 08:16:36.095 - stderr> 	---------------------------------------------------------------------
2019-09-09 08:16:36.095 - stderr> 	|                  |            modules            ||   artifacts   |
2019-09-09 08:16:36.096 - stderr> 	|       conf       | number| search|dwnlded|evicted|| number|dwnlded|
2019-09-09 08:16:36.096 - stderr> 	---------------------------------------------------------------------
2019-09-09 08:16:36.097 - stderr> 	|      default     |  145  |   0   |   0   |   22  ||  123  |   0   |
2019-09-09 08:16:36.097 - stderr> 	---------------------------------------------------------------------
2019-09-09 08:16:36.141 - stderr> :: retrieving :: org.apache.spark#spark-submit-parent-7932b6ca-48dd-4850-83e5-bef1daf10e13
2019-09-09 08:16:36.142 - stderr> 	confs: [default]
2019-09-09 08:16:36.197 - stderr> 	0 artifacts copied, 123 already retrieved (0kB/55ms)
2019-09-09 08:16:36.385 - stdout> 2019-09-09 08:16:36 INFO  IsolatedClientLoader:54 - Downloaded metastore jars to /tmp/hive-v1_2-3f0c8c30-eb51-4b33-8fc9-21b57b3cb06c
2019-09-09 08:16:36.959 - stdout> 2019-09-09 08:16:36 INFO  HiveMetaStore:589 - 0: Opening raw store with implemenation class:org.apache.hadoop.hive.metastore.ObjectStore
2019-09-09 08:16:36.985 - stdout> 2019-09-09 08:16:36 INFO  ObjectStore:289 - ObjectStore, initialize called
2019-09-09 08:16:37.108 - stdout> 2019-09-09 08:16:37 INFO  Persistence:77 - Property hive.metastore.integral.jdo.pushdown unknown - will be ignored
2019-09-09 08:16:37.108 - stdout> 2019-09-09 08:16:37 INFO  Persistence:77 - Property datanucleus.cache.level2 unknown - will be ignored
2019-09-09 08:16:46.805 - stdout> 2019-09-09 08:16:46 INFO  ObjectStore:370 - Setting MetaStore object pin classes with hive.metastore.cache.pinobjtypes="Table,StorageDescriptor,SerDeInfo,Partition,Database,Type,FieldSchema,Order"
2019-09-09 08:16:48.245 - stdout> 2019-09-09 08:16:48 INFO  Datastore:77 - The class "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as "embedded-only" so does not have its own datastore table.
2019-09-09 08:16:48.246 - stdout> 2019-09-09 08:16:48 INFO  Datastore:77 - The class "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as "embedded-only" so does not have its own datastore table.
2019-09-09 08:16:56.252 - stdout> 2019-09-09 08:16:56 INFO  Datastore:77 - The class "org.apache.hadoop.hive.metastore.model.MFieldSchema" is tagged as "embedded-only" so does not have its own datastore table.
2019-09-09 08:16:56.253 - stdout> 2019-09-09 08:16:56 INFO  Datastore:77 - The class "org.apache.hadoop.hive.metastore.model.MOrder" is tagged as "embedded-only" so does not have its own datastore table.
2019-09-09 08:17:02.114 - stdout> 2019-09-09 08:17:02 INFO  MetaStoreDirectSql:139 - Using direct SQL, underlying DB is DERBY
2019-09-09 08:17:02.116 - stdout> 2019-09-09 08:17:02 INFO  ObjectStore:272 - Initialized ObjectStore
2019-09-09 08:17:02.387 - stdout> 2019-09-09 08:17:02 WARN  ObjectStore:6666 - Version information not found in metastore. hive.metastore.schema.verification is not enabled so recording the schema version 1.2.0
2019-09-09 08:17:02.702 - stdout> 2019-09-09 08:17:02 WARN  ObjectStore:568 - Failed to get database default, returning NoSuchObjectException
2019-09-09 08:17:03.121 - stdout> 2019-09-09 08:17:03 INFO  HiveMetaStore:663 - Added admin role in metastore
2019-09-09 08:17:03.138 - stdout> 2019-09-09 08:17:03 INFO  HiveMetaStore:672 - Added public role in metastore
2019-09-09 08:17:03.596 - stdout> 2019-09-09 08:17:03 INFO  HiveMetaStore:712 - No user is added in admin role, since config is empty
2019-09-09 08:17:03.686 - stdout> 2019-09-09 08:17:03 INFO  HiveMetaStore:746 - 0: get_all_databases
2019-09-09 08:17:03.687 - stdout> 2019-09-09 08:17:03 INFO  audit:371 - ugi=jenkins	ip=unknown-ip-addr	cmd=get_all_databases	
2019-09-09 08:17:03.702 - stdout> 2019-09-09 08:17:03 INFO  HiveMetaStore:746 - 0: get_functions: db=default pat=*
2019-09-09 08:17:03.702 - stdout> 2019-09-09 08:17:03 INFO  audit:371 - ugi=jenkins	ip=unknown-ip-addr	cmd=get_functions: db=default pat=*	
2019-09-09 08:17:03.704 - stdout> 2019-09-09 08:17:03 INFO  Datastore:77 - The class "org.apache.hadoop.hive.metastore.model.MResourceUri" is tagged as "embedded-only" so does not have its own datastore table.
2019-09-09 08:17:05.211 - stdout> 2019-09-09 08:17:05 INFO  SessionState:641 - Created local directory: /tmp/0d0c2532-9023-4376-9990-d239a122eb81_resources
2019-09-09 08:17:05.214 - stdout> 2019-09-09 08:17:05 INFO  SessionState:641 - Created HDFS directory: /tmp/hive/jenkins/0d0c2532-9023-4376-9990-d239a122eb81
2019-09-09 08:17:05.218 - stdout> 2019-09-09 08:17:05 INFO  SessionState:641 - Created local directory: /tmp/jenkins/0d0c2532-9023-4376-9990-d239a122eb81
2019-09-09 08:17:05.222 - stdout> 2019-09-09 08:17:05 INFO  SessionState:641 - Created HDFS directory: /tmp/hive/jenkins/0d0c2532-9023-4376-9990-d239a122eb81/_tmp_space.db
2019-09-09 08:17:05.225 - stdout> 2019-09-09 08:17:05 INFO  HiveClientImpl:54 - Warehouse location for Hive client (version 1.2.2) is /home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/warehouse-67f09022-7724-465b-90ac-4a1a2ad4f64d
2019-09-09 08:17:05.235 - stdout> 2019-09-09 08:17:05 INFO  HiveMetaStore:746 - 0: get_database: default
2019-09-09 08:17:05.235 - stdout> 2019-09-09 08:17:05 INFO  audit:371 - ugi=jenkins	ip=unknown-ip-addr	cmd=get_database: default	
2019-09-09 08:17:05.256 - stdout> 2019-09-09 08:17:05 INFO  HiveMetaStore:746 - 0: get_database: global_temp
2019-09-09 08:17:05.256 - stdout> 2019-09-09 08:17:05 INFO  audit:371 - ugi=jenkins	ip=unknown-ip-addr	cmd=get_database: global_temp	
2019-09-09 08:17:05.257 - stdout> 2019-09-09 08:17:05 WARN  ObjectStore:568 - Failed to get database global_temp, returning NoSuchObjectException
2019-09-09 08:17:06.749 - stdout> Traceback (most recent call last):
2019-09-09 08:17:06.749 - stdout>   File "/home/jenkins/workspace/spark-master-test-sbt-hadoop-2.7/sql/hive/target/tmp/org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite/test5652630777716270402.py", line 8, in <module>
2019-09-09 08:17:06.749 - stdout>     spark.sql("create table data_source_tbl_{} using json as select 1 i".format(version_index))
2019-09-09 08:17:06.749 - stdout>   File "/tmp/test-spark/spark-2.3.4/python/lib/pyspark.zip/pyspark/sql/session.py", line 710, in sql
2019-09-09 08:17:06.75 - stdout>   File "/tmp/test-spark/spark-2.3.4/python/lib/py4j-0.10.7-src.zip/py4j/java_gateway.py", line 1257, in __call__
2019-09-09 08:17:06.75 - stdout>   File "/tmp/test-spark/spark-2.3.4/python/lib/pyspark.zip/pyspark/sql/utils.py", line 63, in deco
2019-09-09 08:17:06.75 - stdout>   File "/tmp/test-spark/spark-2.3.4/python/lib/py4j-0.10.7-src.zip/py4j/protocol.py", line 328, in get_return_value
2019-09-09 08:17:06.751 - stdout> py4j.protocol.Py4JJavaError: An error occurred while calling o29.sql.
2019-09-09 08:17:06.751 - stdout> : java.lang.ExceptionInInitializerError
2019-09-09 08:17:06.751 - stdout> 	at org.apache.spark.SparkContext.withScope(SparkContext.scala:693)
2019-09-09 08:17:06.751 - stdout> 	at org.apache.spark.SparkContext.parallelize(SparkContext.scala:710)
2019-09-09 08:17:06.751 - stdout> 	at org.apache.spark.sql.execution.SparkStrategies.singleRowRdd$lzycompute(SparkStrategies.scala:399)
2019-09-09 08:17:06.751 - stdout> 	at org.apache.spark.sql.execution.SparkStrategies.singleRowRdd(SparkStrategies.scala:399)
2019-09-09 08:17:06.752 - stdout> 	at org.apache.spark.sql.execution.SparkStrategies$BasicOperators$.apply(SparkStrategies.scala:536)
2019-09-09 08:17:06.752 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:63)
2019-09-09 08:17:06.752 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:63)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:439)
2019-09-09 08:17:06.752 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:93)
2019-09-09 08:17:06.752 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:78)
2019-09-09 08:17:06.752 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:75)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.Iterator$class.foreach(Iterator.scala:893)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336)
2019-09-09 08:17:06.752 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:75)
2019-09-09 08:17:06.752 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:67)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
2019-09-09 08:17:06.752 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:93)
2019-09-09 08:17:06.752 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:78)
2019-09-09 08:17:06.752 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:75)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.Iterator$class.foreach(Iterator.scala:893)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
2019-09-09 08:17:06.752 - stdout> 	at scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157)
2019-09-09 08:17:06.753 - stdout> 	at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:75)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:67)
2019-09-09 08:17:06.753 - stdout> 	at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
2019-09-09 08:17:06.753 - stdout> 	at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:93)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:78)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2$$anonfun$apply$2.apply(QueryPlanner.scala:75)
2019-09-09 08:17:06.753 - stdout> 	at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
2019-09-09 08:17:06.753 - stdout> 	at scala.collection.TraversableOnce$$anonfun$foldLeft$1.apply(TraversableOnce.scala:157)
2019-09-09 08:17:06.753 - stdout> 	at scala.collection.Iterator$class.foreach(Iterator.scala:893)
2019-09-09 08:17:06.753 - stdout> 	at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
2019-09-09 08:17:06.753 - stdout> 	at scala.collection.TraversableOnce$class.foldLeft(TraversableOnce.scala:157)
2019-09-09 08:17:06.753 - stdout> 	at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1336)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:75)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$2.apply(QueryPlanner.scala:67)
2019-09-09 08:17:06.753 - stdout> 	at scala.collection.Iterator$$anon$12.nextCur(Iterator.scala:434)
2019-09-09 08:17:06.753 - stdout> 	at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:440)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:93)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.execution.QueryExecution.sparkPlan$lzycompute(QueryExecution.scala:72)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.execution.QueryExecution.sparkPlan(QueryExecution.scala:68)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.execution.QueryExecution.executedPlan$lzycompute(QueryExecution.scala:77)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.execution.QueryExecution.executedPlan(QueryExecution.scala:77)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3260)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.Dataset.<init>(Dataset.scala:190)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:75)
2019-09-09 08:17:06.753 - stdout> 	at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
2019-09-09 08:17:06.753 - stdout> 	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
2019-09-09 08:17:06.754 - stdout> 	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
2019-09-09 08:17:06.754 - stdout> 	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
2019-09-09 08:17:06.754 - stdout> 	at java.lang.reflect.Method.invoke(Method.java:498)
2019-09-09 08:17:06.754 - stdout> 	at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
2019-09-09 08:17:06.754 - stdout> 	at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
2019-09-09 08:17:06.754 - stdout> 	at py4j.Gateway.invoke(Gateway.java:282)
2019-09-09 08:17:06.754 - stdout> 	at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
2019-09-09 08:17:06.754 - stdout> 	at py4j.commands.CallCommand.execute(CallCommand.java:79)
2019-09-09 08:17:06.754 - stdout> 	at py4j.GatewayConnection.run(GatewayConnection.java:238)
2019-09-09 08:17:06.754 - stdout> 	at java.lang.Thread.run(Thread.java:748)
2019-09-09 08:17:06.754 - stdout> Caused by: java.util.NoSuchElementException: key not found: groupId
2019-09-09 08:17:06.754 - stdout> 	at scala.collection.MapLike$class.default(MapLike.scala:228)
2019-09-09 08:17:06.754 - stdout> 	at scala.collection.AbstractMap.default(Map.scala:59)
2019-09-09 08:17:06.754 - stdout> 	at scala.collection.MapLike$class.apply(MapLike.scala:141)
2019-09-09 08:17:06.754 - stdout> 	at scala.collection.AbstractMap.apply(Map.scala:59)
2019-09-09 08:17:06.754 - stdout> 	at com.fasterxml.jackson.module.scala.JacksonModule$.version$lzycompute(JacksonModule.scala:27)
2019-09-09 08:17:06.754 - stdout> 	at com.fasterxml.jackson.module.scala.JacksonModule$.version(JacksonModule.scala:26)
2019-09-09 08:17:06.754 - stdout> 	at com.fasterxml.jackson.module.scala.JacksonModule$class.version(JacksonModule.scala:49)
2019-09-09 08:17:06.754 - stdout> 	at com.fasterxml.jackson.module.scala.DefaultScalaModule.version(DefaultScalaModule.scala:19)
2019-09-09 08:17:06.754 - stdout> 	at com.fasterxml.jackson.databind.ObjectMapper.registerModule(ObjectMapper.java:710)
2019-09-09 08:17:06.754 - stdout> 	at org.apache.spark.rdd.RDDOperationScope$.<init>(RDDOperationScope.scala:82)
2019-09-09 08:17:06.754 - stdout> 	at org.apache.spark.rdd.RDDOperationScope$.<clinit>(RDDOperationScope.scala)
2019-09-09 08:17:06.754 - stdout> 	... 69 more
2019-09-09 08:17:06.754 - stdout> 
2019-09-09 08:17:06.821 - stdout> 2019-09-09 08:17:06 INFO  SparkContext:54 - Invoking stop() from shutdown hook
2019-09-09 08:17:06.835 - stdout> 2019-09-09 08:17:06 INFO  MapOutputTrackerMasterEndpoint:54 - MapOutputTrackerMasterEndpoint stopped!
2019-09-09 08:17:06.846 - stdout> 2019-09-09 08:17:06 INFO  MemoryStore:54 - MemoryStore cleared
2019-09-09 08:17:06.846 - stdout> 2019-09-09 08:17:06 INFO  BlockManager:54 - BlockManager stopped
2019-09-09 08:17:06.848 - stdout> 2019-09-09 08:17:06 INFO  BlockManagerMaster:54 - BlockManagerMaster stopped
2019-09-09 08:17:06.853 - stdout> 2019-09-09 08:17:06 INFO  OutputCommitCoordinator$OutputCommitCoordinatorEndpoint:54 - OutputCommitCoordinator stopped!
2019-09-09 08:17:06.856 - stdout> 2019-09-09 08:17:06 INFO  SparkContext:54 - Successfully stopped SparkContext
2019-09-09 08:17:06.857 - stdout> 2019-09-09 08:17:06 INFO  ShutdownHookManager:54 - Shutdown hook called
2019-09-09 08:17:06.858 - stdout> 2019-09-09 08:17:06 INFO  ShutdownHookManager:54 - Deleting directory /tmp/spark-eae4b1da-f853-4e88-aba1-716c25407600/pyspark-d447c291-a72d-4e8d-817a-0f78591683cc
2019-09-09 08:17:06.859 - stdout> 2019-09-09 08:17:06 INFO  ShutdownHookManager:54 - Deleting directory /tmp/hive-v1_2-3f0c8c30-eb51-4b33-8fc9-21b57b3cb06c
2019-09-09 08:17:06.861 - stdout> 2019-09-09 08:17:06 INFO  ShutdownHookManager:54 - Deleting directory /tmp/spark-c02044a6-1cf1-42c1-a617-3e5739b0362e
2019-09-09 08:17:06.862 - stdout> 2019-09-09 08:17:06 INFO  ShutdownHookManager:54 - Deleting directory /tmp/spark-eae4b1da-f853-4e88-aba1-716c25407600
           
	at org.scalatest.Assertions.newAssertionFailedException(Assertions.scala:528)
	at org.scalatest.Assertions.newAssertionFailedException$(Assertions.scala:527)
	at org.scalatest.FunSuite.newAssertionFailedException(FunSuite.scala:1560)
	at org.scalatest.Assertions.fail(Assertions.scala:1089)
	at org.scalatest.Assertions.fail$(Assertions.scala:1085)
	at org.scalatest.FunSuite.fail(FunSuite.scala:1560)
	at org.apache.spark.sql.hive.SparkSubmitTestUtils.runSparkSubmit(SparkSubmitTestUtils.scala:94)
	at org.apache.spark.sql.hive.SparkSubmitTestUtils.runSparkSubmit$(SparkSubmitTestUtils.scala:41)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.runSparkSubmit(HiveExternalCatalogVersionsSuite.scala:50)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.$anonfun$prepare$1(HiveExternalCatalogVersionsSuite.scala:200)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.$anonfun$prepare$1$adapted(HiveExternalCatalogVersionsSuite.scala:183)
	at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
	at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
	at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.prepare(HiveExternalCatalogVersionsSuite.scala:183)
	at org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite.beforeAll(HiveExternalCatalogVersionsSuite.scala:209)
	at org.scalatest.BeforeAndAfterAll.liftedTree1$1(BeforeAndAfterAll.scala:212)
	at org.scalatest.BeforeAndAfterAll.run(BeforeAndAfterAll.scala:210)
	at org.scalatest.BeforeAndAfterAll.run$(BeforeAndAfterAll.scala:208)
	at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:56)
	at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:314)
	at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:507)
	at sbt.ForkMain$Run$2.call(ForkMain.java:296)
	at sbt.ForkMain$Run$2.call(ForkMain.java:286)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)