Exception thrown while executing query:
== Parsed Logical Plan ==
'Project [unresolvedalias('l.a),unresolvedalias('r.b),unresolvedalias('l.p1),unresolvedalias('r.p2)]
+- 'Join Inner, Some(((('l.a = 'r.a) && ('l.p1 = 'r.p1)) && ('l.p2 = 'r.p2)))
 :- 'UnresolvedRelation `t`, Some(l)
 +- 'UnresolvedRelation `t`, Some(r)

== Analyzed Logical Plan ==
a: bigint, b: string, p1: int, p2: string
Project [a#130689L,b#130715,p1#130691,p2#130717]
+- Join Inner, Some((((a#130689L = a#130714L) && (p1#130691 = p1#130716)) && (p2#130692 = p2#130717)))
 :- Subquery l
 : +- Subquery t
 : +- Relation[a#130689L,b#130690,p1#130691,p2#130692] JSONRelation
 +- Subquery r
 +- Subquery t
 +- Relation[a#130714L,b#130715,p1#130716,p2#130717] JSONRelation

== Optimized Logical Plan ==
Project [a#130689L,b#130715,p1#130691,p2#130717]
+- Join Inner, Some((((a#130689L = a#130714L) && (p1#130691 = p1#130716)) && (p2#130692 = p2#130717)))
 :- Project [p1#130691,a#130689L,p2#130692]
 : +- Relation[a#130689L,b#130690,p1#130691,p2#130692] JSONRelation
 +- Relation[a#130714L,b#130715,p1#130716,p2#130717] JSONRelation

== Physical Plan ==
Project [a#130689L,b#130715,p1#130691,p2#130717]
+- BroadcastHashJoin [a#130689L,p1#130691,p2#130692], [a#130714L,p1#130716,p2#130717], BuildRight
 :- Scan JSONRelation[p1#130691,a#130689L,p2#130692] InputPaths: file:/home/jenkins/workspace/spark-branch-1.6-test-maven-pre-yarn-1.2.1/sql/hive/target/tmp/spark-fd60c5cf-3827-4dde-93c8-58c717026dbe
 +- Scan JSONRelation[a#130714L,b#130715,p1#130716,p2#130717] InputPaths: file:/home/jenkins/workspace/spark-branch-1.6-test-maven-pre-yarn-1.2.1/sql/hive/target/tmp/spark-fd60c5cf-3827-4dde-93c8-58c717026dbe
== Exception ==
org.apache.spark.SparkException: Job 5627 cancelled because SparkContext was shut down
org.apache.spark.SparkException: Job 5627 cancelled because SparkContext was shut down
 at org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:806)
 at org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:804)
 at scala.collection.mutable.HashSet.foreach(HashSet.scala:79)
 at org.apache.spark.scheduler.DAGScheduler.cleanUpAfterSchedulerStop(DAGScheduler.scala:804)
 at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onStop(DAGScheduler.scala:1658)
 at org.apache.spark.util.EventLoop.stop(EventLoop.scala:84)
 at org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:1581)
 at org.apache.spark.SparkContext$$anonfun$stop$9.apply$mcV$sp(SparkContext.scala:1740)
 at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1229)
 at org.apache.spark.SparkContext.stop(SparkContext.scala:1739)
 at org.apache.spark.SparkContext$$anonfun$3.apply$mcV$sp(SparkContext.scala:596)
 at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:267)
 at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ShutdownHookManager.scala:239)
 at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:239)
 at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:239)
 at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1765)
 at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply$mcV$sp(ShutdownHookManager.scala:239)
 at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:239)
 at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:239)
 at scala.util.Try$.apply(Try.scala:161)
 at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:239)
 at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:218)
 at java.lang.Thread.run(Thread.java:745)
 at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620)
 at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832)
 at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845)
 at org.apache.spark.SparkContext.runJob(SparkContext.scala:1858)
 at org.apache.spark.SparkContext.runJob(SparkContext.scala:1929)
 at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:927)
 at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
 at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
 at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
 at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
 at org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:166)
 at org.apache.spark.sql.execution.SparkPlan.executeCollectPublic(SparkPlan.scala:174)
 at org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame$$execute$1$1.apply(DataFrame.scala:1499)
 at org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame$$execute$1$1.apply(DataFrame.scala:1499)
 at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:56)
 at org.apache.spark.sql.DataFrame.withNewExecutionId(DataFrame.scala:2086)
 at org.apache.spark.sql.DataFrame.org$apache$spark$sql$DataFrame$$execute$1(DataFrame.scala:1498)
 at org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame$$collect$1.apply(DataFrame.scala:1503)
 at org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame$$collect$1.apply(DataFrame.scala:1503)
 at org.apache.spark.sql.DataFrame.withCallback(DataFrame.scala:2099)
 at org.apache.spark.sql.DataFrame.org$apache$spark$sql$DataFrame$$collect(DataFrame.scala:1503)
 at org.apache.spark.sql.DataFrame.collect(DataFrame.scala:1480)
 at org.apache.spark.sql.QueryTest$.checkAnswer(QueryTest.scala:313)
 at org.apache.spark.sql.QueryTest.checkAnswer(QueryTest.scala:133)
 at org.apache.spark.sql.sources.HadoopFsRelationTest$$anonfun$checkQueries$1.apply$mcV$sp(hadoopFsRelationSuites.scala:94)
 at org.apache.spark.sql.test.SQLTestUtils$class.withTempTable(SQLTestUtils.scala:143)
 at org.apache.spark.sql.sources.HadoopFsRelationTest.withTempTable(hadoopFsRelationSuites.scala:37)
 at org.apache.spark.sql.sources.HadoopFsRelationTest.checkQueries(hadoopFsRelationSuites.scala:93)
 at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(JsonHadoopFsRelationSuite.scala:55)
 at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(JsonHadoopFsRelationSuite.scala:40)
 at org.apache.spark.sql.test.SQLTestUtils$class.withTempDir(SQLTestUtils.scala:136)
 at org.apache.spark.sql.sources.HadoopFsRelationTest.withTempDir(hadoopFsRelationSuites.scala:37)
 at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1.apply$mcV$sp(JsonHadoopFsRelationSuite.scala:40)
 at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1.apply(JsonHadoopFsRelationSuite.scala:40)
 at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1.apply(JsonHadoopFsRelationSuite.scala:40)
 at org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22)
 at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85)
 at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
 at org.scalatest.Transformer.apply(Transformer.scala:22)
 at org.scalatest.Transformer.apply(Transformer.scala:20)
 at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:166)
 at org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:42)
 at org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:163)
 at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
 at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
 at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306)
 at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:175)
 at org.scalatest.FunSuite.runTest(FunSuite.scala:1555)
 at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
 at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
 at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:413)
 at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401)
 at scala.collection.immutable.List.foreach(List.scala:318)
 at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401)
 at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:396)
 at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:483)
 at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:208)
 at org.scalatest.FunSuite.runTests(FunSuite.scala:1555)
 at org.scalatest.Suite$class.run(Suite.scala:1424)
 at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1555)
 at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
 at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
 at org.scalatest.SuperEngine.runImpl(Engine.scala:545)
 at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:212)
 at org.apache.spark.sql.sources.HadoopFsRelationTest.org$scalatest$BeforeAndAfterAll$$super$run(hadoopFsRelationSuites.scala:37)
 at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:257)
 at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:256)
 at org.apache.spark.sql.sources.HadoopFsRelationTest.run(hadoopFsRelationSuites.scala:37)
 at org.scalatest.Suite$class.callExecuteOnSuite$1(Suite.scala:1492)
 at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1528)
 at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1526)
 at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
 at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108)
 at org.scalatest.Suite$class.runNestedSuites(Suite.scala:1526)
 at org.scalatest.tools.DiscoverySuite.runNestedSuites(DiscoverySuite.scala:29)
 at org.scalatest.Suite$class.run(Suite.scala:1421)
 at org.scalatest.tools.DiscoverySuite.run(DiscoverySuite.scala:29)
 at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55)
 at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2563)
 at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2557)
 at scala.collection.immutable.List.foreach(List.scala:318)
 at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:2557)
 at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1044)
 at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1043)
 at org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:2722)
 at org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1043)
 at org.scalatest.tools.Runner$.main(Runner.scala:860)
 at org.scalatest.tools.Runner.main(Runner.scala)


org.scalatest.exceptions.TestFailedException:
Exception thrown while executing query:
== Parsed Logical Plan ==
'Project [unresolvedalias('l.a),unresolvedalias('r.b),unresolvedalias('l.p1),unresolvedalias('r.p2)]
+- 'Join Inner, Some(((('l.a = 'r.a) && ('l.p1 = 'r.p1)) && ('l.p2 = 'r.p2)))
:- 'UnresolvedRelation `t`, Some(l)
+- 'UnresolvedRelation `t`, Some(r)
== Analyzed Logical Plan ==
a: bigint, b: string, p1: int, p2: string
Project [a#130689L,b#130715,p1#130691,p2#130717]
+- Join Inner, Some((((a#130689L = a#130714L) && (p1#130691 = p1#130716)) && (p2#130692 = p2#130717)))
:- Subquery l
: +- Subquery t
: +- Relation[a#130689L,b#130690,p1#130691,p2#130692] JSONRelation
+- Subquery r
+- Subquery t
+- Relation[a#130714L,b#130715,p1#130716,p2#130717] JSONRelation
== Optimized Logical Plan ==
Project [a#130689L,b#130715,p1#130691,p2#130717]
+- Join Inner, Some((((a#130689L = a#130714L) && (p1#130691 = p1#130716)) && (p2#130692 = p2#130717)))
:- Project [p1#130691,a#130689L,p2#130692]
: +- Relation[a#130689L,b#130690,p1#130691,p2#130692] JSONRelation
+- Relation[a#130714L,b#130715,p1#130716,p2#130717] JSONRelation
== Physical Plan ==
Project [a#130689L,b#130715,p1#130691,p2#130717]
+- BroadcastHashJoin [a#130689L,p1#130691,p2#130692], [a#130714L,p1#130716,p2#130717], BuildRight
:- Scan JSONRelation[p1#130691,a#130689L,p2#130692] InputPaths: file:/home/jenkins/workspace/spark-branch-1.6-test-maven-pre-yarn-1.2.1/sql/hive/target/tmp/spark-fd60c5cf-3827-4dde-93c8-58c717026dbe
+- Scan JSONRelation[a#130714L,b#130715,p1#130716,p2#130717] InputPaths: file:/home/jenkins/workspace/spark-branch-1.6-test-maven-pre-yarn-1.2.1/sql/hive/target/tmp/spark-fd60c5cf-3827-4dde-93c8-58c717026dbe
== Exception ==
org.apache.spark.SparkException: Job 5627 cancelled because SparkContext was shut down
org.apache.spark.SparkException: Job 5627 cancelled because SparkContext was shut down
at org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:806)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:804)
at scala.collection.mutable.HashSet.foreach(HashSet.scala:79)
at org.apache.spark.scheduler.DAGScheduler.cleanUpAfterSchedulerStop(DAGScheduler.scala:804)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onStop(DAGScheduler.scala:1658)
at org.apache.spark.util.EventLoop.stop(EventLoop.scala:84)
at org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:1581)
at org.apache.spark.SparkContext$$anonfun$stop$9.apply$mcV$sp(SparkContext.scala:1740)
at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1229)
at org.apache.spark.SparkContext.stop(SparkContext.scala:1739)
at org.apache.spark.SparkContext$$anonfun$3.apply$mcV$sp(SparkContext.scala:596)
at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:267)
at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ShutdownHookManager.scala:239)
at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:239)
at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:239)
at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1765)
at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply$mcV$sp(ShutdownHookManager.scala:239)
at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:239)
at org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:239)
at scala.util.Try$.apply(Try.scala:161)
at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:239)
at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:218)
at java.lang.Thread.run(Thread.java:745)
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1858)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1929)
at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:927)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:316)
at org.apache.spark.rdd.RDD.collect(RDD.scala:926)
at org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:166)
at org.apache.spark.sql.execution.SparkPlan.executeCollectPublic(SparkPlan.scala:174)
at org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame$$execute$1$1.apply(DataFrame.scala:1499)
at org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame$$execute$1$1.apply(DataFrame.scala:1499)
at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:56)
at org.apache.spark.sql.DataFrame.withNewExecutionId(DataFrame.scala:2086)
at org.apache.spark.sql.DataFrame.org$apache$spark$sql$DataFrame$$execute$1(DataFrame.scala:1498)
at org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame$$collect$1.apply(DataFrame.scala:1503)
at org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame$$collect$1.apply(DataFrame.scala:1503)
at org.apache.spark.sql.DataFrame.withCallback(DataFrame.scala:2099)
at org.apache.spark.sql.DataFrame.org$apache$spark$sql$DataFrame$$collect(DataFrame.scala:1503)
at org.apache.spark.sql.DataFrame.collect(DataFrame.scala:1480)
at org.apache.spark.sql.QueryTest$.checkAnswer(QueryTest.scala:313)
at org.apache.spark.sql.QueryTest.checkAnswer(QueryTest.scala:133)
at org.apache.spark.sql.sources.HadoopFsRelationTest$$anonfun$checkQueries$1.apply$mcV$sp(hadoopFsRelationSuites.scala:94)
at org.apache.spark.sql.test.SQLTestUtils$class.withTempTable(SQLTestUtils.scala:143)
at org.apache.spark.sql.sources.HadoopFsRelationTest.withTempTable(hadoopFsRelationSuites.scala:37)
at org.apache.spark.sql.sources.HadoopFsRelationTest.checkQueries(hadoopFsRelationSuites.scala:93)
at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(JsonHadoopFsRelationSuite.scala:55)
at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(JsonHadoopFsRelationSuite.scala:40)
at org.apache.spark.sql.test.SQLTestUtils$class.withTempDir(SQLTestUtils.scala:136)
at org.apache.spark.sql.sources.HadoopFsRelationTest.withTempDir(hadoopFsRelationSuites.scala:37)
at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1.apply$mcV$sp(JsonHadoopFsRelationSuite.scala:40)
at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1.apply(JsonHadoopFsRelationSuite.scala:40)
at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1.apply(JsonHadoopFsRelationSuite.scala:40)
at org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22)
at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85)
at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
at org.scalatest.Transformer.apply(Transformer.scala:22)
at org.scalatest.Transformer.apply(Transformer.scala:20)
at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:166)
at org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:42)
at org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:163)
at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306)
at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:175)
at org.scalatest.FunSuite.runTest(FunSuite.scala:1555)
at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:413)
at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401)
at scala.collection.immutable.List.foreach(List.scala:318)
at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401)
at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:396)
at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:483)
at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:208)
at org.scalatest.FunSuite.runTests(FunSuite.scala:1555)
at org.scalatest.Suite$class.run(Suite.scala:1424)
at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1555)
at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
at org.scalatest.SuperEngine.runImpl(Engine.scala:545)
at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:212)
at org.apache.spark.sql.sources.HadoopFsRelationTest.org$scalatest$BeforeAndAfterAll$$super$run(hadoopFsRelationSuites.scala:37)
at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:257)
at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:256)
at org.apache.spark.sql.sources.HadoopFsRelationTest.run(hadoopFsRelationSuites.scala:37)
at org.scalatest.Suite$class.callExecuteOnSuite$1(Suite.scala:1492)
at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1528)
at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1526)
at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108)
at org.scalatest.Suite$class.runNestedSuites(Suite.scala:1526)
at org.scalatest.tools.DiscoverySuite.runNestedSuites(DiscoverySuite.scala:29)
at org.scalatest.Suite$class.run(Suite.scala:1421)
at org.scalatest.tools.DiscoverySuite.run(DiscoverySuite.scala:29)
at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55)
at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2563)
at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2557)
at scala.collection.immutable.List.foreach(List.scala:318)
at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:2557)
at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1044)
at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1043)
at org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:2722)
at org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1043)
at org.scalatest.tools.Runner$.main(Runner.scala:860)
at org.scalatest.tools.Runner.main(Runner.scala)
at org.scalatest.Assertions$class.newAssertionFailedException(Assertions.scala:495)
at org.scalatest.FunSuite.newAssertionFailedException(FunSuite.scala:1555)
at org.scalatest.Assertions$class.fail(Assertions.scala:1328)
at org.scalatest.FunSuite.fail(FunSuite.scala:1555)
at org.apache.spark.sql.QueryTest.checkAnswer(QueryTest.scala:134)
at org.apache.spark.sql.sources.HadoopFsRelationTest$$anonfun$checkQueries$1.apply$mcV$sp(hadoopFsRelationSuites.scala:94)
at org.apache.spark.sql.test.SQLTestUtils$class.withTempTable(SQLTestUtils.scala:143)
at org.apache.spark.sql.sources.HadoopFsRelationTest.withTempTable(hadoopFsRelationSuites.scala:37)
at org.apache.spark.sql.sources.HadoopFsRelationTest.checkQueries(hadoopFsRelationSuites.scala:93)
at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(JsonHadoopFsRelationSuite.scala:55)
at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(JsonHadoopFsRelationSuite.scala:40)
at org.apache.spark.sql.test.SQLTestUtils$class.withTempDir(SQLTestUtils.scala:136)
at org.apache.spark.sql.sources.HadoopFsRelationTest.withTempDir(hadoopFsRelationSuites.scala:37)
at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1.apply$mcV$sp(JsonHadoopFsRelationSuite.scala:40)
at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1.apply(JsonHadoopFsRelationSuite.scala:40)
at org.apache.spark.sql.sources.JsonHadoopFsRelationSuite$$anonfun$1.apply(JsonHadoopFsRelationSuite.scala:40)
at org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22)
at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85)
at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
at org.scalatest.Transformer.apply(Transformer.scala:22)
at org.scalatest.Transformer.apply(Transformer.scala:20)
at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:166)
at org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:42)
at org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:163)
at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306)
at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:175)
at org.scalatest.FunSuite.runTest(FunSuite.scala:1555)
at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:413)
at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401)
at scala.collection.immutable.List.foreach(List.scala:318)
at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401)
at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:396)
at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:483)
at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:208)
at org.scalatest.FunSuite.runTests(FunSuite.scala:1555)
at org.scalatest.Suite$class.run(Suite.scala:1424)
at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1555)
at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
at org.scalatest.SuperEngine.runImpl(Engine.scala:545)
at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:212)
at org.apache.spark.sql.sources.HadoopFsRelationTest.org$scalatest$BeforeAndAfterAll$$super$run(hadoopFsRelationSuites.scala:37)
at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:257)
at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:256)
at org.apache.spark.sql.sources.HadoopFsRelationTest.run(hadoopFsRelationSuites.scala:37)
at org.scalatest.Suite$class.callExecuteOnSuite$1(Suite.scala:1492)
at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1528)
at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1526)
at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108)
at org.scalatest.Suite$class.runNestedSuites(Suite.scala:1526)
at org.scalatest.tools.DiscoverySuite.runNestedSuites(DiscoverySuite.scala:29)
at org.scalatest.Suite$class.run(Suite.scala:1421)
at org.scalatest.tools.DiscoverySuite.run(DiscoverySuite.scala:29)
at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55)
at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2563)
at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2557)
at scala.collection.immutable.List.foreach(List.scala:318)
at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:2557)
at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1044)
at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1043)
at org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:2722)
at org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1043)
at org.scalatest.tools.Runner$.main(Runner.scala:860)
at org.scalatest.tools.Runner.main(Runner.scala)