org.scalatest.exceptions.TestFailedException: Failed to execute query using catalyst: Error: execute, tree: Exchange hashpartitioning(p_mfgr#153883, 5) +- Scan hive default.part [p_mfgr#153883, p_name#153882, p_retailprice#153888, p_size#153886], HiveTableRelation `default`.`part`, org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, [p_partkey#153881, p_name#153882, p_mfgr#153883, p_brand#153884, p_type#153885, p_size#153886, p_container#153887, p_retailprice#153888, p_comment#153889] org.apache.spark.sql.catalyst.errors.package$TreeNodeException: execute, tree: Exchange hashpartitioning(p_mfgr#153883, 5) +- Scan hive default.part [p_mfgr#153883, p_name#153882, p_retailprice#153888, p_size#153886], HiveTableRelation `default`.`part`, org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, [p_partkey#153881, p_name#153882, p_mfgr#153883, p_brand#153884, p_type#153885, p_size#153886, p_container#153887, p_retailprice#153888, p_comment#153889] at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:56) at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.doExecute(ShuffleExchangeExec.scala:119) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:131) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:127) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152) at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:127) at org.apache.spark.sql.execution.InputAdapter.inputRDDs(WholeStageCodegenExec.scala:374) at org.apache.spark.sql.execution.SortExec.inputRDDs(SortExec.scala:121) at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:610) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:131) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:127) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152) at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:127) at org.apache.spark.sql.execution.window.WindowExec.doExecute(WindowExec.scala:302) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:131) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:127) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152) at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:127) at org.apache.spark.sql.execution.InputAdapter.inputRDDs(WholeStageCodegenExec.scala:374) at org.apache.spark.sql.execution.ProjectExec.inputRDDs(basicPhysicalOperators.scala:41) at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:610) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:131) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:127) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152) at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:127) at org.apache.spark.sql.execution.SparkPlan.getByteArrayRdd(SparkPlan.scala:247) at org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:294) at org.apache.spark.sql.execution.SparkPlan.executeCollectPublic(SparkPlan.scala:324) at org.apache.spark.sql.execution.QueryExecution.hiveResultString(QueryExecution.scala:121) at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6$$anonfun$25$$anonfun$getResult$1$1.apply(HiveComparisonTest.scala:348) at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6$$anonfun$25$$anonfun$getResult$1$1.apply(HiveComparisonTest.scala:348) at org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:78) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:125) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:73) at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6$$anonfun$25.getResult$1(HiveComparisonTest.scala:348) at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6$$anonfun$25.apply(HiveComparisonTest.scala:350) at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6$$anonfun$25.apply(HiveComparisonTest.scala:345) at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234) at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) at scala.collection.TraversableLike$class.map(TraversableLike.scala:234) at scala.collection.AbstractTraversable.map(Traversable.scala:104) at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6.doTest$1(HiveComparisonTest.scala:345) at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6.apply(HiveComparisonTest.scala:462) at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85) at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104) at org.scalatest.Transformer.apply(Transformer.scala:22) at org.scalatest.Transformer.apply(Transformer.scala:20) at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:186) at org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:103) at org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:183) at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:196) at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:196) at org.scalatest.SuperEngine.runTestImpl(Engine.scala:289) at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:196) at org.apache.spark.sql.hive.execution.HiveWindowFunctionQuerySuite.org$scalatest$BeforeAndAfter$$super$runTest(HiveWindowFunctionQuerySuite.scala:35) at org.scalatest.BeforeAndAfter$class.runTest(BeforeAndAfter.scala:203) at org.apache.spark.sql.hive.execution.HiveWindowFunctionQuerySuite.runTest(HiveWindowFunctionQuerySuite.scala:35) at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:229) at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:229) at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:396) at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:384) at scala.collection.immutable.List.foreach(List.scala:392) at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:384) at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:379) at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:461) at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:229) at org.scalatest.FunSuite.runTests(FunSuite.scala:1560) at org.scalatest.Suite$class.run(Suite.scala:1147) at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1560) at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:233) at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:233) at org.scalatest.SuperEngine.runImpl(Engine.scala:521) at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:233) at org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterAll$$super$run(SparkFunSuite.scala:52) at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:213) at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:210) at org.apache.spark.sql.hive.execution.HiveWindowFunctionQuerySuite.org$scalatest$BeforeAndAfter$$super$run(HiveWindowFunctionQuerySuite.scala:35) at org.scalatest.BeforeAndAfter$class.run(BeforeAndAfter.scala:258) at org.apache.spark.sql.hive.execution.HiveWindowFunctionQuerySuite.run(HiveWindowFunctionQuerySuite.scala:35) at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:314) at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:480) at sbt.ForkMain$Run$2.call(ForkMain.java:296) at sbt.ForkMain$Run$2.call(ForkMain.java:286) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.IllegalStateException: Cannot call methods on a stopped SparkContext. This stopped SparkContext was created at: org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6.apply(HiveComparisonTest.scala:456) org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85) org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104) org.scalatest.Transformer.apply(Transformer.scala:22) org.scalatest.Transformer.apply(Transformer.scala:20) org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:186) org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:103) org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:183) org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:196) org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:196) org.scalatest.SuperEngine.runTestImpl(Engine.scala:289) org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:196) org.scalatest.FunSuite.runTest(FunSuite.scala:1560) org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:229) org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:229) org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:396) org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:384) scala.collection.immutable.List.foreach(List.scala:392) org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:384) org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:379) The currently active SparkContext was created at: (No active SparkContext.) at org.apache.spark.SparkContext.assertNotStopped(SparkContext.scala:100) at org.apache.spark.SparkContext.broadcast(SparkContext.scala:1492) at org.apache.spark.sql.hive.HadoopTableReader.<init>(TableReader.scala:89) at org.apache.spark.sql.hive.execution.HiveTableScanExec.org$apache$spark$sql$hive$execution$HiveTableScanExec$$hadoopReader$lzycompute(HiveTableScanExec.scala:105) at org.apache.spark.sql.hive.execution.HiveTableScanExec.org$apache$spark$sql$hive$execution$HiveTableScanExec$$hadoopReader(HiveTableScanExec.scala:105) at org.apache.spark.sql.hive.execution.HiveTableScanExec$$anonfun$10.apply(HiveTableScanExec.scala:188) at org.apache.spark.sql.hive.execution.HiveTableScanExec$$anonfun$10.apply(HiveTableScanExec.scala:188) at org.apache.spark.util.Utils$.withDummyCallSite(Utils.scala:2457) at org.apache.spark.sql.hive.execution.HiveTableScanExec.doExecute(HiveTableScanExec.scala:187) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:131) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:127) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152) at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:127) at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.prepareShuffleDependency(ShuffleExchangeExec.scala:92) at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec$$anonfun$doExecute$1.apply(ShuffleExchangeExec.scala:128) at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec$$anonfun$doExecute$1.apply(ShuffleExchangeExec.scala:119) at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:52) ... 96 more select p_mfgr, p_name, rank() over(distribute by p_mfgr sort by p_name) as r, dense_rank() over(distribute by p_mfgr sort by p_name) as dr, count(p_size) over(distribute by p_mfgr sort by p_name) as cd, p_retailprice, sum(p_retailprice) over (distribute by p_mfgr sort by p_name rows between unbounded preceding and current row) as s1, p_size, p_size - lag(p_size,1,p_size) over(distribute by p_mfgr sort by p_name) as deltaSz from part == Parsed Logical Plan == 'Project ['p_mfgr, 'p_name, 'rank() windowspecdefinition('p_mfgr, 'p_name ASC NULLS FIRST, unspecifiedframe$()) AS r#153872, 'dense_rank() windowspecdefinition('p_mfgr, 'p_name ASC NULLS FIRST, unspecifiedframe$()) AS dr#153873, 'count('p_size) windowspecdefinition('p_mfgr, 'p_name ASC NULLS FIRST, unspecifiedframe$()) AS cd#153874, 'p_retailprice, 'sum('p_retailprice) windowspecdefinition('p_mfgr, 'p_name ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS s1#153875, 'p_size, ('p_size - 'lag('p_size, 1, 'p_size) windowspecdefinition('p_mfgr, 'p_name ASC NULLS FIRST, unspecifiedframe$())) AS deltaSz#153876] +- 'UnresolvedRelation `part` == Analyzed Logical Plan == p_mfgr: string, p_name: string, r: int, dr: int, cd: bigint, p_retailprice: double, s1: double, p_size: int, deltaSz: int Project [p_mfgr#153883, p_name#153882, r#153872, dr#153873, cd#153874L, p_retailprice#153888, s1#153875, p_size#153886, deltaSz#153876] +- Project [p_mfgr#153883, p_name#153882, p_retailprice#153888, p_size#153886, r#153872, dr#153873, cd#153874L, s1#153875, _we4#153898, r#153872, dr#153873, cd#153874L, s1#153875, (p_size#153886 - _we4#153898) AS deltaSz#153876] +- Window [rank(p_name#153882) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS r#153872, dense_rank(p_name#153882) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS dr#153873, count(p_size#153886) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RangeFrame, unboundedpreceding$(), currentrow$())) AS cd#153874L, sum(p_retailprice#153888) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS s1#153875, lag(p_size#153886, 1, p_size#153886) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, -1, -1)) AS _we4#153898], [p_mfgr#153883], [p_name#153882 ASC NULLS FIRST] +- Project [p_mfgr#153883, p_name#153882, p_retailprice#153888, p_size#153886] +- SubqueryAlias `default`.`part` +- HiveTableRelation `default`.`part`, org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, [p_partkey#153881, p_name#153882, p_mfgr#153883, p_brand#153884, p_type#153885, p_size#153886, p_container#153887, p_retailprice#153888, p_comment#153889] == Optimized Logical Plan == Project [p_mfgr#153883, p_name#153882, r#153872, dr#153873, cd#153874L, p_retailprice#153888, s1#153875, p_size#153886, (p_size#153886 - _we4#153898) AS deltaSz#153876] +- Window [rank(p_name#153882) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS r#153872, dense_rank(p_name#153882) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS dr#153873, count(p_size#153886) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RangeFrame, unboundedpreceding$(), currentrow$())) AS cd#153874L, sum(p_retailprice#153888) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS s1#153875, lag(p_size#153886, 1, p_size#153886) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, -1, -1)) AS _we4#153898], [p_mfgr#153883], [p_name#153882 ASC NULLS FIRST] +- Project [p_mfgr#153883, p_name#153882, p_retailprice#153888, p_size#153886] +- HiveTableRelation `default`.`part`, org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, [p_partkey#153881, p_name#153882, p_mfgr#153883, p_brand#153884, p_type#153885, p_size#153886, p_container#153887, p_retailprice#153888, p_comment#153889] == Physical Plan == *(2) Project [p_mfgr#153883, p_name#153882, r#153872, dr#153873, cd#153874L, p_retailprice#153888, s1#153875, p_size#153886, (p_size#153886 - _we4#153898) AS deltaSz#153876] +- Window [rank(p_name#153882) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS r#153872, dense_rank(p_name#153882) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS dr#153873, count(p_size#153886) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RangeFrame, unboundedpreceding$(), currentrow$())) AS cd#153874L, sum(p_retailprice#153888) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS s1#153875, lag(p_size#153886, 1, p_size#153886) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, -1, -1)) AS _we4#153898], [p_mfgr#153883], [p_name#153882 ASC NULLS FIRST] +- *(1) Sort [p_mfgr#153883 ASC NULLS FIRST, p_name#153882 ASC NULLS FIRST], false, 0 +- Exchange hashpartitioning(p_mfgr#153883, 5) +- Scan hive default.part [p_mfgr#153883, p_name#153882, p_retailprice#153888, p_size#153886], HiveTableRelation `default`.`part`, org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, [p_partkey#153881, p_name#153882, p_mfgr#153883, p_brand#153884, p_type#153885, p_size#153886, p_container#153887, p_retailprice#153888, p_comment#153889] == HIVE - 26 row(s) == Manufacturer#1 almond antique burnished rose metallic 1 1 2 1173.15 1173.15 2 0 Manufacturer#1 almond antique burnished rose metallic 1 1 2 1173.15 2346.3 2 0 Manufacturer#1 almond antique chartreuse lavender yellow 3 2 3 1753.76 4100.06 34 32 Manufacturer#1 almond antique salmon chartreuse burlywood 4 3 4 1602.59 5702.650000000001 6 -28 Manufacturer#1 almond aquamarine burnished black steel 5 4 5 1414.42 7117.070000000001 28 22 Manufacturer#1 almond aquamarine pink moccasin thistle 6 5 6 1632.66 8749.730000000001 42 14 Manufacturer#2 almond antique violet chocolate turquoise 1 1 1 1690.68 1690.68 14 0 Manufacturer#2 almond antique violet turquoise frosted 2 2 2 1800.7 3491.38 40 26 Manufacturer#2 almond aquamarine midnight light salmon 3 3 3 2031.98 5523.360000000001 2 -38 Manufacturer#2 almond aquamarine rose maroon antique 4 4 4 1698.66 7222.02 25 23 Manufacturer#2 almond aquamarine sandy cyan gainsboro 5 5 5 1701.6 8923.62 18 -7 Manufacturer#3 almond antique chartreuse khaki white 1 1 1 1671.68 1671.68 17 0 Manufacturer#3 almond antique forest lavender goldenrod 2 2 2 1190.27 2861.95 14 -3 Manufacturer#3 almond antique metallic orange dim 3 3 3 1410.39 4272.34 19 5 Manufacturer#3 almond antique misty red olive 4 4 4 1922.98 6195.32 1 -18 Manufacturer#3 almond antique olive coral navajo 5 5 5 1337.29 7532.61 45 44 Manufacturer#4 almond antique gainsboro frosted violet 1 1 1 1620.67 1620.67 10 0 Manufacturer#4 almond antique violet mint lemon 2 2 2 1375.42 2996.09 39 29 Manufacturer#4 almond aquamarine floral ivory bisque 3 3 3 1206.26 4202.35 27 -12 Manufacturer#4 almond aquamarine yellow dodger mint 4 4 4 1844.92 6047.27 7 -20 Manufacturer#4 almond azure aquamarine papaya violet 5 5 5 1290.35 7337.620000000001 12 5 Manufacturer#5 almond antique blue firebrick mint 1 1 1 1789.69 1789.69 31 0 Manufacturer#5 almond antique medium spring khaki 2 2 2 1611.66 3401.3500000000004 6 -25 Manufacturer#5 almond antique sky peru orange 3 3 3 1788.73 5190.08 2 -4 Manufacturer#5 almond aquamarine dodger light gainsboro 4 4 4 1018.1 6208.18 46 44 Manufacturer#5 almond azure blanched chiffon midnight 5 5 5 1464.48 7672.66 23 -23
sbt.ForkMain$ForkError: org.scalatest.exceptions.TestFailedException:
Failed to execute query using catalyst:
Error: execute, tree:
Exchange hashpartitioning(p_mfgr#153883, 5)
+- Scan hive default.part [p_mfgr#153883, p_name#153882, p_retailprice#153888, p_size#153886], HiveTableRelation `default`.`part`, org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, [p_partkey#153881, p_name#153882, p_mfgr#153883, p_brand#153884, p_type#153885, p_size#153886, p_container#153887, p_retailprice#153888, p_comment#153889]
org.apache.spark.sql.catalyst.errors.package$TreeNodeException: execute, tree:
Exchange hashpartitioning(p_mfgr#153883, 5)
+- Scan hive default.part [p_mfgr#153883, p_name#153882, p_retailprice#153888, p_size#153886], HiveTableRelation `default`.`part`, org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, [p_partkey#153881, p_name#153882, p_mfgr#153883, p_brand#153884, p_type#153885, p_size#153886, p_container#153887, p_retailprice#153888, p_comment#153889]
at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:56)
at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.doExecute(ShuffleExchangeExec.scala:119)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:131)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:127)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:127)
at org.apache.spark.sql.execution.InputAdapter.inputRDDs(WholeStageCodegenExec.scala:374)
at org.apache.spark.sql.execution.SortExec.inputRDDs(SortExec.scala:121)
at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:610)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:131)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:127)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:127)
at org.apache.spark.sql.execution.window.WindowExec.doExecute(WindowExec.scala:302)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:131)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:127)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:127)
at org.apache.spark.sql.execution.InputAdapter.inputRDDs(WholeStageCodegenExec.scala:374)
at org.apache.spark.sql.execution.ProjectExec.inputRDDs(basicPhysicalOperators.scala:41)
at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:610)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:131)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:127)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:127)
at org.apache.spark.sql.execution.SparkPlan.getByteArrayRdd(SparkPlan.scala:247)
at org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:294)
at org.apache.spark.sql.execution.SparkPlan.executeCollectPublic(SparkPlan.scala:324)
at org.apache.spark.sql.execution.QueryExecution.hiveResultString(QueryExecution.scala:121)
at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6$$anonfun$25$$anonfun$getResult$1$1.apply(HiveComparisonTest.scala:348)
at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6$$anonfun$25$$anonfun$getResult$1$1.apply(HiveComparisonTest.scala:348)
at org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:78)
at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:125)
at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:73)
at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6$$anonfun$25.getResult$1(HiveComparisonTest.scala:348)
at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6$$anonfun$25.apply(HiveComparisonTest.scala:350)
at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6$$anonfun$25.apply(HiveComparisonTest.scala:345)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
at scala.collection.AbstractTraversable.map(Traversable.scala:104)
at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6.doTest$1(HiveComparisonTest.scala:345)
at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6.apply(HiveComparisonTest.scala:462)
at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85)
at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
at org.scalatest.Transformer.apply(Transformer.scala:22)
at org.scalatest.Transformer.apply(Transformer.scala:20)
at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:186)
at org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:103)
at org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:183)
at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:196)
at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:196)
at org.scalatest.SuperEngine.runTestImpl(Engine.scala:289)
at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:196)
at org.apache.spark.sql.hive.execution.HiveWindowFunctionQuerySuite.org$scalatest$BeforeAndAfter$$super$runTest(HiveWindowFunctionQuerySuite.scala:35)
at org.scalatest.BeforeAndAfter$class.runTest(BeforeAndAfter.scala:203)
at org.apache.spark.sql.hive.execution.HiveWindowFunctionQuerySuite.runTest(HiveWindowFunctionQuerySuite.scala:35)
at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:229)
at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:229)
at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:396)
at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:384)
at scala.collection.immutable.List.foreach(List.scala:392)
at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:384)
at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:379)
at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:461)
at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:229)
at org.scalatest.FunSuite.runTests(FunSuite.scala:1560)
at org.scalatest.Suite$class.run(Suite.scala:1147)
at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1560)
at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:233)
at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:233)
at org.scalatest.SuperEngine.runImpl(Engine.scala:521)
at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:233)
at org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterAll$$super$run(SparkFunSuite.scala:52)
at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:213)
at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:210)
at org.apache.spark.sql.hive.execution.HiveWindowFunctionQuerySuite.org$scalatest$BeforeAndAfter$$super$run(HiveWindowFunctionQuerySuite.scala:35)
at org.scalatest.BeforeAndAfter$class.run(BeforeAndAfter.scala:258)
at org.apache.spark.sql.hive.execution.HiveWindowFunctionQuerySuite.run(HiveWindowFunctionQuerySuite.scala:35)
at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:314)
at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:480)
at sbt.ForkMain$Run$2.call(ForkMain.java:296)
at sbt.ForkMain$Run$2.call(ForkMain.java:286)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.IllegalStateException: Cannot call methods on a stopped SparkContext.
This stopped SparkContext was created at:
org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6.apply(HiveComparisonTest.scala:456)
org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85)
org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
org.scalatest.Transformer.apply(Transformer.scala:22)
org.scalatest.Transformer.apply(Transformer.scala:20)
org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:186)
org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:103)
org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:183)
org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:196)
org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:196)
org.scalatest.SuperEngine.runTestImpl(Engine.scala:289)
org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:196)
org.scalatest.FunSuite.runTest(FunSuite.scala:1560)
org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:229)
org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:229)
org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:396)
org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:384)
scala.collection.immutable.List.foreach(List.scala:392)
org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:384)
org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:379)
The currently active SparkContext was created at:
(No active SparkContext.)
at org.apache.spark.SparkContext.assertNotStopped(SparkContext.scala:100)
at org.apache.spark.SparkContext.broadcast(SparkContext.scala:1492)
at org.apache.spark.sql.hive.HadoopTableReader.<init>(TableReader.scala:89)
at org.apache.spark.sql.hive.execution.HiveTableScanExec.org$apache$spark$sql$hive$execution$HiveTableScanExec$$hadoopReader$lzycompute(HiveTableScanExec.scala:105)
at org.apache.spark.sql.hive.execution.HiveTableScanExec.org$apache$spark$sql$hive$execution$HiveTableScanExec$$hadoopReader(HiveTableScanExec.scala:105)
at org.apache.spark.sql.hive.execution.HiveTableScanExec$$anonfun$10.apply(HiveTableScanExec.scala:188)
at org.apache.spark.sql.hive.execution.HiveTableScanExec$$anonfun$10.apply(HiveTableScanExec.scala:188)
at org.apache.spark.util.Utils$.withDummyCallSite(Utils.scala:2457)
at org.apache.spark.sql.hive.execution.HiveTableScanExec.doExecute(HiveTableScanExec.scala:187)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:131)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:127)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:127)
at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec.prepareShuffleDependency(ShuffleExchangeExec.scala:92)
at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec$$anonfun$doExecute$1.apply(ShuffleExchangeExec.scala:128)
at org.apache.spark.sql.execution.exchange.ShuffleExchangeExec$$anonfun$doExecute$1.apply(ShuffleExchangeExec.scala:119)
at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:52)
... 96 more
select p_mfgr, p_name,
rank() over(distribute by p_mfgr sort by p_name) as r,
dense_rank() over(distribute by p_mfgr sort by p_name) as dr,
count(p_size) over(distribute by p_mfgr sort by p_name) as cd,
p_retailprice, sum(p_retailprice) over (distribute by p_mfgr sort by p_name
rows between unbounded preceding and current row) as s1,
p_size, p_size - lag(p_size,1,p_size) over(distribute by p_mfgr sort by p_name) as deltaSz
from part
== Parsed Logical Plan ==
'Project ['p_mfgr, 'p_name, 'rank() windowspecdefinition('p_mfgr, 'p_name ASC NULLS FIRST, unspecifiedframe$()) AS r#153872, 'dense_rank() windowspecdefinition('p_mfgr, 'p_name ASC NULLS FIRST, unspecifiedframe$()) AS dr#153873, 'count('p_size) windowspecdefinition('p_mfgr, 'p_name ASC NULLS FIRST, unspecifiedframe$()) AS cd#153874, 'p_retailprice, 'sum('p_retailprice) windowspecdefinition('p_mfgr, 'p_name ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS s1#153875, 'p_size, ('p_size - 'lag('p_size, 1, 'p_size) windowspecdefinition('p_mfgr, 'p_name ASC NULLS FIRST, unspecifiedframe$())) AS deltaSz#153876]
+- 'UnresolvedRelation `part`
== Analyzed Logical Plan ==
p_mfgr: string, p_name: string, r: int, dr: int, cd: bigint, p_retailprice: double, s1: double, p_size: int, deltaSz: int
Project [p_mfgr#153883, p_name#153882, r#153872, dr#153873, cd#153874L, p_retailprice#153888, s1#153875, p_size#153886, deltaSz#153876]
+- Project [p_mfgr#153883, p_name#153882, p_retailprice#153888, p_size#153886, r#153872, dr#153873, cd#153874L, s1#153875, _we4#153898, r#153872, dr#153873, cd#153874L, s1#153875, (p_size#153886 - _we4#153898) AS deltaSz#153876]
+- Window [rank(p_name#153882) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS r#153872, dense_rank(p_name#153882) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS dr#153873, count(p_size#153886) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RangeFrame, unboundedpreceding$(), currentrow$())) AS cd#153874L, sum(p_retailprice#153888) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS s1#153875, lag(p_size#153886, 1, p_size#153886) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, -1, -1)) AS _we4#153898], [p_mfgr#153883], [p_name#153882 ASC NULLS FIRST]
+- Project [p_mfgr#153883, p_name#153882, p_retailprice#153888, p_size#153886]
+- SubqueryAlias `default`.`part`
+- HiveTableRelation `default`.`part`, org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, [p_partkey#153881, p_name#153882, p_mfgr#153883, p_brand#153884, p_type#153885, p_size#153886, p_container#153887, p_retailprice#153888, p_comment#153889]
== Optimized Logical Plan ==
Project [p_mfgr#153883, p_name#153882, r#153872, dr#153873, cd#153874L, p_retailprice#153888, s1#153875, p_size#153886, (p_size#153886 - _we4#153898) AS deltaSz#153876]
+- Window [rank(p_name#153882) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS r#153872, dense_rank(p_name#153882) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS dr#153873, count(p_size#153886) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RangeFrame, unboundedpreceding$(), currentrow$())) AS cd#153874L, sum(p_retailprice#153888) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS s1#153875, lag(p_size#153886, 1, p_size#153886) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, -1, -1)) AS _we4#153898], [p_mfgr#153883], [p_name#153882 ASC NULLS FIRST]
+- Project [p_mfgr#153883, p_name#153882, p_retailprice#153888, p_size#153886]
+- HiveTableRelation `default`.`part`, org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, [p_partkey#153881, p_name#153882, p_mfgr#153883, p_brand#153884, p_type#153885, p_size#153886, p_container#153887, p_retailprice#153888, p_comment#153889]
== Physical Plan ==
*(2) Project [p_mfgr#153883, p_name#153882, r#153872, dr#153873, cd#153874L, p_retailprice#153888, s1#153875, p_size#153886, (p_size#153886 - _we4#153898) AS deltaSz#153876]
+- Window [rank(p_name#153882) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS r#153872, dense_rank(p_name#153882) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS dr#153873, count(p_size#153886) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RangeFrame, unboundedpreceding$(), currentrow$())) AS cd#153874L, sum(p_retailprice#153888) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$())) AS s1#153875, lag(p_size#153886, 1, p_size#153886) windowspecdefinition(p_mfgr#153883, p_name#153882 ASC NULLS FIRST, specifiedwindowframe(RowFrame, -1, -1)) AS _we4#153898], [p_mfgr#153883], [p_name#153882 ASC NULLS FIRST]
+- *(1) Sort [p_mfgr#153883 ASC NULLS FIRST, p_name#153882 ASC NULLS FIRST], false, 0
+- Exchange hashpartitioning(p_mfgr#153883, 5)
+- Scan hive default.part [p_mfgr#153883, p_name#153882, p_retailprice#153888, p_size#153886], HiveTableRelation `default`.`part`, org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe, [p_partkey#153881, p_name#153882, p_mfgr#153883, p_brand#153884, p_type#153885, p_size#153886, p_container#153887, p_retailprice#153888, p_comment#153889]
== HIVE - 26 row(s) ==
Manufacturer#1 almond antique burnished rose metallic 1 1 2 1173.15 1173.15 2 0
Manufacturer#1 almond antique burnished rose metallic 1 1 2 1173.15 2346.3 2 0
Manufacturer#1 almond antique chartreuse lavender yellow 3 2 3 1753.76 4100.06 34 32
Manufacturer#1 almond antique salmon chartreuse burlywood 4 3 4 1602.59 5702.650000000001 6 -28
Manufacturer#1 almond aquamarine burnished black steel 5 4 5 1414.42 7117.070000000001 28 22
Manufacturer#1 almond aquamarine pink moccasin thistle 6 5 6 1632.66 8749.730000000001 42 14
Manufacturer#2 almond antique violet chocolate turquoise 1 1 1 1690.68 1690.68 14 0
Manufacturer#2 almond antique violet turquoise frosted 2 2 2 1800.7 3491.38 40 26
Manufacturer#2 almond aquamarine midnight light salmon 3 3 3 2031.98 5523.360000000001 2 -38
Manufacturer#2 almond aquamarine rose maroon antique 4 4 4 1698.66 7222.02 25 23
Manufacturer#2 almond aquamarine sandy cyan gainsboro 5 5 5 1701.6 8923.62 18 -7
Manufacturer#3 almond antique chartreuse khaki white 1 1 1 1671.68 1671.68 17 0
Manufacturer#3 almond antique forest lavender goldenrod 2 2 2 1190.27 2861.95 14 -3
Manufacturer#3 almond antique metallic orange dim 3 3 3 1410.39 4272.34 19 5
Manufacturer#3 almond antique misty red olive 4 4 4 1922.98 6195.32 1 -18
Manufacturer#3 almond antique olive coral navajo 5 5 5 1337.29 7532.61 45 44
Manufacturer#4 almond antique gainsboro frosted violet 1 1 1 1620.67 1620.67 10 0
Manufacturer#4 almond antique violet mint lemon 2 2 2 1375.42 2996.09 39 29
Manufacturer#4 almond aquamarine floral ivory bisque 3 3 3 1206.26 4202.35 27 -12
Manufacturer#4 almond aquamarine yellow dodger mint 4 4 4 1844.92 6047.27 7 -20
Manufacturer#4 almond azure aquamarine papaya violet 5 5 5 1290.35 7337.620000000001 12 5
Manufacturer#5 almond antique blue firebrick mint 1 1 1 1789.69 1789.69 31 0
Manufacturer#5 almond antique medium spring khaki 2 2 2 1611.66 3401.3500000000004 6 -25
Manufacturer#5 almond antique sky peru orange 3 3 3 1788.73 5190.08 2 -4
Manufacturer#5 almond aquamarine dodger light gainsboro 4 4 4 1018.1 6208.18 46 44
Manufacturer#5 almond azure blanched chiffon midnight 5 5 5 1464.48 7672.66 23 -23
at org.scalatest.Assertions$class.newAssertionFailedException(Assertions.scala:528)
at org.scalatest.FunSuite.newAssertionFailedException(FunSuite.scala:1560)
at org.scalatest.Assertions$class.fail(Assertions.scala:1089)
at org.scalatest.FunSuite.fail(FunSuite.scala:1560)
at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6$$anonfun$25.apply(HiveComparisonTest.scala:363)
at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6$$anonfun$25.apply(HiveComparisonTest.scala:345)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
at scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:234)
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
at scala.collection.TraversableLike$class.map(TraversableLike.scala:234)
at scala.collection.AbstractTraversable.map(Traversable.scala:104)
at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6.doTest$1(HiveComparisonTest.scala:345)
at org.apache.spark.sql.hive.execution.HiveComparisonTest$$anonfun$createQueryTest$6.apply(HiveComparisonTest.scala:462)
at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85)
at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
at org.scalatest.Transformer.apply(Transformer.scala:22)
at org.scalatest.Transformer.apply(Transformer.scala:20)
at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:186)
at org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:103)
at org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:183)
at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:196)
at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:196)
at org.scalatest.SuperEngine.runTestImpl(Engine.scala:289)
at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:196)
at org.apache.spark.sql.hive.execution.HiveWindowFunctionQuerySuite.org$scalatest$BeforeAndAfter$$super$runTest(HiveWindowFunctionQuerySuite.scala:35)
at org.scalatest.BeforeAndAfter$class.runTest(BeforeAndAfter.scala:203)
at org.apache.spark.sql.hive.execution.HiveWindowFunctionQuerySuite.runTest(HiveWindowFunctionQuerySuite.scala:35)
at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:229)
at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:229)
at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:396)
at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:384)
at scala.collection.immutable.List.foreach(List.scala:392)
at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:384)
at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:379)
at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:461)
at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:229)
at org.scalatest.FunSuite.runTests(FunSuite.scala:1560)
at org.scalatest.Suite$class.run(Suite.scala:1147)
at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1560)
at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:233)
at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:233)
at org.scalatest.SuperEngine.runImpl(Engine.scala:521)
at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:233)
at org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterAll$$super$run(SparkFunSuite.scala:52)
at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:213)
at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:210)
at org.apache.spark.sql.hive.execution.HiveWindowFunctionQuerySuite.org$scalatest$BeforeAndAfter$$super$run(HiveWindowFunctionQuerySuite.scala:35)
at org.scalatest.BeforeAndAfter$class.run(BeforeAndAfter.scala:258)
at org.apache.spark.sql.hive.execution.HiveWindowFunctionQuerySuite.run(HiveWindowFunctionQuerySuite.scala:35)
at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:314)
at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:480)
at sbt.ForkMain$Run$2.call(ForkMain.java:296)
at sbt.ForkMain$Run$2.call(ForkMain.java:286)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)