Cannot call methods on a stopped SparkContext. This stopped SparkContext was created at: org.apache.spark.sql.hive.orc.OrcHadoopFsRelationSuite.<init>(OrcHadoopFsRelationSuite.scala:26) sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) java.lang.reflect.Constructor.newInstance(Constructor.java:526) java.lang.Class.newInstance(Class.java:379) org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:641) sbt.ForkMain$Run$2.call(ForkMain.java:294) sbt.ForkMain$Run$2.call(ForkMain.java:284) java.util.concurrent.FutureTask.run(FutureTask.java:262) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) java.lang.Thread.run(Thread.java:745) The currently active SparkContext was created at: org.apache.spark.sql.hive.orc.OrcHadoopFsRelationSuite.<init>(OrcHadoopFsRelationSuite.scala:26) sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) java.lang.reflect.Constructor.newInstance(Constructor.java:526) java.lang.Class.newInstance(Class.java:379) org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:641) sbt.ForkMain$Run$2.call(ForkMain.java:294) sbt.ForkMain$Run$2.call(ForkMain.java:284) java.util.concurrent.FutureTask.run(FutureTask.java:262) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) java.lang.Thread.run(Thread.java:745)
sbt.ForkMain$ForkError: Cannot call methods on a stopped SparkContext.
This stopped SparkContext was created at:
org.apache.spark.sql.hive.orc.OrcHadoopFsRelationSuite.<init>(OrcHadoopFsRelationSuite.scala:26)
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
java.lang.reflect.Constructor.newInstance(Constructor.java:526)
java.lang.Class.newInstance(Class.java:379)
org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:641)
sbt.ForkMain$Run$2.call(ForkMain.java:294)
sbt.ForkMain$Run$2.call(ForkMain.java:284)
java.util.concurrent.FutureTask.run(FutureTask.java:262)
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
java.lang.Thread.run(Thread.java:745)
The currently active SparkContext was created at:
org.apache.spark.sql.hive.orc.OrcHadoopFsRelationSuite.<init>(OrcHadoopFsRelationSuite.scala:26)
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
java.lang.reflect.Constructor.newInstance(Constructor.java:526)
java.lang.Class.newInstance(Class.java:379)
org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:641)
sbt.ForkMain$Run$2.call(ForkMain.java:294)
sbt.ForkMain$Run$2.call(ForkMain.java:284)
java.util.concurrent.FutureTask.run(FutureTask.java:262)
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
java.lang.Thread.run(Thread.java:745)
at org.apache.spark.SparkContext.org$apache$spark$SparkContext$$assertNotStopped(SparkContext.scala:106)
at org.apache.spark.SparkContext.defaultParallelism(SparkContext.scala:2086)
at org.apache.spark.SparkContext.makeRDD$default$2(SparkContext.scala:811)
at org.apache.spark.sql.hive.ParquetPartitioningTest$$anonfun$beforeAll$1.apply$mcVI$sp(parquetSuites.scala:703)
at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:141)
at org.apache.spark.sql.hive.ParquetPartitioningTest.beforeAll(parquetSuites.scala:701)
at org.apache.spark.sql.hive.ParquetSourceSuite.beforeAll(parquetSuites.scala:541)
at org.scalatest.BeforeAndAfterAll$class.beforeAll(BeforeAndAfterAll.scala:187)
at org.apache.spark.sql.hive.ParquetPartitioningTest.beforeAll(parquetSuites.scala:688)
at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253)
at org.apache.spark.sql.hive.ParquetPartitioningTest.run(parquetSuites.scala:688)
at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:462)
at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:671)
at sbt.ForkMain$Run$2.call(ForkMain.java:294)
at sbt.ForkMain$Run$2.call(ForkMain.java:284)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)