Cannot call methods on a stopped SparkContext. This stopped SparkContext was created at: org.apache.spark.sql.hive.orc.OrcHadoopFsRelationSuite.<init>(OrcHadoopFsRelationSuite.scala:26) sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) java.lang.reflect.Constructor.newInstance(Constructor.java:526) java.lang.Class.newInstance(Class.java:379) org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:641) sbt.ForkMain$Run$2.call(ForkMain.java:294) sbt.ForkMain$Run$2.call(ForkMain.java:284) java.util.concurrent.FutureTask.run(FutureTask.java:262) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) java.lang.Thread.run(Thread.java:745) The currently active SparkContext was created at: org.apache.spark.sql.hive.orc.OrcHadoopFsRelationSuite.<init>(OrcHadoopFsRelationSuite.scala:26) sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) java.lang.reflect.Constructor.newInstance(Constructor.java:526) java.lang.Class.newInstance(Class.java:379) org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:641) sbt.ForkMain$Run$2.call(ForkMain.java:294) sbt.ForkMain$Run$2.call(ForkMain.java:284) java.util.concurrent.FutureTask.run(FutureTask.java:262) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) java.lang.Thread.run(Thread.java:745)

sbt.ForkMain$ForkError: Cannot call methods on a stopped SparkContext.
This stopped SparkContext was created at:

org.apache.spark.sql.hive.orc.OrcHadoopFsRelationSuite.<init>(OrcHadoopFsRelationSuite.scala:26)
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
java.lang.reflect.Constructor.newInstance(Constructor.java:526)
java.lang.Class.newInstance(Class.java:379)
org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:641)
sbt.ForkMain$Run$2.call(ForkMain.java:294)
sbt.ForkMain$Run$2.call(ForkMain.java:284)
java.util.concurrent.FutureTask.run(FutureTask.java:262)
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
java.lang.Thread.run(Thread.java:745)

The currently active SparkContext was created at:

org.apache.spark.sql.hive.orc.OrcHadoopFsRelationSuite.<init>(OrcHadoopFsRelationSuite.scala:26)
sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
java.lang.reflect.Constructor.newInstance(Constructor.java:526)
java.lang.Class.newInstance(Class.java:379)
org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:641)
sbt.ForkMain$Run$2.call(ForkMain.java:294)
sbt.ForkMain$Run$2.call(ForkMain.java:284)
java.util.concurrent.FutureTask.run(FutureTask.java:262)
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
java.lang.Thread.run(Thread.java:745)
         
	at org.apache.spark.SparkContext.org$apache$spark$SparkContext$$assertNotStopped(SparkContext.scala:106)
	at org.apache.spark.SparkContext.defaultParallelism(SparkContext.scala:2086)
	at org.apache.spark.SparkContext.makeRDD$default$2(SparkContext.scala:811)
	at org.apache.spark.sql.hive.ParquetPartitioningTest$$anonfun$beforeAll$1.apply$mcVI$sp(parquetSuites.scala:703)
	at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:141)
	at org.apache.spark.sql.hive.ParquetPartitioningTest.beforeAll(parquetSuites.scala:701)
	at org.apache.spark.sql.hive.ParquetSourceSuite.beforeAll(parquetSuites.scala:541)
	at org.scalatest.BeforeAndAfterAll$class.beforeAll(BeforeAndAfterAll.scala:187)
	at org.apache.spark.sql.hive.ParquetPartitioningTest.beforeAll(parquetSuites.scala:688)
	at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:253)
	at org.apache.spark.sql.hive.ParquetPartitioningTest.run(parquetSuites.scala:688)
	at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:462)
	at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:671)
	at sbt.ForkMain$Run$2.call(ForkMain.java:294)
	at sbt.ForkMain$Run$2.call(ForkMain.java:284)
	at java.util.concurrent.FutureTask.run(FutureTask.java:262)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
	at java.lang.Thread.run(Thread.java:745)