org.apache.spark.SparkException: Only one SparkContext may be running in this JVM (see SPARK-2243). To ignore this error, set spark.driver.allowMultipleContexts = true. The currently running SparkContext was created at: org.apache.spark.SharedSparkContext$class.initializeContext(SharedSparkContext.scala:41) org.apache.spark.serializer.KryoSerializerSuite.initializeContext(KryoSerializerSuite.scala:36) org.apache.spark.SharedSparkContext$class.beforeAll(SharedSparkContext.scala:48) org.apache.spark.serializer.KryoSerializerSuite.beforeAll(KryoSerializerSuite.scala:36) org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:212) org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:210) org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:52) org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:314) org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:480) sbt.ForkMain$Run$2.call(ForkMain.java:296) sbt.ForkMain$Run$2.call(ForkMain.java:286) java.util.concurrent.FutureTask.run(FutureTask.java:266) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) java.lang.Thread.run(Thread.java:748)
sbt.ForkMain$ForkError: org.apache.spark.SparkException: Only one SparkContext may be running in this JVM (see SPARK-2243). To ignore this error, set spark.driver.allowMultipleContexts = true. The currently running SparkContext was created at:
org.apache.spark.SharedSparkContext$class.initializeContext(SharedSparkContext.scala:41)
org.apache.spark.serializer.KryoSerializerSuite.initializeContext(KryoSerializerSuite.scala:36)
org.apache.spark.SharedSparkContext$class.beforeAll(SharedSparkContext.scala:48)
org.apache.spark.serializer.KryoSerializerSuite.beforeAll(KryoSerializerSuite.scala:36)
org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:212)
org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:210)
org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:52)
org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:314)
org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:480)
sbt.ForkMain$Run$2.call(ForkMain.java:296)
sbt.ForkMain$Run$2.call(ForkMain.java:286)
java.util.concurrent.FutureTask.run(FutureTask.java:266)
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
java.lang.Thread.run(Thread.java:748)
at org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$2.apply(SparkContext.scala:2504)
at org.apache.spark.SparkContext$$anonfun$assertNoOtherContextIsRunning$2.apply(SparkContext.scala:2500)
at scala.Option.foreach(Option.scala:257)
at org.apache.spark.SparkContext$.assertNoOtherContextIsRunning(SparkContext.scala:2500)
at org.apache.spark.SparkContext$.markPartiallyConstructed(SparkContext.scala:2589)
at org.apache.spark.SparkContext.<init>(SparkContext.scala:85)
at org.apache.spark.SparkContext.<init>(SparkContext.scala:127)
at org.apache.spark.input.WholeTextFileRecordReaderSuite.beforeAll(WholeTextFileRecordReaderSuite.scala:53)
at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:212)
at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:210)
at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:52)
at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:314)
at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:480)
at sbt.ForkMain$Run$2.call(ForkMain.java:296)
at sbt.ForkMain$Run$2.call(ForkMain.java:286)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)