org.scalatest.exceptions.TestFailedException: Stream Thread Died: null java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1326) scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:208) scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:218) scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223) org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:201) org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75) org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:92) org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:76) org.apache.spark.sql.execution.streaming.state.StateStoreCoordinatorRef.deactivateInstances(StateStoreCoordinator.scala:108) org.apache.spark.sql.streaming.StreamingQueryManager.notifyQueryTermination(StreamingQueryManager.scala:335) == Progress == AddKafkaData(topics = Set(topic-9), data = WrappedArray(4, 5, 6), message = ) CheckAnswer: [2],[3],[4],[5],[6],[7] StopStream StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@26c48f0f,Map()) CheckAnswer: [2],[3],[4],[5],[6],[7] StopStream AddKafkaData(topics = Set(topic-9), data = WrappedArray(7, 8), message = ) StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@1e78738f,Map()) CheckAnswer: [2],[3],[4],[5],[6],[7],[8],[9] AssertOnQuery(<condition>, Add partitions) AddKafkaData(topics = Set(topic-9), data = WrappedArray(9, 10, 11, 12, 13, 14, 15, 16), message = ) CheckAnswer: [2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17] == Stream == Output Mode: Append Stream state: {KafkaSource[Subscribe[topic-9]]: {"topic-9":{"8":0,"2":2,"5":1,"4":2,"7":1,"1":3,"9":1,"3":2,"6":1,"0":3}}} Thread state: alive Thread stack trace: java.lang.Thread.sleep(Native Method) org.apache.spark.sql.execution.streaming.StreamExecution$$anonfun$org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches$1.apply$mcZ$sp(StreamExecution.scala:328) org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:56) org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches(StreamExecution.scala:296) org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:212) java.lang.InterruptedException at java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1326) at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:208) at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:218) at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223) at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:201) at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75) at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:92) at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:76) at org.apache.spark.sql.execution.streaming.state.StateStoreCoordinatorRef.deactivateInstances(StateStoreCoordinator.scala:108) at org.apache.spark.sql.streaming.StreamingQueryManager.notifyQueryTermination(StreamingQueryManager.scala:335) at org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches(StreamExecution.scala:377) at org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:212) == Sink == 0: [3] [4] [2] 1: [5] 2: [6] [7] 3: [9] [8] 4: 5: [10] 6: [15] [16] [12] [17] [14] [11] [13] == Plan == == Parsed Logical Plan == SerializeFromObject [input[0, int, false] AS value#1017] +- MapElements <function1>, class scala.Tuple2, [StructField(_1,StringType,true), StructField(_2,StringType,true)], obj#1016: int +- DeserializeToObject newInstance(class scala.Tuple2), obj#1015: scala.Tuple2 +- Project [cast(key#1169 as string) AS key#1006, cast(value#1170 as string) AS value#1007] +- LogicalRDD [key#1169, value#1170, topic#1171, partition#1172, offset#1173L, timestamp#1174, timestampType#1175] == Analyzed Logical Plan == value: int SerializeFromObject [input[0, int, false] AS value#1017] +- MapElements <function1>, class scala.Tuple2, [StructField(_1,StringType,true), StructField(_2,StringType,true)], obj#1016: int +- DeserializeToObject newInstance(class scala.Tuple2), obj#1015: scala.Tuple2 +- Project [cast(key#1169 as string) AS key#1006, cast(value#1170 as string) AS value#1007] +- LogicalRDD [key#1169, value#1170, topic#1171, partition#1172, offset#1173L, timestamp#1174, timestampType#1175] == Optimized Logical Plan == SerializeFromObject [input[0, int, false] AS value#1017] +- MapElements <function1>, class scala.Tuple2, [StructField(_1,StringType,true), StructField(_2,StringType,true)], obj#1016: int +- DeserializeToObject newInstance(class scala.Tuple2), obj#1015: scala.Tuple2 +- Project [cast(key#1169 as string) AS key#1006, cast(value#1170 as string) AS value#1007] +- LogicalRDD [key#1169, value#1170, topic#1171, partition#1172, offset#1173L, timestamp#1174, timestampType#1175] == Physical Plan == *SerializeFromObject [input[0, int, false] AS value#1017] +- *MapElements <function1>, obj#1016: int +- *DeserializeToObject newInstance(class scala.Tuple2), obj#1015: scala.Tuple2 +- *Project [cast(key#1169 as string) AS key#1006, cast(value#1170 as string) AS value#1007] +- Scan ExistingRDD[key#1169,value#1170,topic#1171,partition#1172,offset#1173L,timestamp#1174,timestampType#1175]

sbt.ForkMain$ForkError: org.scalatest.exceptions.TestFailedException: 
Stream Thread Died: null
java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1326)
	scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:208)
	scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:218)
	scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:201)
	org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
	org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:92)
	org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:76)
	org.apache.spark.sql.execution.streaming.state.StateStoreCoordinatorRef.deactivateInstances(StateStoreCoordinator.scala:108)
	org.apache.spark.sql.streaming.StreamingQueryManager.notifyQueryTermination(StreamingQueryManager.scala:335)


== Progress ==
   AddKafkaData(topics = Set(topic-9), data = WrappedArray(4, 5, 6), message = )
   CheckAnswer: [2],[3],[4],[5],[6],[7]
   StopStream
   StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@26c48f0f,Map())
   CheckAnswer: [2],[3],[4],[5],[6],[7]
   StopStream
   AddKafkaData(topics = Set(topic-9), data = WrappedArray(7, 8), message = )
   StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@1e78738f,Map())
   CheckAnswer: [2],[3],[4],[5],[6],[7],[8],[9]
   AssertOnQuery(<condition>, Add partitions)
   AddKafkaData(topics = Set(topic-9), data = WrappedArray(9, 10, 11, 12, 13, 14, 15, 16), message = )
   CheckAnswer: [2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17]

== Stream ==
Output Mode: Append
Stream state: {KafkaSource[Subscribe[topic-9]]: {"topic-9":{"8":0,"2":2,"5":1,"4":2,"7":1,"1":3,"9":1,"3":2,"6":1,"0":3}}}
Thread state: alive
Thread stack trace: java.lang.Thread.sleep(Native Method)
org.apache.spark.sql.execution.streaming.StreamExecution$$anonfun$org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches$1.apply$mcZ$sp(StreamExecution.scala:328)
org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:56)
org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches(StreamExecution.scala:296)
org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:212)
java.lang.InterruptedException
	at java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1326)
	at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:208)
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:218)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:201)
	at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
	at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:92)
	at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:76)
	at org.apache.spark.sql.execution.streaming.state.StateStoreCoordinatorRef.deactivateInstances(StateStoreCoordinator.scala:108)
	at org.apache.spark.sql.streaming.StreamingQueryManager.notifyQueryTermination(StreamingQueryManager.scala:335)
	at org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches(StreamExecution.scala:377)
	at org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:212)


== Sink ==
0: [3] [4] [2]
1: [5]
2: [6] [7]
3: [9] [8]
4: 
5: [10]
6: [15] [16] [12] [17] [14] [11] [13]


== Plan ==
== Parsed Logical Plan ==
SerializeFromObject [input[0, int, false] AS value#1017]
+- MapElements <function1>, class scala.Tuple2, [StructField(_1,StringType,true), StructField(_2,StringType,true)], obj#1016: int
   +- DeserializeToObject newInstance(class scala.Tuple2), obj#1015: scala.Tuple2
      +- Project [cast(key#1169 as string) AS key#1006, cast(value#1170 as string) AS value#1007]
         +- LogicalRDD [key#1169, value#1170, topic#1171, partition#1172, offset#1173L, timestamp#1174, timestampType#1175]

== Analyzed Logical Plan ==
value: int
SerializeFromObject [input[0, int, false] AS value#1017]
+- MapElements <function1>, class scala.Tuple2, [StructField(_1,StringType,true), StructField(_2,StringType,true)], obj#1016: int
   +- DeserializeToObject newInstance(class scala.Tuple2), obj#1015: scala.Tuple2
      +- Project [cast(key#1169 as string) AS key#1006, cast(value#1170 as string) AS value#1007]
         +- LogicalRDD [key#1169, value#1170, topic#1171, partition#1172, offset#1173L, timestamp#1174, timestampType#1175]

== Optimized Logical Plan ==
SerializeFromObject [input[0, int, false] AS value#1017]
+- MapElements <function1>, class scala.Tuple2, [StructField(_1,StringType,true), StructField(_2,StringType,true)], obj#1016: int
   +- DeserializeToObject newInstance(class scala.Tuple2), obj#1015: scala.Tuple2
      +- Project [cast(key#1169 as string) AS key#1006, cast(value#1170 as string) AS value#1007]
         +- LogicalRDD [key#1169, value#1170, topic#1171, partition#1172, offset#1173L, timestamp#1174, timestampType#1175]

== Physical Plan ==
*SerializeFromObject [input[0, int, false] AS value#1017]
+- *MapElements <function1>, obj#1016: int
   +- *DeserializeToObject newInstance(class scala.Tuple2), obj#1015: scala.Tuple2
      +- *Project [cast(key#1169 as string) AS key#1006, cast(value#1170 as string) AS value#1007]
         +- Scan ExistingRDD[key#1169,value#1170,topic#1171,partition#1172,offset#1173L,timestamp#1174,timestampType#1175]
         
         
	at org.scalatest.Assertions$class.newAssertionFailedException(Assertions.scala:495)
	at org.scalatest.FunSuite.newAssertionFailedException(FunSuite.scala:1555)
	at org.scalatest.Assertions$class.fail(Assertions.scala:1328)
	at org.scalatest.FunSuite.fail(FunSuite.scala:1555)
	at org.apache.spark.sql.streaming.StreamTest$class.failTest$1(StreamTest.scala:344)
	at org.apache.spark.sql.streaming.StreamTest$class.liftedTree1$1(StreamTest.scala:572)
	at org.apache.spark.sql.streaming.StreamTest$class.testStream(StreamTest.scala:353)
	at org.apache.spark.sql.kafka010.KafkaSourceTest.testStream(KafkaSourceSuite.scala:46)
	at org.apache.spark.sql.kafka010.KafkaSourceSuite.org$apache$spark$sql$kafka010$KafkaSourceSuite$$testFromEarliestOffsets(KafkaSourceSuite.scala:835)
	at org.apache.spark.sql.kafka010.KafkaSourceSuite$$anonfun$24$$anonfun$apply$6.apply$mcV$sp(KafkaSourceSuite.scala:383)
	at org.apache.spark.sql.kafka010.KafkaSourceSuite$$anonfun$24$$anonfun$apply$6.apply(KafkaSourceSuite.scala:381)
	at org.apache.spark.sql.kafka010.KafkaSourceSuite$$anonfun$24$$anonfun$apply$6.apply(KafkaSourceSuite.scala:381)
	at org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22)
	at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85)
	at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
	at org.scalatest.Transformer.apply(Transformer.scala:22)
	at org.scalatest.Transformer.apply(Transformer.scala:20)
	at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:166)
	at org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:68)
	at org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:163)
	at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
	at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
	at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306)
	at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:175)
	at org.apache.spark.sql.kafka010.KafkaSourceTest.org$scalatest$BeforeAndAfterEach$$super$runTest(KafkaSourceSuite.scala:46)
	at org.scalatest.BeforeAndAfterEach$class.runTest(BeforeAndAfterEach.scala:255)
	at org.apache.spark.sql.kafka010.KafkaSourceTest.runTest(KafkaSourceSuite.scala:46)
	at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
	at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
	at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:413)
	at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401)
	at scala.collection.immutable.List.foreach(List.scala:381)
	at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401)
	at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:396)
	at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:483)
	at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:208)
	at org.scalatest.FunSuite.runTests(FunSuite.scala:1555)
	at org.scalatest.Suite$class.run(Suite.scala:1424)
	at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1555)
	at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
	at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
	at org.scalatest.SuperEngine.runImpl(Engine.scala:545)
	at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:212)
	at org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterAll$$super$run(SparkFunSuite.scala:31)
	at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:257)
	at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:256)
	at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:31)
	at org.scalatest.tools.Framework.org$scalatest$tools$Framework$$runSuite(Framework.scala:357)
	at org.scalatest.tools.Framework$ScalaTestTask.execute(Framework.scala:502)
	at sbt.ForkMain$Run$2.call(ForkMain.java:296)
	at sbt.ForkMain$Run$2.call(ForkMain.java:286)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:745)