&#010;Stream Thread Died: null&#010;java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1326)&#010; scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:208)&#010; scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:218)&#010; scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)&#010; org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:201)&#010; org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)&#010; org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:92)&#010; org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:76)&#010; org.apache.spark.sql.execution.streaming.state.StateStoreCoordinatorRef.deactivateInstances(StateStoreCoordinator.scala:108)&#010; org.apache.spark.sql.streaming.StreamingQueryManager.notifyQueryTermination(StreamingQueryManager.scala:335)&#010;&#010;&#010;== Progress ==&#010; AssertOnQuery(<condition>, )&#010; AddKafkaData(topics = Set(stress4, stress6, stress2, stress1, stress5, stress3), data = Range(0, 1), message = Add topic stress7)&#010; CheckAnswer: [1],[2]&#010; StopStream&#010; StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@27de9336,Map())&#010; CheckAnswer: [1],[2]&#010; StopStream&#010; AddKafkaData(topics = Set(stress4, stress6, stress2, stress8, stress1, stress5, stress3), data = Range(2, 3), message = Add topic stress9)&#010; AddKafkaData(topics = Set(stress4, stress6, stress2, stress8, stress1, stress5, stress3), data = Range(4, 5, 6, 7), message = Add partition)&#010; AddKafkaData(topics = Set(stress4, stress6, stress2, stress8, stress1, stress5, stress3, stress10), data = Range(8, 9, 10, 11, 12, 13), message = Add topic stress11)&#010; StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@4c2812aa,Map())&#010; AddKafkaData(topics = Set(stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10), data = Range(14, 15, 16), message = Add topic stress13)&#010; CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17]&#010; CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17]&#010; StopStream&#010; StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@53d94a6b,Map())&#010; CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17]&#010; StopStream&#010; AddKafkaData(topics = Set(stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10), data = Range(), message = Add partition)&#010; AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10), data = Range(17, 18, 19, 20, 21), message = Add topic stress15)&#010; StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@75051d05,Map())&#010; AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10), data = Range(22, 23, 24, 25, 26, 27, 28, 29, 30), message = )&#010; CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31]&#010; StopStream&#010; StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@468dbd07,Map())&#010; CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31]&#010; StopStream&#010; AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10), data = Range(31), message = )&#010; StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@7bd6c35e,Map())&#010; AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10), data = Range(32, 33, 34), message = Add partition)&#010; CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35]&#010; AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10, stress16), data = Range(35), message = Add topic stress17)&#010; CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36]&#010; StopStream&#010; StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@1131fcfd,Map())&#010; CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36]&#010; StopStream&#010; AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10, stress16), data = Range(36, 37, 38, 39, 40, 41, 42, 43), message = )&#010; StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@42d4de20,Map())&#010; AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress18, stress3, stress10, stress16), data = Range(44, 45, 46, 47, 48, 49, 50), message = Add topic stress19)&#010; AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress3, stress10, stress16), data = Range(51, 52, 53, 54, 55, 56, 57), message = Add topic stress21)&#010; CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36],[37],[38],[39],[40],[41],[42],[43],[44],[45],[46],[47],[48],[49],[50],[51],[52],[53],[54],[55],[56],[57],[58]&#010; StopStream&#010; StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@32fb4a87,Map())&#010; CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36],[37],[38],[39],[40],[41],[42],[43],[44],[45],[46],[47],[48],[49],[50],[51],[52],[53],[54],[55],[56],[57],[58]&#010; StopStream&#010; AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress3, stress10, stress16), data = Range(58, 59, 60, 61, 62, 63, 64), message = Add partition)&#010; AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress3, stress10, stress16), data = Range(65, 66, 67), message = )&#010; AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress3, stress10, stress16), data = Range(68, 69, 70, 71, 72, 73, 74, 75, 76), message = )&#010; StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@3d837af7,Map())&#010; AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress3, stress10, stress16), data = Range(), message = )&#010; AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress22, stress3, stress10, stress16), data = Range(77, 78, 79), message = Add topic stress23)&#010; CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36],[37],[38],[39],[40],[41],[42],[43],[44],[45],[46],[47],[48],[49],[50],[51],[52],[53],[54],[55],[56],[57],[58],[59],[60],[61],[62],[63],[64],[65],[66],[67],[68],[69],[70],[71],[72],[73],[74],[75],[76],[77],[78],[79],[80]&#010; StopStream&#010; StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@36e8fefe,Map())&#010; AddKafkaData(topics = Set(stress14, stress24, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress22, stress3, stress10, stress16), data = Range(80, 81, 82, 83, 84, 85, 86, 87, 88), message = Add topic stress25)&#010; AddKafkaData(topics = Set(stress14, stress24, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress26, stress22, stress3, stress10, stress16), data = Range(89), message = Add topic stress27)&#010; AddKafkaData(topics = Set(stress14, stress24, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress26, stress22, stress3, stress10, stress16), data = Range(90, 91, 92, 93, 94), message = Add partition)&#010; AddKafkaData(topics = Set(stress14, stress24, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress26, stress22, stress3, stress10, stress16), data = Range(95, 96, 97, 98, 99, 100), message = )&#010; CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36],[37],[38],[39],[40],[41],[42],[43],[44],[45],[46],[47],[48],[49],[50],[51],[52],[53],[54],[55],[56],[57],[58],[59],[60],[61],[62],[63],[64],[65],[66],[67],[68],[69],[70],[71],[72],[73],[74],[75],[76],[77],[78],[79],[80],[81],[82],[83],[84],[85],[86],[87],[88],[89],[90],[91],[92],[93],[94],[95],[96],[97],[98],[99],[100],[101]&#010; AddKafkaData(topics = Set(stress14, stress24, stress4, stress28, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress26, stress22, stress3, stress10, stress16), data = Range(101, 102, 103, 104, 105, 106, 107, 108), message = Add topic stress29)&#010; CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36],[37],[38],[39],[40],[41],[42],[43],[44],[45],[46],[47],[48],[49],[50],[51],[52],[53],[54],[55],[56],[57],[58],[59],[60],[61],[62],[63],[64],[65],[66],[67],[68],[69],[70],[71],[72],[73],[74],[75],[76],[77],[78],[79],[80],[81],[82],[83],[84],[85],[86],[87],[88],[89],[90],[91],[92],[93],[94],[95],[96],[97],[98],[99],[100],[101],[102],[103],[104],[105],[106],[107],[108],[109]&#010; StopStream&#010; StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@30ffc3c0,Map())&#010; CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36],[37],[38],[39],[40],[41],[42],[43],[44],[45],[46],[47],[48],[49],[50],[51],[52],[53],[54],[55],[56],[57],[58],[59],[60],[61],[62],[63],[64],[65],[66],[67],[68],[69],[70],[71],[72],[73],[74],[75],[76],[77],[78],[79],[80],[81],[82],[83],[84],[85],[86],[87],[88],[89],[90],[91],[92],[93],[94],[95],[96],[97],[98],[99],[100],[101],[102],[103],[104],[105],[106],[107],[108],[109]&#010;&#010;== Stream ==&#010;Output Mode: Append&#010;Stream state: {KafkaSource[SubscribePattern[stress.*]]: {"stress5":{"23":0,"8":0,"17":0,"11":0,"2":2,"20":0,"5":0,"14":0,"4":0,"13":0,"22":0,"7":0,"16":0,"10":0,"1":1,"19":0,"9":0,"18":0,"12":0,"3":0,"21":0,"15":0,"6":1,"24":0,"0":2},"stress8":{"17":0,"8":0,"11":0,"20":0,"2":0,"5":0,"14":0,"13":0,"4":0,"16":0,"7":1,"1":1,"10":1,"19":0,"18":0,"9":1,"21":0,"3":1,"12":1,"15":0,"6":1,"0":0},"stress16":{"8":0,"11":0,"2":0,"5":0,"13":0,"4":0,"7":0,"1":0,"10":0,"9":0,"3":0,"12":0,"6":0,"0":0},"stress22":{"2":2,"4":1,"1":1,"3":1,"0":1},"stress10":{"8":0,"11":0,"2":1,"5":2,"14":1,"13":1,"4":2,"7":0,"1":1,"10":0,"9":0,"3":1,"12":0,"6":1,"0":3},"stress1":{"8":0,"11":0,"2":2,"5":1,"14":1,"13":1,"4":1,"7":1,"1":3,"10":1,"9":1,"3":1,"12":1,"15":0,"6":0,"0":1},"stress28":{"2":0,"1":0,"0":0},"stress4":{"17":0,"8":0,"11":1,"2":2,"5":2,"14":0,"13":0,"4":1,"16":0,"7":0,"1":2,"10":0,"18":0,"9":0,"3":2,"12":0,"15":0,"6":0,"0":0},"stress3":{"8":1,"11":1,"2":2,"5":1,"14":0,"13":0,"4":1,"16":1,"7":1,"1":2,"10":1,"9":1,"3":3,"12":2,"15":1,"6":1,"0":3},"stress18":{"8":0,"2":0,"5":1,"4":1,"7":0,"1":1,"3":0,"6":0,"0":0},"stress6":{"8":1,"11":0,"2":2,"5":2,"14":0,"13":0,"4":2,"7":1,"1":2,"10":0,"9":0,"3":2,"12":0,"6":2,"0":2},"stress24":{"8":0,"2":0,"5":0,"4":0,"7":0,"1":0,"3":0,"6":0,"0":0},"stress12":{"17":0,"8":2,"11":0,"2":1,"5":1,"14":0,"13":0,"4":1,"16":0,"7":3,"1":2,"10":2,"9":2,"3":2,"12":1,"15":0,"6":3,"0":2},"stress26":{"2":0,"4":0,"1":0,"3":0,"0":0},"stress14":{"17":0,"8":1,"11":1,"2":1,"5":0,"14":0,"13":0,"4":0,"16":0,"7":0,"1":0,"10":0,"9":0,"3":0,"12":0,"15":0,"6":0,"0":0},"stress2":{"17":0,"8":1,"11":0,"20":0,"2":3,"5":0,"14":0,"13":0,"4":1,"16":0,"7":0,"1":1,"10":0,"19":0,"18":0,"9":0,"21":0,"3":1,"12":0,"15":0,"6":1,"0":3},"stress20":{"8":0,"11":0,"2":0,"5":0,"4":0,"7":0,"10":0,"1":0,"9":0,"3":0,"6":0,"0":0}}}&#010;Thread state: alive&#010;Thread stack trace: java.lang.Thread.sleep(Native Method)&#010;org.apache.spark.sql.execution.streaming.StreamExecution$$anonfun$org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches$1.apply$mcZ$sp(StreamExecution.scala:326)&#010;org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:56)&#010;org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches(StreamExecution.scala:294)&#010;org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:211)&#010;java.lang.InterruptedException&#010; at java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1326)&#010; at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:208)&#010; at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:218)&#010; at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)&#010; at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:201)&#010; at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)&#010; at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:92)&#010; at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:76)&#010; at org.apache.spark.sql.execution.streaming.state.StateStoreCoordinatorRef.deactivateInstances(StateStoreCoordinator.scala:108)&#010; at org.apache.spark.sql.streaming.StreamingQueryManager.notifyQueryTermination(StreamingQueryManager.scala:335)&#010; at org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches(StreamExecution.scala:375)&#010; at org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:211)&#010;&#010;&#010;== Sink ==&#010;0: &#010;1: [2] [1]&#010;2: [12] [3] [6] [7] [13] [14] [10] [9] [5] [11] [4] [8]&#010;3: &#010;4: [15] [16] [17]&#010;5: [20] [18] [21] [22] [19]&#010;6: [23]&#010;7: [26] [25] [27] [28] [31] [24] [29] [30]&#010;8: [32]&#010;9: &#010;10: &#010;11: &#010;12: &#010;13: [33] [35] [34]&#010;14: [36]&#010;15: [41] [38] [37] [43] [44] [39] [40] [42]&#010;16: [46] [45]&#010;17: [48] [51] [49] [50] [47]&#010;18: [52] [53]&#010;19: [54] [58] [57] [56] [55]&#010;20: [71] [67] [64] [77] [61] [73] [59] [75] [70] [63] [72] [74] [60] [69] [76] [62] [68] [66] [65]&#010;21: [78]&#010;22: [79] [80]&#010;23: [82] [81]&#010;24: [88] [83] [89] [85] [86] [84] [87]&#010;25: [90]&#010;26: &#010;27: &#010;28: [92] [94] [93] [91] [95]&#010;29: [97]&#010;30: [96] [101] [98] [99] [100]&#010;31: [102] [103]&#010;32: [105] [106] [107] [109] [104] [108]&#010;&#010;&#010;== Plan ==&#010;null&#010; &#010;


      org.scalatest.exceptions.TestFailedException: 
Stream Thread Died: null
java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1326)
	scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:208)
	scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:218)
	scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:201)
	org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
	org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:92)
	org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:76)
	org.apache.spark.sql.execution.streaming.state.StateStoreCoordinatorRef.deactivateInstances(StateStoreCoordinator.scala:108)
	org.apache.spark.sql.streaming.StreamingQueryManager.notifyQueryTermination(StreamingQueryManager.scala:335)


== Progress ==
   AssertOnQuery(<condition>, )
   AddKafkaData(topics = Set(stress4, stress6, stress2, stress1, stress5, stress3), data = Range(0, 1), message = Add topic stress7)
   CheckAnswer: [1],[2]
   StopStream
   StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@27de9336,Map())
   CheckAnswer: [1],[2]
   StopStream
   AddKafkaData(topics = Set(stress4, stress6, stress2, stress8, stress1, stress5, stress3), data = Range(2, 3), message = Add topic stress9)
   AddKafkaData(topics = Set(stress4, stress6, stress2, stress8, stress1, stress5, stress3), data = Range(4, 5, 6, 7), message = Add partition)
   AddKafkaData(topics = Set(stress4, stress6, stress2, stress8, stress1, stress5, stress3, stress10), data = Range(8, 9, 10, 11, 12, 13), message = Add topic stress11)
   StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@4c2812aa,Map())
   AddKafkaData(topics = Set(stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10), data = Range(14, 15, 16), message = Add topic stress13)
   CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17]
   CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17]
   StopStream
   StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@53d94a6b,Map())
   CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17]
   StopStream
   AddKafkaData(topics = Set(stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10), data = Range(), message = Add partition)
   AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10), data = Range(17, 18, 19, 20, 21), message = Add topic stress15)
   StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@75051d05,Map())
   AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10), data = Range(22, 23, 24, 25, 26, 27, 28, 29, 30), message = )
   CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31]
   StopStream
   StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@468dbd07,Map())
   CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31]
   StopStream
   AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10), data = Range(31), message = )
   StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@7bd6c35e,Map())
   AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10), data = Range(32, 33, 34), message = Add partition)
   CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35]
   AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10, stress16), data = Range(35), message = Add topic stress17)
   CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36]
   StopStream
   StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@1131fcfd,Map())
   CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36]
   StopStream
   AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress3, stress10, stress16), data = Range(36, 37, 38, 39, 40, 41, 42, 43), message = )
   StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@42d4de20,Map())
   AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress8, stress1, stress5, stress18, stress3, stress10, stress16), data = Range(44, 45, 46, 47, 48, 49, 50), message = Add topic stress19)
   AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress3, stress10, stress16), data = Range(51, 52, 53, 54, 55, 56, 57), message = Add topic stress21)
   CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36],[37],[38],[39],[40],[41],[42],[43],[44],[45],[46],[47],[48],[49],[50],[51],[52],[53],[54],[55],[56],[57],[58]
   StopStream
   StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@32fb4a87,Map())
   CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36],[37],[38],[39],[40],[41],[42],[43],[44],[45],[46],[47],[48],[49],[50],[51],[52],[53],[54],[55],[56],[57],[58]
   StopStream
   AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress3, stress10, stress16), data = Range(58, 59, 60, 61, 62, 63, 64), message = Add partition)
   AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress3, stress10, stress16), data = Range(65, 66, 67), message = )
   AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress3, stress10, stress16), data = Range(68, 69, 70, 71, 72, 73, 74, 75, 76), message = )
   StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@3d837af7,Map())
   AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress3, stress10, stress16), data = Range(), message = )
   AddKafkaData(topics = Set(stress14, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress22, stress3, stress10, stress16), data = Range(77, 78, 79), message = Add topic stress23)
   CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36],[37],[38],[39],[40],[41],[42],[43],[44],[45],[46],[47],[48],[49],[50],[51],[52],[53],[54],[55],[56],[57],[58],[59],[60],[61],[62],[63],[64],[65],[66],[67],[68],[69],[70],[71],[72],[73],[74],[75],[76],[77],[78],[79],[80]
   StopStream
   StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@36e8fefe,Map())
   AddKafkaData(topics = Set(stress14, stress24, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress22, stress3, stress10, stress16), data = Range(80, 81, 82, 83, 84, 85, 86, 87, 88), message = Add topic stress25)
   AddKafkaData(topics = Set(stress14, stress24, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress26, stress22, stress3, stress10, stress16), data = Range(89), message = Add topic stress27)
   AddKafkaData(topics = Set(stress14, stress24, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress26, stress22, stress3, stress10, stress16), data = Range(90, 91, 92, 93, 94), message = Add partition)
   AddKafkaData(topics = Set(stress14, stress24, stress4, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress26, stress22, stress3, stress10, stress16), data = Range(95, 96, 97, 98, 99, 100), message = )
   CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36],[37],[38],[39],[40],[41],[42],[43],[44],[45],[46],[47],[48],[49],[50],[51],[52],[53],[54],[55],[56],[57],[58],[59],[60],[61],[62],[63],[64],[65],[66],[67],[68],[69],[70],[71],[72],[73],[74],[75],[76],[77],[78],[79],[80],[81],[82],[83],[84],[85],[86],[87],[88],[89],[90],[91],[92],[93],[94],[95],[96],[97],[98],[99],[100],[101]
   AddKafkaData(topics = Set(stress14, stress24, stress4, stress28, stress6, stress12, stress2, stress20, stress8, stress1, stress5, stress18, stress26, stress22, stress3, stress10, stress16), data = Range(101, 102, 103, 104, 105, 106, 107, 108), message = Add topic stress29)
   CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36],[37],[38],[39],[40],[41],[42],[43],[44],[45],[46],[47],[48],[49],[50],[51],[52],[53],[54],[55],[56],[57],[58],[59],[60],[61],[62],[63],[64],[65],[66],[67],[68],[69],[70],[71],[72],[73],[74],[75],[76],[77],[78],[79],[80],[81],[82],[83],[84],[85],[86],[87],[88],[89],[90],[91],[92],[93],[94],[95],[96],[97],[98],[99],[100],[101],[102],[103],[104],[105],[106],[107],[108],[109]
   StopStream
   StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@30ffc3c0,Map())
   CheckAnswer: [1],[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23],[24],[25],[26],[27],[28],[29],[30],[31],[32],[33],[34],[35],[36],[37],[38],[39],[40],[41],[42],[43],[44],[45],[46],[47],[48],[49],[50],[51],[52],[53],[54],[55],[56],[57],[58],[59],[60],[61],[62],[63],[64],[65],[66],[67],[68],[69],[70],[71],[72],[73],[74],[75],[76],[77],[78],[79],[80],[81],[82],[83],[84],[85],[86],[87],[88],[89],[90],[91],[92],[93],[94],[95],[96],[97],[98],[99],[100],[101],[102],[103],[104],[105],[106],[107],[108],[109]

== Stream ==
Output Mode: Append
Stream state: {KafkaSource[SubscribePattern[stress.*]]: {"stress5":{"23":0,"8":0,"17":0,"11":0,"2":2,"20":0,"5":0,"14":0,"4":0,"13":0,"22":0,"7":0,"16":0,"10":0,"1":1,"19":0,"9":0,"18":0,"12":0,"3":0,"21":0,"15":0,"6":1,"24":0,"0":2},"stress8":{"17":0,"8":0,"11":0,"20":0,"2":0,"5":0,"14":0,"13":0,"4":0,"16":0,"7":1,"1":1,"10":1,"19":0,"18":0,"9":1,"21":0,"3":1,"12":1,"15":0,"6":1,"0":0},"stress16":{"8":0,"11":0,"2":0,"5":0,"13":0,"4":0,"7":0,"1":0,"10":0,"9":0,"3":0,"12":0,"6":0,"0":0},"stress22":{"2":2,"4":1,"1":1,"3":1,"0":1},"stress10":{"8":0,"11":0,"2":1,"5":2,"14":1,"13":1,"4":2,"7":0,"1":1,"10":0,"9":0,"3":1,"12":0,"6":1,"0":3},"stress1":{"8":0,"11":0,"2":2,"5":1,"14":1,"13":1,"4":1,"7":1,"1":3,"10":1,"9":1,"3":1,"12":1,"15":0,"6":0,"0":1},"stress28":{"2":0,"1":0,"0":0},"stress4":{"17":0,"8":0,"11":1,"2":2,"5":2,"14":0,"13":0,"4":1,"16":0,"7":0,"1":2,"10":0,"18":0,"9":0,"3":2,"12":0,"15":0,"6":0,"0":0},"stress3":{"8":1,"11":1,"2":2,"5":1,"14":0,"13":0,"4":1,"16":1,"7":1,"1":2,"10":1,"9":1,"3":3,"12":2,"15":1,"6":1,"0":3},"stress18":{"8":0,"2":0,"5":1,"4":1,"7":0,"1":1,"3":0,"6":0,"0":0},"stress6":{"8":1,"11":0,"2":2,"5":2,"14":0,"13":0,"4":2,"7":1,"1":2,"10":0,"9":0,"3":2,"12":0,"6":2,"0":2},"stress24":{"8":0,"2":0,"5":0,"4":0,"7":0,"1":0,"3":0,"6":0,"0":0},"stress12":{"17":0,"8":2,"11":0,"2":1,"5":1,"14":0,"13":0,"4":1,"16":0,"7":3,"1":2,"10":2,"9":2,"3":2,"12":1,"15":0,"6":3,"0":2},"stress26":{"2":0,"4":0,"1":0,"3":0,"0":0},"stress14":{"17":0,"8":1,"11":1,"2":1,"5":0,"14":0,"13":0,"4":0,"16":0,"7":0,"1":0,"10":0,"9":0,"3":0,"12":0,"15":0,"6":0,"0":0},"stress2":{"17":0,"8":1,"11":0,"20":0,"2":3,"5":0,"14":0,"13":0,"4":1,"16":0,"7":0,"1":1,"10":0,"19":0,"18":0,"9":0,"21":0,"3":1,"12":0,"15":0,"6":1,"0":3},"stress20":{"8":0,"11":0,"2":0,"5":0,"4":0,"7":0,"10":0,"1":0,"9":0,"3":0,"6":0,"0":0}}}
Thread state: alive
Thread stack trace: java.lang.Thread.sleep(Native Method)
org.apache.spark.sql.execution.streaming.StreamExecution$$anonfun$org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches$1.apply$mcZ$sp(StreamExecution.scala:326)
org.apache.spark.sql.execution.streaming.ProcessingTimeExecutor.execute(TriggerExecutor.scala:56)
org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches(StreamExecution.scala:294)
org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:211)
java.lang.InterruptedException
	at java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1326)
	at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:208)
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:218)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:201)
	at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
	at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:92)
	at org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:76)
	at org.apache.spark.sql.execution.streaming.state.StateStoreCoordinatorRef.deactivateInstances(StateStoreCoordinator.scala:108)
	at org.apache.spark.sql.streaming.StreamingQueryManager.notifyQueryTermination(StreamingQueryManager.scala:335)
	at org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches(StreamExecution.scala:375)
	at org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:211)


== Sink ==
0: 
1: [2] [1]
2: [12] [3] [6] [7] [13] [14] [10] [9] [5] [11] [4] [8]
3: 
4: [15] [16] [17]
5: [20] [18] [21] [22] [19]
6: [23]
7: [26] [25] [27] [28] [31] [24] [29] [30]
8: [32]
9: 
10: 
11: 
12: 
13: [33] [35] [34]
14: [36]
15: [41] [38] [37] [43] [44] [39] [40] [42]
16: [46] [45]
17: [48] [51] [49] [50] [47]
18: [52] [53]
19: [54] [58] [57] [56] [55]
20: [71] [67] [64] [77] [61] [73] [59] [75] [70] [63] [72] [74] [60] [69] [76] [62] [68] [66] [65]
21: [78]
22: [79] [80]
23: [82] [81]
24: [88] [83] [89] [85] [86] [84] [87]
25: [90]
26: 
27: 
28: [92] [94] [93] [91] [95]
29: [97]
30: [96] [101] [98] [99] [100]
31: [102] [103]
32: [105] [106] [107] [109] [104] [108]


== Plan ==
null
         
         
      at org.scalatest.Assertions$class.newAssertionFailedException(Assertions.scala:495)
      at org.scalatest.FunSuite.newAssertionFailedException(FunSuite.scala:1555)
      at org.scalatest.Assertions$class.fail(Assertions.scala:1328)
      at org.scalatest.FunSuite.fail(FunSuite.scala:1555)
      at org.apache.spark.sql.streaming.StreamTest$class.failTest$1(StreamTest.scala:344)
      at org.apache.spark.sql.streaming.StreamTest$class.liftedTree1$1(StreamTest.scala:572)
      at org.apache.spark.sql.streaming.StreamTest$class.testStream(StreamTest.scala:353)
      at org.apache.spark.sql.kafka010.KafkaSourceTest.testStream(KafkaSourceSuite.scala:46)
      at org.apache.spark.sql.streaming.StreamTest$class.runStressTest(StreamTest.scala:666)
      at org.apache.spark.sql.kafka010.KafkaSourceTest.runStressTest(KafkaSourceSuite.scala:46)
      at org.apache.spark.sql.kafka010.KafkaSourceStressSuite$$anonfun$16.apply$mcV$sp(KafkaSourceSuite.scala:897)
      at org.apache.spark.sql.kafka010.KafkaSourceStressSuite$$anonfun$16.apply(KafkaSourceSuite.scala:877)
      at org.apache.spark.sql.kafka010.KafkaSourceStressSuite$$anonfun$16.apply(KafkaSourceSuite.scala:877)
      at org.scalatest.Transformer$$anonfun$apply$1.apply$mcV$sp(Transformer.scala:22)
      at org.scalatest.OutcomeOf$class.outcomeOf(OutcomeOf.scala:85)
      at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
      at org.scalatest.Transformer.apply(Transformer.scala:22)
      at org.scalatest.Transformer.apply(Transformer.scala:20)
      at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:166)
      at org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:68)
      at org.scalatest.FunSuiteLike$class.invokeWithFixture$1(FunSuiteLike.scala:163)
      at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
      at org.scalatest.FunSuiteLike$$anonfun$runTest$1.apply(FunSuiteLike.scala:175)
      at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306)
      at org.scalatest.FunSuiteLike$class.runTest(FunSuiteLike.scala:175)
      at org.apache.spark.sql.kafka010.KafkaSourceTest.org$scalatest$BeforeAndAfterEach$$super$runTest(KafkaSourceSuite.scala:46)
      at org.scalatest.BeforeAndAfterEach$class.runTest(BeforeAndAfterEach.scala:255)
      at org.apache.spark.sql.kafka010.KafkaSourceTest.runTest(KafkaSourceSuite.scala:46)
      at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
      at org.scalatest.FunSuiteLike$$anonfun$runTests$1.apply(FunSuiteLike.scala:208)
      at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:413)
      at org.scalatest.SuperEngine$$anonfun$traverseSubNodes$1$1.apply(Engine.scala:401)
      at scala.collection.immutable.List.foreach(List.scala:381)
      at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401)
      at org.scalatest.SuperEngine.org$scalatest$SuperEngine$$runTestsInBranch(Engine.scala:396)
      at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:483)
      at org.scalatest.FunSuiteLike$class.runTests(FunSuiteLike.scala:208)
      at org.scalatest.FunSuite.runTests(FunSuite.scala:1555)
      at org.scalatest.Suite$class.run(Suite.scala:1424)
      at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1555)
      at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
      at org.scalatest.FunSuiteLike$$anonfun$run$1.apply(FunSuiteLike.scala:212)
      at org.scalatest.SuperEngine.runImpl(Engine.scala:545)
      at org.scalatest.FunSuiteLike$class.run(FunSuiteLike.scala:212)
      at org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterAll$$super$run(SparkFunSuite.scala:31)
      at org.scalatest.BeforeAndAfterAll$class.liftedTree1$1(BeforeAndAfterAll.scala:257)
      at org.scalatest.BeforeAndAfterAll$class.run(BeforeAndAfterAll.scala:256)
      at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:31)
      at org.scalatest.Suite$class.callExecuteOnSuite$1(Suite.scala:1492)
      at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1528)
      at org.scalatest.Suite$$anonfun$runNestedSuites$1.apply(Suite.scala:1526)
      at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
      at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186)
      at org.scalatest.Suite$class.runNestedSuites(Suite.scala:1526)
      at org.scalatest.tools.DiscoverySuite.runNestedSuites(DiscoverySuite.scala:29)
      at org.scalatest.Suite$class.run(Suite.scala:1421)
      at org.scalatest.tools.DiscoverySuite.run(DiscoverySuite.scala:29)
      at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:55)
      at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2563)
      at org.scalatest.tools.Runner$$anonfun$doRunRunRunDaDoRunRun$3.apply(Runner.scala:2557)
      at scala.collection.immutable.List.foreach(List.scala:381)
      at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:2557)
      at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1044)
      at org.scalatest.tools.Runner$$anonfun$runOptionallyWithPassFailReporter$2.apply(Runner.scala:1043)
      at org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:2722)
      at org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1043)
      at org.scalatest.tools.Runner$.main(Runner.scala:860)
      at org.scalatest.tools.Runner.main(Runner.scala)