[ 
https://issues.apache.org/jira/browse/SPARK-21248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16067425#comment-16067425
 ] 

Apache Spark commented on SPARK-21248:
--------------------------------------

User 'zsxwing' has created a pull request for this issue:
https://github.com/apache/spark/pull/18461

> Flaky test: o.a.s.sql.kafka010.KafkaSourceSuite.assign from specific offsets 
> (failOnDataLoss: true)
> ---------------------------------------------------------------------------------------------------
>
>                 Key: SPARK-21248
>                 URL: https://issues.apache.org/jira/browse/SPARK-21248
>             Project: Spark
>          Issue Type: Bug
>          Components: Structured Streaming
>    Affects Versions: 2.3.0
>            Reporter: Shixiong Zhu
>
> {code}
> org.scalatest.exceptions.TestFailedException:  Stream Thread Died: null 
> java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1326)
>   scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:208)  
> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:218)  
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)  
> org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:201)  
> org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)  
> org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:92)  
> org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:76)  
> org.apache.spark.sql.execution.streaming.state.StateStoreCoordinatorRef.deactivateInstances(StateStoreCoordinator.scala:108)
>   
> org.apache.spark.sql.streaming.StreamingQueryManager.notifyQueryTermination(StreamingQueryManager.scala:335)
>    == Progress ==    AssertOnQuery(<condition>, )    CheckAnswer: 
> [-20],[-21],[-22],[0],[1],[2],[11],[12],[22]    StopStream    
> StartStream(ProcessingTime(0),org.apache.spark.util.SystemClock@6c63901,Map())
>     CheckAnswer: [-20],[-21],[-22],[0],[1],[2],[11],[12],[22]    
> AddKafkaData(topics = Set(topic-7), data = WrappedArray(30, 31, 32, 33, 34), 
> message = )    CheckAnswer: 
> [-20],[-21],[-22],[0],[1],[2],[11],[12],[22],[30],[31],[32],[33],[34]    
> StopStream  == Stream == Output Mode: Append Stream state: not started Thread 
> state: dead  java.lang.InterruptedException  at 
> java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireSharedNanos(AbstractQueuedSynchronizer.java:1326)
>   at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:208) 
>  at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:218)  at 
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)  at 
> org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:201)  at 
> org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)  at 
> org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:92)  at 
> org.apache.spark.rpc.RpcEndpointRef.askSync(RpcEndpointRef.scala:76)  at 
> org.apache.spark.sql.execution.streaming.state.StateStoreCoordinatorRef.deactivateInstances(StateStoreCoordinator.scala:108)
>   at 
> org.apache.spark.sql.streaming.StreamingQueryManager.notifyQueryTermination(StreamingQueryManager.scala:335)
>   at 
> org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runBatches(StreamExecution.scala:375)
>   at 
> org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:211)
>    == Sink == 0: [-20] [-21] [-22] [22] [11] [12] [0] [1] [2] 1: [30] 2: [33] 
> [31] [32] [34]   == Plan ==   
> {code}
> See 
> https://amplab.cs.berkeley.edu/jenkins/job/spark-master-test-sbt-hadoop-2.6/3173/testReport/junit/org.apache.spark.sql.kafka010/KafkaSourceSuite/assign_from_specific_offsets__failOnDataLoss__true_/



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to