HeartSaVioR edited a comment on pull request #30427:
URL: https://github.com/apache/spark/pull/30427#issuecomment-730855993


   > If we process history data or some simulation data, the event time could 
be far different to processing time. For example, if we process some data from 
2010 to 2019, now the gap is current time - 2010-xx-xx...?
   
   You understand it correctly, though that's just a one of use cases. Given 
they are running "streaming workload", one of the main goals is to capture the 
recent outputs (e.g. trends). Watermark would still work for such use cases as 
well, but what to plot to provide values even on the situation remains the 
question. (What would be the "ideal" timestamp to calculate the gap in this 
case?)
   
   EDIT: for that case, adjusting range on y axis would probably help, 
otherwise we only see the "line" plotted nearly linear like what I commented 
above.


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]



---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to