Hi all,
we have two environments for spark streaming job, which consumes Kafka
topic to do calculation.
Advertising
Now in one environment, spark streaming job consume an non-standard data
from kafka and throw an excepiton(not catch it in code), then the
sreaming job is down.
But in another environment, spark streaming job also throw an exception(
same exception message in log file), but the streaming job is still
running and consume other data continuously.
Is there some parameters or configuration for this problem? Why one job
is down and another job is still running.
---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org