;Sea"<261810...@qq.com>;
: Re: Exception in using updateStateByKey
Can you make the value for spark.cleaner.ttl larger ?Cheers
On Mon, Apr 27, 2015 at 7:13 AM, Sea <261810...@qq.com> wrote:
my hadoop version is 2.2.0?? the hdfs-audit.log is too large?? The problem is
that?
I make it to 240, it happens again when 240 seconds is reached.
-- --
??: "261810726";<261810...@qq.com>;
: 2015??4??27??(??) 10:24
??: "Ted Yu";
: ?? Exception in using updateStateByKey
Which hadoop release are you using ?
Can you check hdfs audit log to see who / when deleted spark/ck/hdfsaudit/
receivedData/0/log-1430139541443-1430139601443 ?
Cheers
On Mon, Apr 27, 2015 at 6:21 AM, Sea <261810...@qq.com> wrote:
> Hi, all:
> I use function updateStateByKey in Spark Streaming,
Hi, all:
I use function updateStateByKey in Spark Streaming, I need to store the states
for one minite, I set "spark.cleaner.ttl" to 120, the duration is 2 seconds,
but it throws Exception
Caused by:
org.apache.hadoop.ipc.RemoteException(java.io.FileNotFoundException): File does
not exist