UNOFFICIAL

Josh,

Zookeepr - 3.4.5-cdh4.3.0
Accumulo - 1.5.0
Hadoop - cdh 4.3.0

In the accumulo console getting

ERROR RemoteException(...LeaseExpiredException): Lease mismatch on 
/accumulo/wal/<ip>+9997/<uid> owned by DFSClient_NONMAPREDUCE_699577321_12 but 
is accessed by DFSClient_NONMAPREDUCE_903051502_12

We can scan the table without issues and can load rows directly, ie not using 
bulk import.  

A bit more information - we recently extended how we manage old tablets in the 
system. We load data by date, creating splits for each day and then ageoff 
using the ageoff filters.  This leaves empty tablets so we now merge these old 
tablets together to effectively remove them.  I mention it because I'm not sure 
if this might have introduced another issue.  

Matt

-----Original Message-----
From: Josh Elser [mailto:[email protected]]
Sent: Monday, 17 February 2014 11:32
To: [email protected]
Subject: Re: Failing to BulkIngest [SEC=UNOFFICIAL]

Matt,

Can you provide Hadoop, ZK and Accumulo versions? Does the cluster appear to be 
functional otherwise (can you scan that table you're bulk importing to? any 
other errors on the monitor? etc)

On 2/16/14, 7:07 PM, Dickson, Matt MR wrote:
> *UNOFFICIAL*
>
> I have a situation where bulk ingests are failing with a "Thread "shell"
> stuck on IO to xxx:9999:99999 ...
>  From the management console the table we are loading to has no 
> compactions running, yet we ran "./accumulo 
> org.apache.accumulo.server.fate.Admin print and can see 50,000 lines 
> stating
> txid: xxxx     status:IN_PROGRESS op: CompactRange     locked: []
> locking: [R:3n]     top: Compact:Range
> Does this mean there are actually compactions running or old 
> comapaction locks still hanging around that will be preventing the builk 
> ingest to run?
> Thanks in advance,
> Matt

Reply via email to