Hi, so I'm slowly getting an overview of the code, here. I haven't really understood the problem yet, though.
DataInput and DataOutput cannot handle BigDecimal, which seems to be somehwere close to the root cause of the problem. The error is being triggered in HBaseServer on line 1642 param.readFields(dis); which calls org.apache.hadoop.io.writable, which implements write and readFields(), and I assume is being implemented by HbaseObjectWritable#readFields. In HbaseObjectWritable#readObject DataInput in then gets checked for being a primitive data type and read accordingly. Now if I interpret Bytes#valueOf() correctly, it just takes a BigDecimal value and converts _just_ the value to byte[] and not the whole object. So what readObject finds here, should be interpreted as byte[] and happily passed on. The first method, that should even care about parsing this to BigDecimal would then be BigDecimalColumnInterpreter#getValue() To test this, I decided to overwrite write and readFields, as I inherit them from Exec, anyway, however I have no understanding of how these methods work. I put in a few printlns to get a feeling for it, but turns out it is never even beling called, at all. 2012/9/10 anil gupta <[email protected]> > Hi Julian, > > I am using only cdh4 libraries. I use the jars present under hadoop and > hbase installed dir. In my last email i gave you some more pointers. Try to > follow them and see what happens. > If then also it doesn't works for you, then i will try to write an utility > to test BigDecimalColumnInterpreter on your setup also. > > Thanks, > Anil > > On Mon, Sep 10, 2012 at 9:36 AM, Julian Wissmann > <[email protected]>wrote: > > > Hi, > > > > I haven't really gotten to working on this, since last wednesday. > > Checked readFields() and write() today, but don't really see, why I would > > need to reimplement those. Admittedly I'm not that into the whole Hbase > > codebase, yet, so there is a good chance I'm missing something, here. > > > > Also, Anil, what hbase library are you coding this against? > > It does seem like madness, that even though, we're both using this > > identically it does not work for me. > > > > Cheers, > > > > Julian > > > > 2012/9/6 anil gupta <[email protected]> > > > > > Yes, we do. :) > > > Let me know the outcome. If you look at the BD ColumnInterpreter, > > getValue > > > method is converting the byte array into BigDecimal. So you should not > > have > > > any problem. The BD ColumnInterpreter is pretty similar to > > > LongColumnInterpreter. > > > > > > Here is the code snippet for getValue() method which will convert > Byte[] > > to > > > BigDecimal: > > > > > > @Override > > > public BigDecimal getValue(byte[] paramArrayOfByte1, byte[] > > > paramArrayOfByte2, > > > KeyValue kv) throws IOException { > > > if ((kv == null || kv.getValue() == null)) > > > return null; > > > return Bytes.toBigDecimal(kv.getValue()); > > > } > > > > > > Thanks, > > > Anil > > > > > > > > > On Thu, Sep 6, 2012 at 11:43 AM, Julian Wissmann > > > <[email protected]>wrote: > > > > > > > 0.92.1 from cdh4. I assume we use the same thing. > > > > > > > > 2012/9/6 anil gupta <[email protected]> > > > > > > > > > I am using HBase0.92.1. Which version you are using? > > > > > > > > > > > > > > > On Thu, Sep 6, 2012 at 10:19 AM, anil gupta <[email protected] > > > > > > wrote: > > > > > > > > > > > Hi Julian, > > > > > > > > > > > > You need to add the column qualifier explicitly in the scanner. > You > > > > have > > > > > > only added the column family in the scanner. > > > > > > I am also assuming that you are writing a ByteArray of BigDecimal > > > > object > > > > > > as value of these cells in HBase. Is that right? > > > > > > > > > > > > Thanks, > > > > > > Anil > > > > > > > > > > > > > > > > > > On Thu, Sep 6, 2012 at 2:28 AM, Julian Wissmann < > > > > > [email protected]>wrote: > > > > > > > > > > > >> Hi, anil, > > > > > >> > > > > > >> I presume you mean something like this: > > > > > >> Scan scan = new Scan(_start, _end); > > > > > >> scan.addFamily(family.getBytes()); > > > > > >> final ColumnInterpreter<BigDecimal, BigDecimal> ci = new > > > > > >> mypackage.BigDecimalColumnInterpreter(); > > > > > >> AggregationClient ag = new org.apache.hadoop.hbase. > > > > > >> client.coprocessor.AggregationClient(config); > > > > > >> BigDecimal sum = ag.sum(Bytes.toBytes(tableName), new > > > > > >> BigDecimalColumnInterpreter(), scan); > > > > > >> > > > > > >> > > > > > >> When I call this,with the Endpoint in place and loaded as a > jar, I > > > get > > > > > the > > > > > >> above error. > > > > > >> When I call it without the endpoint loaded as coprocessor, > > though, I > > > > get > > > > > >> this: > > > > > >> > > > > > >> java.util.concurrent.ExecutionException: > > > > > >> org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed > > > after > > > > > >> attempts=10, exceptions: > > > > > >> Thu Sep 06 11:07:39 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:07:40 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:07:41 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:07:42 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:07:44 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:07:46 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:07:50 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:07:54 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:08:02 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:08:18 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> > > > > > >> at > > > java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:252) > > > > > >> at java.util.concurrent.FutureTask.get(FutureTask.java:111) > > > > > >> at > > > > > >> > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.processExecs(HConnectionManager.java:1434) > > > > > >> at > > > > > > > org.apache.hadoop.hbase.client.HTable.coprocessorExec(HTable.java:1263) > > > > > >> at > > > > > >> > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.coprocessor.AggregationClient.sum(AggregationClient.java:259) > > > > > >> at > > > > > >> > > > > > >> > > > > > > > > > > > > > > > helpers.HbaseReaderBigDecimal.getWeeksumSCAN(HbaseReaderBigDecimal.java:355) > > > > > >> at > > helpers.HbaseReaderBigDecimal.main(HbaseReaderBigDecimal.java:76) > > > > > >> Caused by: > > org.apache.hadoop.hbase.client.RetriesExhaustedException: > > > > > >> Failed > > > > > >> after attempts=10, exceptions: > > > > > >> Thu Sep 06 11:07:39 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:07:40 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:07:41 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:07:42 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:07:44 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:07:46 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:07:50 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:07:54 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:08:02 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> Thu Sep 06 11:08:18 CEST 2012, > > > > > >> org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@7bd6747b, > > > > > >> java.io.IOException: > > > > > >> IPC server unable to read call parameters: Error in readFields > > > > > >> > > > > > >> at > > > > > >> > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.ServerCallable.withRetries(ServerCallable.java:183) > > > > > >> at > > > > > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker.invoke(ExecRPCInvoker.java:79) > > > > > >> at $Proxy7.getSum(Unknown Source) > > > > > >> at > > > > > >> > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.coprocessor.AggregationClient$4.call(AggregationClient.java:263) > > > > > >> at > > > > > >> > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.coprocessor.AggregationClient$4.call(AggregationClient.java:260) > > > > > >> at > > > > > >> > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation$4.call(HConnectionManager.java:1422) > > > > > >> at > > > java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) > > > > > >> at java.util.concurrent.FutureTask.run(FutureTask.java:166) > > > > > >> at > > > > > >> > > > > > >> > > > > > > > > > > > > > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) > > > > > >> at > > > > > >> > > > > > >> > > > > > > > > > > > > > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) > > > > > >> at java.lang.Thread.run(Thread.java:679) > > > > > >> > > > > > >> Interestingly, the RegionServer-log doesn't give me any errors, > > > here, > > > > > just > > > > > >> some info stuff, that suggests it was actually doing something. > > > > > >> Did you implement readFields() and write(), then as Ted > suggested? > > > > > >> > > > > > >> > > > > > >> 2012/9/5 anil gupta <[email protected]> > > > > > >> > > > > > >> > Hi Julian, > > > > > >> > > > > > > >> > Sorry for wrong reference to the aggregation client class in > my > > > > > previous > > > > > >> > email. Here is the right > > > > > >> > > > > class:org.apache.hadoop.hbase.client.coprocessor.AggregationClient. > > > > > >> > > > > > > >> > HTH, > > > > > >> > Anil > > > > > >> > > > > > > >> > On Wed, Sep 5, 2012 at 2:04 PM, anil gupta < > > [email protected] > > > > > > > > > >> wrote: > > > > > >> > > > > > > >> > > Hi Julian, > > > > > >> > > > > > > > >> > > I have been running the same class on my distributed cluster > > for > > > > > >> > > aggregation. It has been working fine. The only difference > is > > > > that i > > > > > >> use > > > > > >> > > the methods provided > > > > incom.intuit.ihub.hbase.poc.aggregation.client. > > > > > >> > > AggregationClient class. IMHO, you don't need to define an > > > > Endpoint > > > > > >> for > > > > > >> > > using the BigDecimalColumnInterpreter. > > > > > >> > > > > > > > >> > > You need to use methods of AggregationClient: > > > > > >> > > sum(Bytes.toBytes(tableName), ci, scan), avg(final byte[] > > > > tableName, > > > > > >> > final > > > > > >> > > ColumnInterpreter<R, S> ci, Scan scan), etc. > > > > > >> > > > > > > > >> > > In the above method you just need to pass the > > > > > >> > BigDecimalColumnInterpreter, > > > > > >> > > Scan object and Byte Array of TableName. It should work. Let > > me > > > > know > > > > > >> if > > > > > >> > it > > > > > >> > > doesn't work this way.* > > > > > >> > > * > > > > > >> > > Thanks, > > > > > >> > > Anil Gupta > > > > > >> > > * > > > > > >> > > * > > > > > >> > > > > > > > >> > > On Wed, Sep 5, 2012 at 1:30 PM, Julian Wissmann < > > > > > >> > [email protected]>wrote: > > > > > >> > > > > > > > >> > >> Thank you! > > > > > >> > >> So this looks like the missing link here. > > > > > >> > >> I'll see if I can get it working, tomorrow morning. > > > > > >> > >> > > > > > >> > >> Cheers > > > > > >> > >> > > > > > >> > >> 2012/9/5 Ted Yu <[email protected]> > > > > > >> > >> > > > > > >> > >> > I added one review comment on > > > > > >> > >> > HBASE-6669< > > https://issues.apache.org/jira/browse/HBASE-6669> > > > > > >> > >> > . > > > > > >> > >> > > > > > > >> > >> > Thanks Julian for reminding me. > > > > > >> > >> > > > > > > >> > >> > On Wed, Sep 5, 2012 at 12:49 PM, Julian Wissmann > > > > > >> > >> > <[email protected]>wrote: > > > > > >> > >> > > > > > > >> > >> > > I get supplied with doubles from sensors, but in the > end > > I > > > > > loose > > > > > >> too > > > > > >> > >> much > > > > > >> > >> > > precision if I do my aggregations on double, otherwise > > I'd > > > go > > > > > for > > > > > >> > it. > > > > > >> > >> > > I use 0.92.1, from Cloudera CDH4. > > > > > >> > >> > > I've done some initial testing with > LongColumnInterpreter > > > on > > > > a > > > > > >> > dataset > > > > > >> > >> > that > > > > > >> > >> > > I've generated, to do some testing and get accustomed > to > > > > stuff, > > > > > >> but > > > > > >> > >> that > > > > > >> > >> > > worked like a charm after some initial stupidity on my > > > side. > > > > > >> > >> > > So now I'm trying to do some testing with the real > data, > > > > which > > > > > >> comes > > > > > >> > >> in > > > > > >> > >> > as > > > > > >> > >> > > double and gets parsed to BigDecimal before writing. > > > > > >> > >> > > > > > > > >> > >> > > 2012/9/5 Ted Yu <[email protected]> > > > > > >> > >> > > > > > > > >> > >> > > > And your HBase version is ? > > > > > >> > >> > > > > > > > > >> > >> > > > Since you use Double.parseDouble(), looks like it > would > > > be > > > > > more > > > > > >> > >> > efficient > > > > > >> > >> > > > to develop DoubleColumnInterpreter. > > > > > >> > >> > > > > > > > > >> > >> > > > On Wed, Sep 5, 2012 at 12:07 PM, Julian Wissmann > > > > > >> > >> > > > <[email protected]>wrote: > > > > > >> > >> > > > > > > > > >> > >> > > > > Hi, > > > > > >> > >> > > > > the schema looks like this: > > > > > >> > >> > > > > RowKey: id,timerange_timestamp,offset (String) > > > > > >> > >> > > > > Qualifier: Offset (long) > > > > > >> > >> > > > > Timestamp: timestamp (long) > > > > > >> > >> > > > > Value:number (BigDecimal) > > > > > >> > >> > > > > > > > > > >> > >> > > > > Or as code when I read data from csv:byte[] value = > > > > > >> > >> > > > > > > > > > >> Bytes.toBytes(BigDecimal.valueOf(Double.parseDouble(cData[2]))); > > > > > >> > >> > > > > > > > > > >> > >> > > > > Cheers, > > > > > >> > >> > > > > > > > > > >> > >> > > > > Julian > > > > > >> > >> > > > > > > > > > >> > >> > > > > 2012/9/5 Ted Yu <[email protected]> > > > > > >> > >> > > > > > > > > > >> > >> > > > > > You haven't told us the schema of your table yet. > > > > > >> > >> > > > > > Your table should have column whose value can be > > > > > >> interpreted > > > > > >> > by > > > > > >> > >> > > > > > BigDecimalColumnInterpreter. > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > Cheers > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > On Wed, Sep 5, 2012 at 9:17 AM, Julian Wissmann < > > > > > >> > >> > > > > [email protected] > > > > > >> > >> > > > > > >wrote: > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > Hi, > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > I am currently experimenting with the > > > > > >> > >> BigDecimalColumnInterpreter > > > > > >> > >> > > > from > > > > > >> > >> > > > > > > > https://issues.apache.org/jira/browse/HBASE-6669 > > . > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > I was thinking the best way for me to work with > > it > > > > > would > > > > > >> be > > > > > >> > to > > > > > >> > >> > use > > > > > >> > >> > > > the > > > > > >> > >> > > > > > Java > > > > > >> > >> > > > > > > class and just use that as is. > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > Imported it into my project and tried to work > > with > > > it > > > > > as > > > > > >> is, > > > > > >> > >> by > > > > > >> > >> > > just > > > > > >> > >> > > > > > > instantiating the ColumnInterpreter as > > > > > >> > >> > BigDecimalColumnInterpreter. > > > > > >> > >> > > > > Okay, > > > > > >> > >> > > > > > > threw errors and also complained about not > > knowing > > > > > where > > > > > >> to > > > > > >> > >> find > > > > > >> > >> > > > such a > > > > > >> > >> > > > > > > class. > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > So I did some reading and found out, that I'd > > need > > > to > > > > > >> have > > > > > >> > an > > > > > >> > >> > > > Endpoint > > > > > >> > >> > > > > > for > > > > > >> > >> > > > > > > it. So I imported AggregateImplementation and > > > > > >> > >> AggregateProtocol > > > > > >> > >> > > into > > > > > >> > >> > > > my > > > > > >> > >> > > > > > > workspace, renamed them, and refactored them > > where > > > > > >> necessary > > > > > >> > >> to > > > > > >> > >> > > take > > > > > >> > >> > > > > > > BigDecimal. Re-exported the jar, then and had > > > another > > > > > >> try. > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > So when I call: > > > > > >> > >> > > > > > > ------ > > > > > >> > >> > > > > > > final Scan scan = new Scan((metricID + "," + > > > > > >> > >> > > > > basetime_begin).getBytes(), > > > > > >> > >> > > > > > > (metricID + "," + basetime_end).getBytes()); > > > > > >> > >> > > > > > > scan.addFamily(family.getBytes()); > > > > > >> > >> > > > > > > final ColumnInterpreter<BigDecimal, BigDecimal> > > ci > > > = > > > > > new > > > > > >> > >> > > > > > > BigDecimalColumnInterpreter(); > > > > > >> > >> > > > > > > Map<byte[], BigDecimal> results = > > > > > >> > >> > > > > > > table.coprocessorExec(BigDecimalProtocol.class, > > > null, > > > > > >> null, > > > > > >> > >> > > > > > > new > > > Batch.Call<BigDecimalProtocol,BigDecimal>() { > > > > > >> > >> > > > > > > public BigDecimal call(BigDecimalProtocol > > > > > >> > >> instance)throws > > > > > >> > >> > > > > > > IOException{ > > > > > >> > >> > > > > > > return instance.getMax(ci, scan); > > > > > >> > >> > > > > > > } > > > > > >> > >> > > > > > > }); > > > > > >> > >> > > > > > > ------ > > > > > >> > >> > > > > > > I get errors in the log again, that it can't > find > > > > > >> > >> > > > > > > BigDecimalColumnInterpreter... okay, so I tried > > > > > >> > >> > > > > > > ------ > > > > > >> > >> > > > > > > Scan scan = new Scan((metricID + "," + > > > > > >> > >> > basetime_begin).getBytes(), > > > > > >> > >> > > > > > > (metricID + "," + basetime_end).getBytes()); > > > > > >> > >> > > > > > > scan.addFamily(family.getBytes()); > > > > > >> > >> > > > > > > final ColumnInterpreter<BigDecimal, BigDecimal> > > ci > > > = > > > > > new > > > > > >> > >> > > > > > > BigDecimalColumnInterpreter(); > > > > > >> > >> > > > > > > AggregationClient ag = new > > > AggregationClient(config); > > > > > >> > >> > > > > > > BigDecimal max = > ag.max(Bytes.toBytes(tableName), > > > ci, > > > > > >> scan); > > > > > >> > >> > > > > > > ------ > > > > > >> > >> > > > > > > I don't get errors recored in the log anymore, > > but > > > a > > > > > >> load of > > > > > >> > >> Java > > > > > >> > >> > > > error > > > > > >> > >> > > > > > > output: > > > > > >> > >> > > > > > > ------ > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > java.util.concurrent.ExecutionException: > > > > > >> > >> > > > > > > > > > > > org.apache.hadoop.hbase.client.RetriesExhaustedException: > > > > > >> > >> Failed > > > > > >> > >> > > > after > > > > > >> > >> > > > > > > attempts=10, exceptions: > > > > > >> > >> > > > > > > Wed Sep 05 18:13:43 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:44 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:45 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:46 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:49 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:51 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:55 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:59 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:14:07 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:14:23 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > >> > java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:252) > > > > > >> > >> > > > > > > at > > > > > >> java.util.concurrent.FutureTask.get(FutureTask.java:111) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.processExecs(HConnectionManager.java:1434) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > >> > > > > > org.apache.hadoop.hbase.client.HTable.coprocessorExec(HTable.java:1263) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.coprocessor.AggregationClient.sum(AggregationClient.java:259) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > helpers.HbaseReaderBigDecimal.getWeeksumSCAN(HbaseReaderBigDecimal.java:360) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > helpers.HbaseReaderBigDecimal.main(HbaseReaderBigDecimal.java:81) > > > > > >> > >> > > > > > > Caused by: > > > > > >> > >> > > > org.apache.hadoop.hbase.client.RetriesExhaustedException: > > > > > >> > >> > > > > > Failed > > > > > >> > >> > > > > > > after attempts=10, exceptions: > > > > > >> > >> > > > > > > Wed Sep 05 18:13:43 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:44 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:45 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:46 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:49 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:51 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:55 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:59 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:14:07 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:14:23 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.ServerCallable.withRetries(ServerCallable.java:183) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker.invoke(ExecRPCInvoker.java:79) > > > > > >> > >> > > > > > > at $Proxy7.getSum(Unknown Source) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.coprocessor.AggregationClient$4.call(AggregationClient.java:263) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.coprocessor.AggregationClient$4.call(AggregationClient.java:260) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation$4.call(HConnectionManager.java:1422) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > >> > java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) > > > > > >> > >> > > > > > > at > > > > > >> java.util.concurrent.FutureTask.run(FutureTask.java:166) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) > > > > > >> > >> > > > > > > at java.lang.Thread.run(Thread.java:679) > > > > > >> > >> > > > > > > > > > > > org.apache.hadoop.hbase.client.RetriesExhaustedException: > > > > > >> > >> Failed > > > > > >> > >> > > > after > > > > > >> > >> > > > > > > attempts=10, exceptions: > > > > > >> > >> > > > > > > Wed Sep 05 18:13:43 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:44 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:45 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:46 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:49 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:51 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:55 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:13:59 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:14:07 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > Wed Sep 05 18:14:23 CEST 2012, > > > > > >> > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker$1@50502819 > > > > , > > > > > >> > >> > > > > > > java.io.IOException: > > > > > >> > >> > > > > > > IPC server unable to read call parameters: > Error > > in > > > > > >> > readFields > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.ServerCallable.withRetries(ServerCallable.java:183) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > org.apache.hadoop.hbase.ipc.ExecRPCInvoker.invoke(ExecRPCInvoker.java:79) > > > > > >> > >> > > > > > > at $Proxy7.getSum(Unknown Source) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.coprocessor.AggregationClient$4.call(AggregationClient.java:263) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.coprocessor.AggregationClient$4.call(AggregationClient.java:260) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation$4.call(HConnectionManager.java:1422) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > >> > java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) > > > > > >> > >> > > > > > > at > > > > > >> java.util.concurrent.FutureTask.run(FutureTask.java:166) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) > > > > > >> > >> > > > > > > at > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) > > > > > >> > >> > > > > > > at java.lang.Thread.run(Thread.java:679) > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > ------ > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > I'm not really sure about what I'm doing wrong. > > > Does > > > > > >> anyone > > > > > >> > >> have > > > > > >> > >> > a > > > > > >> > >> > > > hint > > > > > >> > >> > > > > > > towards the right direction? > > > > > >> > >> > > > > > > > > > > > >> > >> > > > > > > > > > > >> > >> > > > > > > > > > >> > >> > > > > > > > > >> > >> > > > > > > > >> > >> > > > > > > >> > >> > > > > > >> > > > > > > > >> > > > > > > > >> > > > > > > > >> > > -- > > > > > >> > > Thanks & Regards, > > > > > >> > > Anil Gupta > > > > > >> > > > > > > > >> > > > > > > >> > > > > > > >> > > > > > > >> > -- > > > > > >> > Thanks & Regards, > > > > > >> > Anil Gupta > > > > > >> > > > > > > >> > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > > > Thanks & Regards, > > > > > > Anil Gupta > > > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > > Thanks & Regards, > > > > > Anil Gupta > > > > > > > > > > > > > > > > > > > > > -- > > > Thanks & Regards, > > > Anil Gupta > > > > > > > > > -- > Thanks & Regards, > Anil Gupta >
