Can I just ignore this error? I assume it is only a scanner error. What
caused this? Memory, disk space, max-file, I checked all these, but seems
very unlikly caused these conditions.

Regards,
Yan

2009/3/1 Ryan Smith <[email protected]>

> Hi guys,
>
> http://pastebin.com/m1c7b01da
>
> I am also getting the ScannerTimeoutException on HBase/Hadoop 0.19.0.  I am
> loading millions of records in 20GB of data in a mapreduce job.  Just
> trying
> to map (write) one text file to one hbase table.  This job would complete
> fine on 0.18.x   This is the first time ive tried to run this on 0.19.0.
> Ive repeated this test under several 0.19.0 clouds and i keep getting this
> ScannerTimeoutException.   I am going to test it once again on 0.18.x to
> make sure that is the problem.
>
> -alphaomega
>
>
>
> On Sun, Mar 1, 2009 at 3:39 AM, Liu Yan <[email protected]> wrote:
>
> > I have 2 tables, each with about 350K rows (I did "count" on them).
> >
> > For the first table, the scan kept running for about 5 hours before
> failing
> > out. For the second table, it took about 5 minutes to see the same error.
> >
> > How to increase the timeout, I searched for "timeout" or "scan" keyword
> > under the ./conf folder, didn't find anything too helpful.
> >
> > Regards,
> > Yan
> >
> > 2009/3/1 stack <[email protected]>
> >
> > > Before you hit the exception, had it been scanning fine?
> > > The exception from the regionserver is a complaint that an outstanding
> > > scanner didn't check in within the configured client timeout.  Was
> > > something
> > > happening on the client?  Was the cell being fetched extra large?  You
> > > could
> > > up the timeout and see if that helps.
> > >
> > > St.Ack
> > >
> > >
> > > On Sat, Feb 28, 2009 at 8:00 AM, Liu Yan <[email protected]>
> wrote:
> > >
> > > > hi,
> > > >
> > > > I tried to scan a table with "scan 'table_name'" command in the hbase
> > > > shell,
> > > > and hit the following exception:
> > > >
> > > > {{{
> > > > NativeException: java.lang.RuntimeException:
> > > > org.apache.hadoop.hbase.client.ScannerTimeoutException
> > > >    from org/apache/hadoop/hbase/client/HTable.java:1704:in `hasNext'
> > > >    from sun.reflect.GeneratedMethodAccessor10:-1:in `invoke'
> > > >    from sun/reflect/DelegatingMethodAccessorImpl.java:25:in `invoke'
> > > >    from java/lang/reflect/Method.java:597:in `invoke'
> > > >    from org/jruby/javasupport/JavaMethod.java:250:in
> > > > `invokeWithExceptionHandling'
> > > >    from org/jruby/javasupport/JavaMethod.java:219:in `invoke'
> > > >    from org/jruby/javasupport/JavaClass.java:416:in `execute'
> > > >    from
> > > org/jruby/internal/runtime/methods/SimpleCallbackMethod.java:67:in
> > > > `call'
> > > >    from org/jruby/internal/runtime/methods/DynamicMethod.java:70:in
> > > `call'
> > > >    from org/jruby/runtime/CallSite.java:295:in `call'
> > > >    from org/jruby/evaluator/ASTInterpreter.java:646:in `callNode'
> > > >    from org/jruby/evaluator/ASTInterpreter.java:324:in `evalInternal'
> > > >    from org/jruby/evaluator/ASTInterpreter.java:1790:in `whileNode'
> > > >    from org/jruby/evaluator/ASTInterpreter.java:505:in `evalInternal'
> > > >    from org/jruby/evaluator/ASTInterpreter.java:620:in `blockNode'
> > > >    from org/jruby/evaluator/ASTInterpreter.java:318:in `evalInternal'
> > > > ... 121 levels...
> > > >    from
> > > >
> > > >
> > >
> >
> ruby.usr.local.hbase_minus_0_dot_19_dot_0.bin.hirbInvokermethod__32$RUBY$startOpt:-1:in
> > > > `call'
> > > >    from org/jruby/internal/runtime/methods/DynamicMethod.java:74:in
> > > `call'
> > > >    from org/jruby/internal/runtime/methods/CompiledMethod.java:48:in
> > > `call'
> > > >    from org/jruby/runtime/CallSite.java:123:in `cacheAndCall'
> > > >    from org/jruby/runtime/CallSite.java:298:in `call'
> > > >    from
> > > >
> > > >
> > >
> >
> ruby/usr/local/hbase_minus_0_dot_19_dot_0/bin//usr/local/hbase/bin/../bin/hirb.rb:429:in
> > > > `__file__'
> > > >    from
> > > >
> > > >
> > >
> >
> ruby/usr/local/hbase_minus_0_dot_19_dot_0/bin//usr/local/hbase/bin/../bin/hirb.rb:-1:in
> > > > `__file__'
> > > >    from
> > > >
> > > >
> > >
> >
> ruby/usr/local/hbase_minus_0_dot_19_dot_0/bin//usr/local/hbase/bin/../bin/hirb.rb:-1:in
> > > > `load'
> > > >    from org/jruby/Ruby.java:512:in `runScript'
> > > >    from org/jruby/Ruby.java:432:in `runNormally'
> > > >    from org/jruby/Ruby.java:312:in `runFromMain'
> > > >    from org/jruby/Main.java:144:in `run'
> > > >    from org/jruby/Main.java:89:in `run'
> > > >    from org/jruby/Main.java:80:in `main'
> > > >    from /usr/local/hbase/bin/../bin/hirb.rb:334:in `scan'
> > > >    from (hbase):3:in `binding
> > > > }}}
> > > >
> > > > In the regionserver log file, I found this:
> > > >
> > > > {{{
> > > > 2009-02-28 10:50:48,685 INFO
> > > org.apache.hadoop.hbase.regionserver.HRegion:
> > > > compaction completed on region
> > > > 1002_profiles,113495530684109376,1235659992234 in 0sec
> > > > 2009-02-28 10:50:54,779 INFO
> > > > org.apache.hadoop.hbase.regionserver.HRegionServer: Scanner
> > > > 310687422691357774 lease expired2009-02-28 10:50:55,869 ERROR
> > > > org.apache.hadoop.hbase.regionserver.HRegionServer
> > > > : org.apache.hadoop.hbase.UnknownScannerException: Name:
> > > 310687422691357774
> > > > 2009-02-28 10:50:55,871 INFO org.apache.hadoop.ipc.HBaseServer: IPC
> > > Server
> > > > handler 7 on 60020, call next(310687422691357774, 30) from
> > > > 10.254.51.127:54821: error
> > > > : org.apache.hadoop.hbase.UnknownScannerException: Name:
> > > 310687422691357774
> > > > org.apache.hadoop.hbase.UnknownScannerException: Name:
> > > > 310687422691357774        at
> > > > org.apache.hadoop.hbase.regionserver.HRegionServer.next(HRegionServer
> > > > .java:1568)
> > > >        at sun.reflect.GeneratedMethodAccessor11.invoke(Unknown
> Source)
> > > >        at
> > > >
> > > >
> > >
> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> > > >        at java.lang.reflect.Method.invoke(Method.java:597)        at
> > > > org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:632)
> > > >        at
> > > >
> > org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:895)
> > > > 2009-02-28 10:51:08,687 INFO
> > > org.apache.hadoop.hbase.regionserver.HRegion:
> > > > starting  compaction on region
> > > > 1002_profiles,266291625891696512,1235713612502
> > > > 2009-02-28 10:51:08,688 DEBUG
> > > org.apache.hadoop.hbase.regionserver.HStore:
> > > > 2131509311/fetl: no store files to compact
> > > > 2009-02-28 10:51:08,689 DEBUG
> > > org.apache.hadoop.hbase.regionserver.HStore:
> > > > 2131509311/pre_fetl: no store files to compact
> > > > 2009-02-28 10:51:08,691 DEBUG
> > > org.apache.hadoop.hbase.regionserver.HStore:
> > > > 2131509311/scored: no store files to compact
> > > > 2009-02-28 10:51:08,692 DEBUG
> > > org.apache.hadoop.hbase.regionserver.HStore:
> > > > 2131509311/reverse_edge: no store files to compact
> > > > 2009-02-28 10:51:08,692 DEBUG
> > > org.apache.hadoop.hbase.regionserver.HStore:
> > > > 2131509311/inferred: no store files to compact
> > > > }}}
> > > >
> > > > I didn't find any exception in the master log file.
> > > >
> > > > Can any one help?
> > > >
> > > > Regards,
> > > > Yan
> > > >
> > >
> >
>

Reply via email to