I looked into the source code of org/apache/hadoop/hbase/fs/HFileSystem.java
and whenever I execute hbase-daemon.sh stop master (or regionserver), the 
method.getName() is "close",
but the org/apache/hadoop/hdfs/protocol/ClientProtocol.java do not have method 
named "close",
thus it result in error "object is not an instance of declaring class"

Could someone that familiar with hbase-0.96.0 hadoop2 can tell me if this 
problem need to be fixed? And how to fix it?

  private static ClientProtocol createReorderingProxy(final ClientProtocol cp, 
final ReorderBlocks lrb, final Configuration conf) {
    return (ClientProtocol) Proxy.newProxyInstance
        (cp.getClass().getClassLoader(),
            new Class[]{ClientProtocol.class, Closeable.class},
            new InvocationHandler() {
              public Object invoke(Object proxy, Method method,
                                   Object[] args) throws Throwable {
                try {
                  // method.invoke will failed if 
method.getName().equals("close")
                  // because ClientProtocol do not have method "close"
                  Object res = method.invoke(cp, args);
                  if (res != null && args != null && args.length == 3
                      && "getBlockLocations".equals(method.getName())
                      && res instanceof LocatedBlocks
                      && args[0] instanceof String
                      && args[0] != null) {
                    lrb.reorderBlocks(conf, (LocatedBlocks) res, (String) 
args[0]);
                  }
                  return res;
                } catch  (InvocationTargetException ite) {
                  // We will have this for all the exception, checked on not, 
sent
                  //  by any layer, including the functional exception
                  Throwable cause = ite.getCause();
                  if (cause == null){
                    throw new RuntimeException(
                      "Proxy invocation failed and getCause is null", ite);
                  }
                  if (cause instanceof UndeclaredThrowableException) {
                    Throwable causeCause = cause.getCause();
                    if (causeCause == null) {
                      throw new RuntimeException("UndeclaredThrowableException 
had null cause!");
                    }
                    cause = cause.getCause();
                  }
                  throw cause;
                }
              }
            });
  }



-----Original Message-----
From: MA11 YTHung1
Sent: Thursday, November 21, 2013 9:57 AM
To: [email protected]
Subject: RE: hbase 0.96 stop master receive ERROR ipc.RPC: RPC.stopProxy called 
on non proxy.

Additional information:

I replace all files with prefix hadoop in hbase-0.96.0-hadoop2/lib with 
hadoop-2.2.0 libraries.

the ls -l of hbase-0.96.0-hadoop2/lib as below:

-rw-r--r-- 1 hadoop users    62983 Sep 17 16:13 activation-1.1.jar
-rw-r--r-- 1 hadoop users     4467 Sep 17 23:29 aopalliance-1.0.jar
-rw-r--r-- 1 hadoop users    43033 Sep 17 16:13 asm-3.1.jar
-rw-r--r-- 1 hadoop users   263268 Sep 17 16:27 avro-1.5.3.jar
-rw-r--r-- 1 hadoop users   188671 Sep 17 16:12 commons-beanutils-1.7.0.jar
-rw-r--r-- 1 hadoop users   206035 Sep 17 16:13 commons-beanutils-core-1.8.0.jar
-rw-r--r-- 1 hadoop users    41123 Sep 17 16:12 commons-cli-1.2.jar
-rw-r--r-- 1 hadoop users   259600 Sep 17 16:13 commons-codec-1.7.jar
-rw-r--r-- 1 hadoop users   575389 Sep 17 16:12 commons-collections-3.2.1.jar
-rw-r--r-- 1 hadoop users   238681 Sep 17 16:27 commons-compress-1.4.jar
-rw-r--r-- 1 hadoop users   298829 Sep 17 16:13 commons-configuration-1.6.jar
-rw-r--r-- 1 hadoop users    24239 Sep 17 23:28 commons-daemon-1.0.13.jar
-rw-r--r-- 1 hadoop users   143602 Sep 17 16:12 commons-digester-1.8.jar
-rw-r--r-- 1 hadoop users   112341 Sep 17 16:13 commons-el-1.0.jar
-rw-r--r-- 1 hadoop users   305001 Sep 17 16:12 commons-httpclient-3.1.jar
-rw-r--r-- 1 hadoop users   185140 Sep 17 16:13 commons-io-2.4.jar
-rw-r--r-- 1 hadoop users   284220 Sep 17 16:12 commons-lang-2.6.jar
-rw-r--r-- 1 hadoop users    60686 Sep 17 16:12 commons-logging-1.1.1.jar
-rw-r--r-- 1 hadoop users   988514 Sep 17 16:13 commons-math-2.2.jar
-rw-r--r-- 1 hadoop users   273370 Sep 17 16:27 commons-net-3.1.jar
-rw-r--r-- 1 hadoop users  3566844 Sep 17 16:15 core-3.1.1.jar
-rw-r--r-- 1 hadoop users    15322 Sep 17 16:12 findbugs-annotations-1.3.9-1.jar
-rw-r--r-- 1 hadoop users    21817 Sep 17 23:29 gmbal-api-only-3.0.0-b023.jar
-rw-r--r-- 1 hadoop users   684337 Sep 17 23:29 grizzly-framework-2.1.1.jar
-rw-r--r-- 1 hadoop users   210846 Sep 17 23:29 
grizzly-framework-2.1.1-tests.jar
-rw-r--r-- 1 hadoop users   248346 Sep 17 23:29 grizzly-http-2.1.1.jar
-rw-r--r-- 1 hadoop users   193583 Sep 17 23:29 grizzly-http-server-2.1.1.jar
-rw-r--r-- 1 hadoop users   336878 Sep 17 23:29 grizzly-http-servlet-2.1.1.jar
-rw-r--r-- 1 hadoop users     8072 Sep 17 23:29 grizzly-rcm-2.1.1.jar
-rw-r--r-- 1 hadoop users  1795932 Sep 17 16:13 guava-12.0.1.jar
-rw-r--r-- 1 hadoop users   710492 Sep 17 23:29 guice-3.0.jar
-rw-r--r-- 1 hadoop users    65012 Sep 17 23:29 guice-servlet-3.0.jar
-rw-r--r-- 1 hadoop users    16778 Nov 20 17:39 hadoop-annotations-2.2.0.jar
-rw-r--r-- 1 hadoop users    49750 Nov 20 17:40 hadoop-auth-2.2.0.jar
-rw-r--r-- 1 hadoop users     2576 Oct 12 06:20 hadoop-client-2.1.0-beta.jar
-rw-r--r-- 1 hadoop users  2735584 Nov 20 17:50 hadoop-common-2.2.0.jar
-rw-r--r-- 1 hadoop users  5242252 Nov 21 08:48 hadoop-hdfs-2.2.0.jar
-rw-r--r-- 1 hadoop users  1988460 Nov 21 08:48 hadoop-hdfs-2.2.0-tests.jar
-rw-r--r-- 1 hadoop users   482042 Nov 21 08:49 
hadoop-mapreduce-client-app-2.2.0.jar
-rw-r--r-- 1 hadoop users   656365 Nov 21 08:49 
hadoop-mapreduce-client-common-2.2.0.jar
-rw-r--r-- 1 hadoop users  1455001 Nov 21 08:50 
hadoop-mapreduce-client-core-2.2.0.jar
-rw-r--r-- 1 hadoop users    35216 Nov 21 08:50 
hadoop-mapreduce-client-jobclient-2.2.0.jar
-rw-r--r-- 1 hadoop users  1434852 Nov 21 08:50 
hadoop-mapreduce-client-jobclient-2.2.0-tests.jar
-rw-r--r-- 1 hadoop users    21537 Nov 21 08:51 
hadoop-mapreduce-client-shuffle-2.2.0.jar
-rw-r--r-- 1 hadoop users  1158936 Nov 21 08:51 hadoop-yarn-api-2.2.0.jar
-rw-r--r-- 1 hadoop users    94728 Nov 21 08:51 hadoop-yarn-client-2.2.0.jar
-rw-r--r-- 1 hadoop users  1301627 Nov 21 08:51 hadoop-yarn-common-2.2.0.jar
-rw-r--r-- 1 hadoop users   175554 Nov 21 08:52 
hadoop-yarn-server-common-2.2.0.jar
-rw-r--r-- 1 hadoop users   467638 Nov 21 08:52 
hadoop-yarn-server-nodemanager-2.2.0.jar
-rw-r--r-- 1 hadoop users   825853 Oct 12 06:28 hbase-client-0.96.0-hadoop2.jar
-rw-r--r-- 1 hadoop users   354845 Oct 12 06:28 hbase-common-0.96.0-hadoop2.jar
-rw-r--r-- 1 hadoop users   132690 Oct 12 06:28 
hbase-common-0.96.0-hadoop2-tests.jar
-rw-r--r-- 1 hadoop users    97428 Oct 12 06:28 
hbase-examples-0.96.0-hadoop2.jar
-rw-r--r-- 1 hadoop users    72765 Oct 12 06:28 
hbase-hadoop2-compat-0.96.0-hadoop2.jar
-rw-r--r-- 1 hadoop users    32096 Oct 12 06:28 
hbase-hadoop-compat-0.96.0-hadoop2.jar
-rw-r--r-- 1 hadoop users    12174 Oct 12 06:28 hbase-it-0.96.0-hadoop2.jar
-rw-r--r-- 1 hadoop users   288784 Oct 12 06:28 
hbase-it-0.96.0-hadoop2-tests.jar
-rw-r--r-- 1 hadoop users    94784 Oct 12 06:28 
hbase-prefix-tree-0.96.0-hadoop2.jar
-rw-r--r-- 1 hadoop users  3134214 Oct 12 06:28 
hbase-protocol-0.96.0-hadoop2.jar
-rw-r--r-- 1 hadoop users  3058804 Oct 12 06:28 hbase-server-0.96.0-hadoop2.jar
-rw-r--r-- 1 hadoop users  3150292 Oct 12 06:28 
hbase-server-0.96.0-hadoop2-tests.jar
-rw-r--r-- 1 hadoop users    12554 Oct 12 06:28 hbase-shell-0.96.0-hadoop2.jar
-rw-r--r-- 1 hadoop users    10941 Oct 12 06:28 
hbase-testing-util-0.96.0-hadoop2.jar
-rw-r--r-- 1 hadoop users  2276333 Oct 12 06:28 hbase-thrift-0.96.0-hadoop2.jar
-rw-r--r-- 1 hadoop users    95975 Sep 17 16:15 high-scale-lib-1.1.1.jar
-rw-r--r-- 1 hadoop users    31020 Sep 17 16:14 htrace-core-2.01.jar
-rw-r--r-- 1 hadoop users   352585 Sep 17 16:15 httpclient-4.1.3.jar
-rw-r--r-- 1 hadoop users   181201 Sep 17 16:15 httpcore-4.1.3.jar
-rw-r--r-- 1 hadoop users   227517 Sep 17 16:13 jackson-core-asl-1.8.8.jar
-rw-r--r-- 1 hadoop users    17884 Sep 17 16:13 jackson-jaxrs-1.8.8.jar
-rw-r--r-- 1 hadoop users   669065 Sep 17 16:13 jackson-mapper-asl-1.8.8.jar
-rw-r--r-- 1 hadoop users    32353 Sep 17 16:13 jackson-xc-1.8.8.jar
-rw-r--r-- 1 hadoop users    20642 Sep 17 16:15 jamon-runtime-2.3.1.jar
-rw-r--r-- 1 hadoop users   408133 Sep 17 16:13 jasper-compiler-5.5.23.jar
-rw-r--r-- 1 hadoop users    76844 Sep 17 16:13 jasper-runtime-5.5.23.jar
-rw-r--r-- 1 hadoop users     2497 Sep 17 23:29 javax.inject-1.jar
-rw-r--r-- 1 hadoop users    83586 Sep 17 23:29 javax.servlet-3.0.jar
-rw-r--r-- 1 hadoop users   105134 Sep 17 16:27 jaxb-api-2.2.2.jar
-rw-r--r-- 1 hadoop users   890168 Sep 17 16:13 jaxb-impl-2.2.3-1.jar
-rw-r--r-- 1 hadoop users   129217 Sep 17 23:29 jersey-client-1.8.jar
-rw-r--r-- 1 hadoop users   458233 Sep 17 16:13 jersey-core-1.8.jar
-rw-r--r-- 1 hadoop users    17585 Sep 17 23:29 jersey-grizzly2-1.8.jar
-rw-r--r-- 1 hadoop users    14712 Sep 17 23:29 jersey-guice-1.8.jar
-rw-r--r-- 1 hadoop users   147933 Sep 17 16:13 jersey-json-1.8.jar
-rw-r--r-- 1 hadoop users   694352 Sep 17 16:13 jersey-server-1.8.jar
-rw-r--r-- 1 hadoop users    28034 Sep 17 23:29 
jersey-test-framework-core-1.8.jar
-rw-r--r-- 1 hadoop users    12907 Sep 17 23:29 
jersey-test-framework-grizzly2-1.8.jar
-rw-r--r-- 1 hadoop users   321806 Sep 17 16:27 jets3t-0.6.1.jar
-rw-r--r-- 1 hadoop users    75963 Sep 17 16:13 jettison-1.3.1.jar
-rw-r--r-- 1 hadoop users   539912 Sep 17 16:13 jetty-6.1.26.jar
-rw-r--r-- 1 hadoop users    18891 Sep 17 16:15 jetty-sslengine-6.1.26.jar
-rw-r--r-- 1 hadoop users   177131 Sep 17 16:13 jetty-util-6.1.26.jar
-rw-r--r-- 1 hadoop users 13832273 Sep 17 16:15 jruby-complete-1.6.8.jar
-rw-r--r-- 1 hadoop users   185746 Sep 17 16:27 jsch-0.1.42.jar
-rw-r--r-- 1 hadoop users  1024680 Sep 17 16:13 jsp-2.1-6.1.14.jar
-rw-r--r-- 1 hadoop users   134910 Sep 17 16:13 jsp-api-2.1-6.1.14.jar
-rw-r--r-- 1 hadoop users   100636 Sep 17 16:27 jsp-api-2.1.jar
-rw-r--r-- 1 hadoop users    33015 Sep 17 16:13 jsr305-1.3.9.jar
-rw-r--r-- 1 hadoop users   245039 Sep 17 16:12 junit-4.11.jar
-rw-r--r-- 1 hadoop users   347531 Sep 17 16:15 libthrift-0.9.0.jar
-rw-r--r-- 1 hadoop users   489884 Sep 17 16:12 log4j-1.2.17.jar
-rw-r--r-- 1 hadoop users    42212 Sep 17 23:29 management-api-3.0.0-b012.jar
-rw-r--r-- 1 hadoop users    82445 Sep 17 16:14 metrics-core-2.1.2.jar
drwxr-xr-x 3 hadoop users     4096 Nov 21 09:10 native
-rw-r--r-- 1 hadoop users  1206119 Sep 18 04:00 netty-3.6.6.Final.jar
-rw-r--r-- 1 hadoop users    29555 Sep 17 16:27 paranamer-2.3.jar
-rw-r--r-- 1 hadoop users   533455 Sep 17 16:13 protobuf-java-2.5.0.jar
drwxr-xr-x 5 hadoop users     4096 Sep 28 10:37 ruby
-rw-r--r-- 1 hadoop users   132368 Sep 17 16:13 servlet-api-2.5-6.1.14.jar
-rw-r--r-- 1 hadoop users   105112 Sep 17 16:12 servlet-api-2.5.jar
-rw-r--r-- 1 hadoop users    25962 Sep 17 16:14 slf4j-api-1.6.4.jar
-rw-r--r-- 1 hadoop users     9748 Oct  3 07:15 slf4j-log4j12-1.6.4.jar
-rw-r--r-- 1 hadoop users   995720 Sep 17 16:27 snappy-java-1.0.3.2.jar
-rw-r--r-- 1 hadoop users    26514 Sep 17 16:13 stax-api-1.0.1.jar
-rw-r--r-- 1 hadoop users    15010 Sep 17 16:13 xmlenc-0.52.jar
-rw-r--r-- 1 hadoop users    94672 Sep 17 16:27 xz-1.0.jar
-rw-r--r-- 1 hadoop users   779974 Sep 17 16:14 zookeeper-3.4.5.jar

Best regards,
Henry

-----Original Message-----
From: MA11 YTHung1
Sent: Thursday, November 21, 2013 9:51 AM
To: [email protected]
Subject: RE: hbase 0.96 stop master receive ERROR ipc.RPC: RPC.stopProxy called 
on non proxy.

I'm using hadoop-2.2.0 stable

-----Original Message-----
From: Jimmy Xiang [mailto:[email protected]]
Sent: Thursday, November 21, 2013 9:49 AM
To: user
Subject: Re: hbase 0.96 stop master receive ERROR ipc.RPC: RPC.stopProxy called 
on non proxy.

Which version of Hadoop do you use?


On Wed, Nov 20, 2013 at 5:43 PM, Henry Hung <[email protected]> wrote:

> Hi All,
>
> When stopping master or regionserver, I found some ERROR and WARN in
> the log files, are these errors can cause problem in hbase:
>
> 13/11/21 09:31:16 INFO zookeeper.ClientCnxn: EventThread shut down
> 13/11/21 09:35:36 ERROR ipc.RPC: RPC.stopProxy called on non proxy.
> java.lang.IllegalArgumentException: object is not an instance of
> declaring class
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>         at java.lang.reflect.Method.invoke(Method.java:597)
>         at
> org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:266)
>         at $Proxy18.close(Unknown Source)
>         at org.apache.hadoop.ipc.RPC.stopProxy(RPC.java:621)
>         at
> org.apache.hadoop.hdfs.DFSClient.closeConnectionToNamenode(DFSClient.java:738)
>         at org.apache.hadoop.hdfs.DFSClient.close(DFSClient.java:794)
>         at
> org.apache.hadoop.hdfs.DistributedFileSystem.close(DistributedFileSystem.java:847)
>         at
> org.apache.hadoop.fs.FileSystem$Cache.closeAll(FileSystem.java:2524)
>         at
> org.apache.hadoop.fs.FileSystem$Cache$ClientFinalizer.run(FileSystem.java:2541)
>         at
> org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.j
> ava:54)
> 13/11/21 09:35:36 WARN util.ShutdownHookManager: ShutdownHook
> 'ClientFinalizer' failed, org.apache.hadoop.HadoopIllegalArgumentException:
> Cannot close proxy - is not Closeable or does not provide closeable
> invocation handler class $Proxy18
> org.apache.hadoop.HadoopIllegalArgumentException: Cannot close proxy -
> is not Closeable or does not provide closeable invocation handler
> class
> $Proxy18
>         at org.apache.hadoop.ipc.RPC.stopProxy(RPC.java:639)
>         at
> org.apache.hadoop.hdfs.DFSClient.closeConnectionToNamenode(DFSClient.java:738)
>         at org.apache.hadoop.hdfs.DFSClient.close(DFSClient.java:794)
>         at
> org.apache.hadoop.hdfs.DistributedFileSystem.close(DistributedFileSystem.java:847)
>         at
> org.apache.hadoop.fs.FileSystem$Cache.closeAll(FileSystem.java:2524)
>         at
> org.apache.hadoop.fs.FileSystem$Cache$ClientFinalizer.run(FileSystem.java:2541)
>         at
> org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.j
> ava:54)
>
> Best regards,
> Henry
>
> ________________________________
> The privileged confidential information contained in this email is
> intended for use only by the addressees as indicated by the original
> sender of this email. If you are not the addressee indicated in this
> email or are not responsible for delivery of the email to such a
> person, please kindly reply to the sender indicating this fact and
> delete all copies of it from your computer and network server
> immediately. Your cooperation is highly appreciated. It is advised
> that any unauthorized use of confidential information of Winbond is
> strictly prohibited; and any information in this email irrelevant to
> the official business of Winbond shall be deemed as neither given nor 
> endorsed by Winbond.
>

The privileged confidential information contained in this email is intended for 
use only by the addressees as indicated by the original sender of this email. 
If you are not the addressee indicated in this email or are not responsible for 
delivery of the email to such a person, please kindly reply to the sender 
indicating this fact and delete all copies of it from your computer and network 
server immediately. Your cooperation is highly appreciated. It is advised that 
any unauthorized use of confidential information of Winbond is strictly 
prohibited; and any information in this email irrelevant to the official 
business of Winbond shall be deemed as neither given nor endorsed by Winbond.

The privileged confidential information contained in this email is intended for 
use only by the addressees as indicated by the original sender of this email. 
If you are not the addressee indicated in this email or are not responsible for 
delivery of the email to such a person, please kindly reply to the sender 
indicating this fact and delete all copies of it from your computer and network 
server immediately. Your cooperation is highly appreciated. It is advised that 
any unauthorized use of confidential information of Winbond is strictly 
prohibited; and any information in this email irrelevant to the official 
business of Winbond shall be deemed as neither given nor endorsed by Winbond.

The privileged confidential information contained in this email is intended for 
use only by the addressees as indicated by the original sender of this email. 
If you are not the addressee indicated in this email or are not responsible for 
delivery of the email to such a person, please kindly reply to the sender 
indicating this fact and delete all copies of it from your computer and network 
server immediately. Your cooperation is highly appreciated. It is advised that 
any unauthorized use of confidential information of Winbond is strictly 
prohibited; and any information in this email irrelevant to the official 
business of Winbond shall be deemed as neither given nor endorsed by Winbond.

Reply via email to