The following JIRA has been integrated to branch 2.2 :
HADOOP-10132 RPC#stopProxy() should log the class of proxy when
IllegalArgumentException is encountered

FYI


On Mon, Nov 25, 2013 at 9:56 PM, Ted Yu <[email protected]> wrote:

> Update:
> Henry tried my patch attached to HBASE-10029
>
> From master log, it seems my patch worked.
>
> I will get back to this thread after further testing / code review.
>
> Cheers
>
> On Nov 25, 2013, at 6:05 PM, Henry Hung <[email protected]> wrote:
>
> > @Ted:
> >
> > I create the JIRA, is the information sufficient?
> > https://issues.apache.org/jira/browse/HBASE-10029
> >
> > Best regards,
> > Henry
> >
> > -----Original Message-----
> > From: Ted Yu [mailto:[email protected]]
> > Sent: Tuesday, November 26, 2013 9:30 AM
> > To: [email protected]
> > Subject: Re: hbase 0.96 stop master receive ERROR ipc.RPC: RPC.stopProxy
> called on non proxy.
> >
> > Henry:
> > Thanks for the additional information.
> >
> > Looks like HA namenode with QJM is not covered by current code.
> >
> > Mind filing a JIRA with summary of this thread ?
> >
> > Cheers
> >
> >
> > On Tue, Nov 26, 2013 at 9:12 AM, Henry Hung <[email protected]> wrote:
> >
> >> @Ted
> >> Yes, I use the hadoop-hdfs-2.2.0.jar.
> >>
> >> BTW, how do you certain that the namenode class is
> >> ClientNamenodeProtocolTranslatorPB?
> >>
> >> From the NameNodeProxies, I can only assume the
> >> ClientNamenodeProtocolTranslatorPB is used only when connecting to
> >> single hadoop namenode.
> >>
> >>  public static <T> ProxyAndInfo<T> createNonHAProxy(
> >>      Configuration conf, InetSocketAddress nnAddr, Class<T> xface,
> >>      UserGroupInformation ugi, boolean withRetries) throws IOException {
> >>    Text dtService = SecurityUtil.buildTokenService(nnAddr);
> >>
> >>    T proxy;
> >>    if (xface == ClientProtocol.class) {
> >>      proxy = (T) createNNProxyWithClientProtocol(nnAddr, conf, ugi,
> >>          withRetries);
> >>
> >>
> >> But I'm using HA configuration using QJM, so the my guess is the
> >> createProxy will go to the HA case because I provide
> >> failoverProxyProviderClass with
> >>
> "org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider".
> >>
> >>  public static <T> ProxyAndInfo<T> createProxy(Configuration conf,
> >>      URI nameNodeUri, Class<T> xface) throws IOException {
> >>    Class<FailoverProxyProvider<T>> failoverProxyProviderClass =
> >>        getFailoverProxyProviderClass(conf, nameNodeUri, xface);
> >>
> >>    if (failoverProxyProviderClass == null) {
> >>      // Non-HA case
> >>      return createNonHAProxy(conf, NameNode.getAddress(nameNodeUri),
> >> xface,
> >>          UserGroupInformation.getCurrentUser(), true);
> >>    } else {
> >>      // HA case
> >>      FailoverProxyProvider<T> failoverProxyProvider = NameNodeProxies
> >>          .createFailoverProxyProvider(conf,
> >> failoverProxyProviderClass, xface,
> >>              nameNodeUri);
> >>      Conf config = new Conf(conf);
> >>      T proxy = (T) RetryProxy.create(xface, failoverProxyProvider,
> >> RetryPolicies
> >>          .failoverOnNetworkException(RetryPolicies.TRY_ONCE_THEN_FAIL,
> >>              config.maxFailoverAttempts, config.failoverSleepBaseMillis,
> >>              config.failoverSleepMaxMillis));
> >>
> >>      Text dtService =
> HAUtil.buildTokenServiceForLogicalUri(nameNodeUri);
> >>      return new ProxyAndInfo<T>(proxy, dtService);
> >>    }
> >>  }
> >>
> >> Here is the snippet of my hdfs-site.xml:
> >>
> >>  <property>
> >>    <name>dfs.nameservices</name>
> >>    <value>hadoopdev</value>
> >>  </property>
> >>  <property>
> >>    <name>dfs.ha.namenodes.hadoopdev</name>
> >>    <value>nn1,nn2</value>
> >>  </property>
> >>  <property>
> >>    <name>dfs.namenode.rpc-address.hadoopdev.nn1</name>
> >>    <value>fphd9.ctpilot1.com:9000</value>
> >>  </property>
> >>  <property>
> >>    <name>dfs.namenode.http-address.hadoopdev.nn1</name>
> >>    <value>fphd9.ctpilot1.com:50070</value>
> >>  </property>
> >>  <property>
> >>    <name>dfs.namenode.rpc-address.hadoopdev.nn2</name>
> >>    <value>fphd10.ctpilot1.com:9000</value>
> >>  </property>
> >>  <property>
> >>    <name>dfs.namenode.http-address.hadoopdev.nn2</name>
> >>    <value>fphd10.ctpilot1.com:50070</value>
> >>  </property>
> >>  <property>
> >>    <name>dfs.namenode.shared.edits.dir</name>
> >>    <value>qjournal://fphd8.ctpilot1.com:8485;fphd9.ctpilot1.com:8485;
> >> fphd10.ctpilot1.com:8485/hadoopdev</value>
> >>  </property>
> >>  <property>
> >>    <name>dfs.client.failover.proxy.provider.hadoopdev</name>
> >>
> >>
> <value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
> >>  </property>
> >>  <property>
> >>    <name>dfs.ha.fencing.methods</name>
> >>    <value>shell(/bin/true)</value>
> >>  </property>
> >>  <property>
> >>    <name>dfs.journalnode.edits.dir</name>
> >>    <value>/data/hadoop/hadoop-data-2/journal</value>
> >>  </property>
> >>  <property>
> >>    <name>dfs.ha.automatic-failover.enabled</name>
> >>    <value>true</value>
> >>  </property>
> >>  <property>
> >>    <name>ha.zookeeper.quorum</name>
> >>    <value>fphd1.ctpilot1.com:2222</value>
> >>  </property>
> >>
> >> -----Original Message-----
> >> From: Ted Yu [mailto:[email protected]]
> >> Sent: Tuesday, November 26, 2013 1:56 AM
> >> To: [email protected]
> >> Subject: Re: hbase 0.96 stop master receive ERROR ipc.RPC:
> >> RPC.stopProxy called on non proxy.
> >>
> >> Here is the caller to createReorderingProxy():
> >>
> >>      ClientProtocol cp1 = createReorderingProxy(namenode, lrb, conf);
> >>
> >> where namenode
> >> is org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB
> :
> >>
> >> public class ClientNamenodeProtocolTranslatorPB implements
> >>
> >>    ProtocolMetaInterface, ClientProtocol, Closeable,
> >> ProtocolTranslator {
> >>
> >> In createReorderingProxy() :
> >>
> >>            new Class[]{ClientProtocol.class, Closeable.class},
> >>
> >> We ask for Closeable interface.
> >>
> >>
> >> Did the error persist after you replaced with the hadoop-hdfs-2.2.0.jar
> ?
> >> Meaning, did you start HBase using the new hadoop jars ?
> >>
> >> Cheers
> >>
> >>
> >> On Mon, Nov 25, 2013 at 1:04 PM, Henry Hung <[email protected]>
> wrote:
> >>
> >>> I looked into the source code of
> >>> org/apache/hadoop/hbase/fs/HFileSystem.java
> >>> and whenever I execute hbase-daemon.sh stop master (or
> >>> regionserver), the
> >>> method.getName() is "close",
> >>> but the org/apache/hadoop/hdfs/protocol/ClientProtocol.java do not
> >>> have method named "close", thus it result in error "object is not an
> >>> instance of declaring class"
> >>>
> >>> Could someone that familiar with hbase-0.96.0 hadoop2 can tell me if
> >>> this problem need to be fixed? And how to fix it?
> >>>
> >>>  private static ClientProtocol createReorderingProxy(final
> >>> ClientProtocol cp, final ReorderBlocks lrb, final Configuration conf) {
> >>>    return (ClientProtocol) Proxy.newProxyInstance
> >>>        (cp.getClass().getClassLoader(),
> >>>            new Class[]{ClientProtocol.class, Closeable.class},
> >>>            new InvocationHandler() {
> >>>              public Object invoke(Object proxy, Method method,
> >>>                                   Object[] args) throws Throwable {
> >>>                try {
> >>>                  // method.invoke will failed if
> >>> method.getName().equals("close")
> >>>                  // because ClientProtocol do not have method "close"
> >>>                  Object res = method.invoke(cp, args);
> >>>                  if (res != null && args != null && args.length == 3
> >>>                      && "getBlockLocations".equals(method.getName())
> >>>                      && res instanceof LocatedBlocks
> >>>                      && args[0] instanceof String
> >>>                      && args[0] != null) {
> >>>                    lrb.reorderBlocks(conf, (LocatedBlocks) res,
> >>> (String) args[0]);
> >>>                  }
> >>>                  return res;
> >>>                } catch  (InvocationTargetException ite) {
> >>>                  // We will have this for all the exception,
> >>> checked on not, sent
> >>>                  //  by any layer, including the functional exception
> >>>                  Throwable cause = ite.getCause();
> >>>                  if (cause == null){
> >>>                    throw new RuntimeException(
> >>>                      "Proxy invocation failed and getCause is
> >>> null",
> >> ite);
> >>>                  }
> >>>                  if (cause instanceof UndeclaredThrowableException) {
> >>>                    Throwable causeCause = cause.getCause();
> >>>                    if (causeCause == null) {
> >>>                      throw new
> >>> RuntimeException("UndeclaredThrowableException had null cause!");
> >>>                    }
> >>>                    cause = cause.getCause();
> >>>                  }
> >>>                  throw cause;
> >>>                }
> >>>              }
> >>>            });
> >>>  }
> >>>
> >>>
> >>>
> >>> -----Original Message-----
> >>> From: MA11 YTHung1
> >>> Sent: Thursday, November 21, 2013 9:57 AM
> >>> To: [email protected]
> >>> Subject: RE: hbase 0.96 stop master receive ERROR ipc.RPC:
> >>> RPC.stopProxy called on non proxy.
> >>>
> >>> Additional information:
> >>>
> >>> I replace all files with prefix hadoop in hbase-0.96.0-hadoop2/lib
> >>> with
> >>> hadoop-2.2.0 libraries.
> >>>
> >>> the ls -l of hbase-0.96.0-hadoop2/lib as below:
> >>>
> >>> -rw-r--r-- 1 hadoop users    62983 Sep 17 16:13 activation-1.1.jar
> >>> -rw-r--r-- 1 hadoop users     4467 Sep 17 23:29 aopalliance-1.0.jar
> >>> -rw-r--r-- 1 hadoop users    43033 Sep 17 16:13 asm-3.1.jar
> >>> -rw-r--r-- 1 hadoop users   263268 Sep 17 16:27 avro-1.5.3.jar
> >>> -rw-r--r-- 1 hadoop users   188671 Sep 17 16:12
> >> commons-beanutils-1.7.0.jar
> >>> -rw-r--r-- 1 hadoop users   206035 Sep 17 16:13
> >>> commons-beanutils-core-1.8.0.jar
> >>> -rw-r--r-- 1 hadoop users    41123 Sep 17 16:12 commons-cli-1.2.jar
> >>> -rw-r--r-- 1 hadoop users   259600 Sep 17 16:13 commons-codec-1.7.jar
> >>> -rw-r--r-- 1 hadoop users   575389 Sep 17 16:12
> >>> commons-collections-3.2.1.jar
> >>> -rw-r--r-- 1 hadoop users   238681 Sep 17 16:27
> commons-compress-1.4.jar
> >>> -rw-r--r-- 1 hadoop users   298829 Sep 17 16:13
> >>> commons-configuration-1.6.jar
> >>> -rw-r--r-- 1 hadoop users    24239 Sep 17 23:28
> commons-daemon-1.0.13.jar
> >>> -rw-r--r-- 1 hadoop users   143602 Sep 17 16:12
> commons-digester-1.8.jar
> >>> -rw-r--r-- 1 hadoop users   112341 Sep 17 16:13 commons-el-1.0.jar
> >>> -rw-r--r-- 1 hadoop users   305001 Sep 17 16:12
> >> commons-httpclient-3.1.jar
> >>> -rw-r--r-- 1 hadoop users   185140 Sep 17 16:13 commons-io-2.4.jar
> >>> -rw-r--r-- 1 hadoop users   284220 Sep 17 16:12 commons-lang-2.6.jar
> >>> -rw-r--r-- 1 hadoop users    60686 Sep 17 16:12
> commons-logging-1.1.1.jar
> >>> -rw-r--r-- 1 hadoop users   988514 Sep 17 16:13 commons-math-2.2.jar
> >>> -rw-r--r-- 1 hadoop users   273370 Sep 17 16:27 commons-net-3.1.jar
> >>> -rw-r--r-- 1 hadoop users  3566844 Sep 17 16:15 core-3.1.1.jar
> >>> -rw-r--r-- 1 hadoop users    15322 Sep 17 16:12
> >>> findbugs-annotations-1.3.9-1.jar
> >>> -rw-r--r-- 1 hadoop users    21817 Sep 17 23:29
> >>> gmbal-api-only-3.0.0-b023.jar
> >>> -rw-r--r-- 1 hadoop users   684337 Sep 17 23:29
> >> grizzly-framework-2.1.1.jar
> >>> -rw-r--r-- 1 hadoop users   210846 Sep 17 23:29
> >>> grizzly-framework-2.1.1-tests.jar
> >>> -rw-r--r-- 1 hadoop users   248346 Sep 17 23:29 grizzly-http-2.1.1.jar
> >>> -rw-r--r-- 1 hadoop users   193583 Sep 17 23:29
> >>> grizzly-http-server-2.1.1.jar
> >>> -rw-r--r-- 1 hadoop users   336878 Sep 17 23:29
> >>> grizzly-http-servlet-2.1.1.jar
> >>> -rw-r--r-- 1 hadoop users     8072 Sep 17 23:29 grizzly-rcm-2.1.1.jar
> >>> -rw-r--r-- 1 hadoop users  1795932 Sep 17 16:13 guava-12.0.1.jar
> >>> -rw-r--r-- 1 hadoop users   710492 Sep 17 23:29 guice-3.0.jar
> >>> -rw-r--r-- 1 hadoop users    65012 Sep 17 23:29 guice-servlet-3.0.jar
> >>> -rw-r--r-- 1 hadoop users    16778 Nov 20 17:39
> >>> hadoop-annotations-2.2.0.jar
> >>> -rw-r--r-- 1 hadoop users    49750 Nov 20 17:40 hadoop-auth-2.2.0.jar
> >>> -rw-r--r-- 1 hadoop users     2576 Oct 12 06:20
> >>> hadoop-client-2.1.0-beta.jar
> >>> -rw-r--r-- 1 hadoop users  2735584 Nov 20 17:50
> >>> hadoop-common-2.2.0.jar
> >>> -rw-r--r-- 1 hadoop users  5242252 Nov 21 08:48
> >>> hadoop-hdfs-2.2.0.jar
> >>> -rw-r--r-- 1 hadoop users  1988460 Nov 21 08:48
> >> hadoop-hdfs-2.2.0-tests.jar
> >>> -rw-r--r-- 1 hadoop users   482042 Nov 21 08:49
> >>> hadoop-mapreduce-client-app-2.2.0.jar
> >>> -rw-r--r-- 1 hadoop users   656365 Nov 21 08:49
> >>> hadoop-mapreduce-client-common-2.2.0.jar
> >>> -rw-r--r-- 1 hadoop users  1455001 Nov 21 08:50
> >>> hadoop-mapreduce-client-core-2.2.0.jar
> >>> -rw-r--r-- 1 hadoop users    35216 Nov 21 08:50
> >>> hadoop-mapreduce-client-jobclient-2.2.0.jar
> >>> -rw-r--r-- 1 hadoop users  1434852 Nov 21 08:50
> >>> hadoop-mapreduce-client-jobclient-2.2.0-tests.jar
> >>> -rw-r--r-- 1 hadoop users    21537 Nov 21 08:51
> >>> hadoop-mapreduce-client-shuffle-2.2.0.jar
> >>> -rw-r--r-- 1 hadoop users  1158936 Nov 21 08:51
> hadoop-yarn-api-2.2.0.jar
> >>> -rw-r--r-- 1 hadoop users    94728 Nov 21 08:51
> >>> hadoop-yarn-client-2.2.0.jar
> >>> -rw-r--r-- 1 hadoop users  1301627 Nov 21 08:51
> >>> hadoop-yarn-common-2.2.0.jar
> >>> -rw-r--r-- 1 hadoop users   175554 Nov 21 08:52
> >>> hadoop-yarn-server-common-2.2.0.jar
> >>> -rw-r--r-- 1 hadoop users   467638 Nov 21 08:52
> >>> hadoop-yarn-server-nodemanager-2.2.0.jar
> >>> -rw-r--r-- 1 hadoop users   825853 Oct 12 06:28
> >>> hbase-client-0.96.0-hadoop2.jar
> >>> -rw-r--r-- 1 hadoop users   354845 Oct 12 06:28
> >>> hbase-common-0.96.0-hadoop2.jar
> >>> -rw-r--r-- 1 hadoop users   132690 Oct 12 06:28
> >>> hbase-common-0.96.0-hadoop2-tests.jar
> >>> -rw-r--r-- 1 hadoop users    97428 Oct 12 06:28
> >>> hbase-examples-0.96.0-hadoop2.jar
> >>> -rw-r--r-- 1 hadoop users    72765 Oct 12 06:28
> >>> hbase-hadoop2-compat-0.96.0-hadoop2.jar
> >>> -rw-r--r-- 1 hadoop users    32096 Oct 12 06:28
> >>> hbase-hadoop-compat-0.96.0-hadoop2.jar
> >>> -rw-r--r-- 1 hadoop users    12174 Oct 12 06:28
> >> hbase-it-0.96.0-hadoop2.jar
> >>> -rw-r--r-- 1 hadoop users   288784 Oct 12 06:28
> >>> hbase-it-0.96.0-hadoop2-tests.jar
> >>> -rw-r--r-- 1 hadoop users    94784 Oct 12 06:28
> >>> hbase-prefix-tree-0.96.0-hadoop2.jar
> >>> -rw-r--r-- 1 hadoop users  3134214 Oct 12 06:28
> >>> hbase-protocol-0.96.0-hadoop2.jar
> >>> -rw-r--r-- 1 hadoop users  3058804 Oct 12 06:28
> >>> hbase-server-0.96.0-hadoop2.jar
> >>> -rw-r--r-- 1 hadoop users  3150292 Oct 12 06:28
> >>> hbase-server-0.96.0-hadoop2-tests.jar
> >>> -rw-r--r-- 1 hadoop users    12554 Oct 12 06:28
> >>> hbase-shell-0.96.0-hadoop2.jar
> >>> -rw-r--r-- 1 hadoop users    10941 Oct 12 06:28
> >>> hbase-testing-util-0.96.0-hadoop2.jar
> >>> -rw-r--r-- 1 hadoop users  2276333 Oct 12 06:28
> >>> hbase-thrift-0.96.0-hadoop2.jar
> >>> -rw-r--r-- 1 hadoop users    95975 Sep 17 16:15
> high-scale-lib-1.1.1.jar
> >>> -rw-r--r-- 1 hadoop users    31020 Sep 17 16:14 htrace-core-2.01.jar
> >>> -rw-r--r-- 1 hadoop users   352585 Sep 17 16:15 httpclient-4.1.3.jar
> >>> -rw-r--r-- 1 hadoop users   181201 Sep 17 16:15 httpcore-4.1.3.jar
> >>> -rw-r--r-- 1 hadoop users   227517 Sep 17 16:13
> >> jackson-core-asl-1.8.8.jar
> >>> -rw-r--r-- 1 hadoop users    17884 Sep 17 16:13 jackson-jaxrs-1.8.8.jar
> >>> -rw-r--r-- 1 hadoop users   669065 Sep 17 16:13
> >>> jackson-mapper-asl-1.8.8.jar
> >>> -rw-r--r-- 1 hadoop users    32353 Sep 17 16:13 jackson-xc-1.8.8.jar
> >>> -rw-r--r-- 1 hadoop users    20642 Sep 17 16:15 jamon-runtime-2.3.1.jar
> >>> -rw-r--r-- 1 hadoop users   408133 Sep 17 16:13
> >> jasper-compiler-5.5.23.jar
> >>> -rw-r--r-- 1 hadoop users    76844 Sep 17 16:13
> jasper-runtime-5.5.23.jar
> >>> -rw-r--r-- 1 hadoop users     2497 Sep 17 23:29 javax.inject-1.jar
> >>> -rw-r--r-- 1 hadoop users    83586 Sep 17 23:29 javax.servlet-3.0.jar
> >>> -rw-r--r-- 1 hadoop users   105134 Sep 17 16:27 jaxb-api-2.2.2.jar
> >>> -rw-r--r-- 1 hadoop users   890168 Sep 17 16:13 jaxb-impl-2.2.3-1.jar
> >>> -rw-r--r-- 1 hadoop users   129217 Sep 17 23:29 jersey-client-1.8.jar
> >>> -rw-r--r-- 1 hadoop users   458233 Sep 17 16:13 jersey-core-1.8.jar
> >>> -rw-r--r-- 1 hadoop users    17585 Sep 17 23:29 jersey-grizzly2-1.8.jar
> >>> -rw-r--r-- 1 hadoop users    14712 Sep 17 23:29 jersey-guice-1.8.jar
> >>> -rw-r--r-- 1 hadoop users   147933 Sep 17 16:13 jersey-json-1.8.jar
> >>> -rw-r--r-- 1 hadoop users   694352 Sep 17 16:13 jersey-server-1.8.jar
> >>> -rw-r--r-- 1 hadoop users    28034 Sep 17 23:29
> >>> jersey-test-framework-core-1.8.jar
> >>> -rw-r--r-- 1 hadoop users    12907 Sep 17 23:29
> >>> jersey-test-framework-grizzly2-1.8.jar
> >>> -rw-r--r-- 1 hadoop users   321806 Sep 17 16:27 jets3t-0.6.1.jar
> >>> -rw-r--r-- 1 hadoop users    75963 Sep 17 16:13 jettison-1.3.1.jar
> >>> -rw-r--r-- 1 hadoop users   539912 Sep 17 16:13 jetty-6.1.26.jar
> >>> -rw-r--r-- 1 hadoop users    18891 Sep 17 16:15
> >> jetty-sslengine-6.1.26.jar
> >>> -rw-r--r-- 1 hadoop users   177131 Sep 17 16:13 jetty-util-6.1.26.jar
> >>> -rw-r--r-- 1 hadoop users 13832273 Sep 17 16:15
> jruby-complete-1.6.8.jar
> >>> -rw-r--r-- 1 hadoop users   185746 Sep 17 16:27 jsch-0.1.42.jar
> >>> -rw-r--r-- 1 hadoop users  1024680 Sep 17 16:13 jsp-2.1-6.1.14.jar
> >>> -rw-r--r-- 1 hadoop users   134910 Sep 17 16:13 jsp-api-2.1-6.1.14.jar
> >>> -rw-r--r-- 1 hadoop users   100636 Sep 17 16:27 jsp-api-2.1.jar
> >>> -rw-r--r-- 1 hadoop users    33015 Sep 17 16:13 jsr305-1.3.9.jar
> >>> -rw-r--r-- 1 hadoop users   245039 Sep 17 16:12 junit-4.11.jar
> >>> -rw-r--r-- 1 hadoop users   347531 Sep 17 16:15 libthrift-0.9.0.jar
> >>> -rw-r--r-- 1 hadoop users   489884 Sep 17 16:12 log4j-1.2.17.jar
> >>> -rw-r--r-- 1 hadoop users    42212 Sep 17 23:29
> >>> management-api-3.0.0-b012.jar
> >>> -rw-r--r-- 1 hadoop users    82445 Sep 17 16:14 metrics-core-2.1.2.jar
> >>> drwxr-xr-x 3 hadoop users     4096 Nov 21 09:10 native
> >>> -rw-r--r-- 1 hadoop users  1206119 Sep 18 04:00 netty-3.6.6.Final.jar
> >>> -rw-r--r-- 1 hadoop users    29555 Sep 17 16:27 paranamer-2.3.jar
> >>> -rw-r--r-- 1 hadoop users   533455 Sep 17 16:13 protobuf-java-2.5.0.jar
> >>> drwxr-xr-x 5 hadoop users     4096 Sep 28 10:37 ruby
> >>> -rw-r--r-- 1 hadoop users   132368 Sep 17 16:13
> >> servlet-api-2.5-6.1.14.jar
> >>> -rw-r--r-- 1 hadoop users   105112 Sep 17 16:12 servlet-api-2.5.jar
> >>> -rw-r--r-- 1 hadoop users    25962 Sep 17 16:14 slf4j-api-1.6.4.jar
> >>> -rw-r--r-- 1 hadoop users     9748 Oct  3 07:15 slf4j-log4j12-1.6.4.jar
> >>> -rw-r--r-- 1 hadoop users   995720 Sep 17 16:27 snappy-java-1.0.3.2.jar
> >>> -rw-r--r-- 1 hadoop users    26514 Sep 17 16:13 stax-api-1.0.1.jar
> >>> -rw-r--r-- 1 hadoop users    15010 Sep 17 16:13 xmlenc-0.52.jar
> >>> -rw-r--r-- 1 hadoop users    94672 Sep 17 16:27 xz-1.0.jar
> >>> -rw-r--r-- 1 hadoop users   779974 Sep 17 16:14 zookeeper-3.4.5.jar
> >>>
> >>> Best regards,
> >>> Henry
> >>>
> >>> -----Original Message-----
> >>> From: MA11 YTHung1
> >>> Sent: Thursday, November 21, 2013 9:51 AM
> >>> To: [email protected]
> >>> Subject: RE: hbase 0.96 stop master receive ERROR ipc.RPC:
> >>> RPC.stopProxy called on non proxy.
> >>>
> >>> I'm using hadoop-2.2.0 stable
> >>>
> >>> -----Original Message-----
> >>> From: Jimmy Xiang [mailto:[email protected]]
> >>> Sent: Thursday, November 21, 2013 9:49 AM
> >>> To: user
> >>> Subject: Re: hbase 0.96 stop master receive ERROR ipc.RPC:
> >>> RPC.stopProxy called on non proxy.
> >>>
> >>> Which version of Hadoop do you use?
> >>>
> >>>
> >>> On Wed, Nov 20, 2013 at 5:43 PM, Henry Hung <[email protected]>
> wrote:
> >>>
> >>>> Hi All,
> >>>>
> >>>> When stopping master or regionserver, I found some ERROR and WARN
> >>>> in the log files, are these errors can cause problem in hbase:
> >>>>
> >>>> 13/11/21 09:31:16 INFO zookeeper.ClientCnxn: EventThread shut down
> >>>> 13/11/21 09:35:36 ERROR ipc.RPC: RPC.stopProxy called on non proxy.
> >>>> java.lang.IllegalArgumentException: object is not an instance of
> >>>> declaring class
> >>>>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >>>>        at
> >>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl
> >>> .j
> >>> ava:39)
> >>>>        at
> >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcce
> >>> ss
> >>> orImpl.java:25)
> >>>>        at java.lang.reflect.Method.invoke(Method.java:597)
> >>>>        at
> >>>> org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:266)
> >>>>        at $Proxy18.close(Unknown Source)
> >>>>        at org.apache.hadoop.ipc.RPC.stopProxy(RPC.java:621)
> >>>>        at
> >>> org.apache.hadoop.hdfs.DFSClient.closeConnectionToNamenode(DFSClient
> >>> .j
> >>> ava:738)
> >>>>        at org.apache.hadoop.hdfs.DFSClient.close(DFSClient.java:794)
> >>>>        at
> >>> org.apache.hadoop.hdfs.DistributedFileSystem.close(DistributedFileSy
> >>> st
> >>> em.java:847)
> >>>>        at
> >>>> org.apache.hadoop.fs.FileSystem$Cache.closeAll(FileSystem.java:2524)
> >>>>        at
> >>> org.apache.hadoop.fs.FileSystem$Cache$ClientFinalizer.run(FileSystem
> >>> .j
> >>> ava:2541)
> >>>>        at
> >>>> org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManag
> >>>> er
> >>>> .j
> >>>> ava:54)
> >>>> 13/11/21 09:35:36 WARN util.ShutdownHookManager: ShutdownHook
> >>>> 'ClientFinalizer' failed,
> >>> org.apache.hadoop.HadoopIllegalArgumentException:
> >>>> Cannot close proxy - is not Closeable or does not provide
> >>>> closeable invocation handler class $Proxy18
> >>>> org.apache.hadoop.HadoopIllegalArgumentException: Cannot close
> >>>> proxy
> >>>> - is not Closeable or does not provide closeable invocation
> >>>> handler class
> >>>> $Proxy18
> >>>>        at org.apache.hadoop.ipc.RPC.stopProxy(RPC.java:639)
> >>>>        at
> >>> org.apache.hadoop.hdfs.DFSClient.closeConnectionToNamenode(DFSClient
> >>> .j
> >>> ava:738)
> >>>>        at org.apache.hadoop.hdfs.DFSClient.close(DFSClient.java:794)
> >>>>        at
> >>> org.apache.hadoop.hdfs.DistributedFileSystem.close(DistributedFileSy
> >>> st
> >>> em.java:847)
> >>>>        at
> >>>> org.apache.hadoop.fs.FileSystem$Cache.closeAll(FileSystem.java:2524)
> >>>>        at
> >>> org.apache.hadoop.fs.FileSystem$Cache$ClientFinalizer.run(FileSystem
> >>> .j
> >>> ava:2541)
> >>>>        at
> >>>> org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManag
> >>>> er
> >>>> .j
> >>>> ava:54)
> >>>>
> >>>> Best regards,
> >>>> Henry
> >>>>
> >>>> ________________________________
> >>>> The privileged confidential information contained in this email is
> >>>> intended for use only by the addressees as indicated by the
> >>>> original sender of this email. If you are not the addressee
> >>>> indicated in this email or are not responsible for delivery of the
> >>>> email to such a person, please kindly reply to the sender
> >>>> indicating this fact and delete all copies of it from your
> >>>> computer and network server immediately. Your cooperation is
> >>>> highly appreciated. It is advised that any unauthorized use of
> >>>> confidential information of Winbond is strictly prohibited; and
> >>>> any information in this email irrelevant to the official business
> >>>> of Winbond shall be deemed as neither given nor
> >>> endorsed by Winbond.
> >>>
> >>> The privileged confidential information contained in this email is
> >>> intended for use only by the addressees as indicated by the original
> >>> sender of this email. If you are not the addressee indicated in this
> >>> email or are not responsible for delivery of the email to such a
> >>> person, please kindly reply to the sender indicating this fact and
> >>> delete all copies of it from your computer and network server
> >>> immediately. Your cooperation is highly appreciated. It is advised
> >>> that any unauthorized use of confidential information of Winbond is
> >>> strictly prohibited; and any information in this email irrelevant to
> >>> the official business of Winbond shall be deemed as neither given
> >>> nor
> >> endorsed by Winbond.
> >>>
> >>> The privileged confidential information contained in this email is
> >>> intended for use only by the addressees as indicated by the original
> >>> sender of this email. If you are not the addressee indicated in this
> >>> email or are not responsible for delivery of the email to such a
> >>> person, please kindly reply to the sender indicating this fact and
> >>> delete all copies of it from your computer and network server
> >>> immediately. Your cooperation is highly appreciated. It is advised
> >>> that any unauthorized use of confidential information of Winbond is
> >>> strictly prohibited; and any information in this email irrelevant to
> >>> the official business of Winbond shall be deemed as neither given
> >>> nor
> >> endorsed by Winbond.
> >>>
> >>> The privileged confidential information contained in this email is
> >>> intended for use only by the addressees as indicated by the original
> >>> sender of this email. If you are not the addressee indicated in this
> >>> email or are not responsible for delivery of the email to such a
> >>> person, please kindly reply to the sender indicating this fact and
> >>> delete all copies of it from your computer and network server
> >>> immediately. Your cooperation is highly appreciated. It is advised
> >>> that any unauthorized use of confidential information of Winbond is
> >>> strictly prohibited; and any information in this email irrelevant to
> >>> the official business of Winbond shall be deemed as neither given
> >>> nor
> >> endorsed by Winbond.
> >>
> >> The privileged confidential information contained in this email is
> >> intended for use only by the addressees as indicated by the original
> >> sender of this email. If you are not the addressee indicated in this
> >> email or are not responsible for delivery of the email to such a
> >> person, please kindly reply to the sender indicating this fact and
> >> delete all copies of it from your computer and network server
> >> immediately. Your cooperation is highly appreciated. It is advised
> >> that any unauthorized use of confidential information of Winbond is
> >> strictly prohibited; and any information in this email irrelevant to
> >> the official business of Winbond shall be deemed as neither given nor
> endorsed by Winbond.
> >
> > The privileged confidential information contained in this email is
> intended for use only by the addressees as indicated by the original sender
> of this email. If you are not the addressee indicated in this email or are
> not responsible for delivery of the email to such a person, please kindly
> reply to the sender indicating this fact and delete all copies of it from
> your computer and network server immediately. Your cooperation is highly
> appreciated. It is advised that any unauthorized use of confidential
> information of Winbond is strictly prohibited; and any information in this
> email irrelevant to the official business of Winbond shall be deemed as
> neither given nor endorsed by Winbond.
>

Reply via email to