[ 
https://issues.apache.org/jira/browse/HDDS-1700?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anu Engineer resolved HDDS-1700.
--------------------------------
    Resolution: Cannot Reproduce

> RPC Payload too large on datanode startup in kubernetes
> -------------------------------------------------------
>
>                 Key: HDDS-1700
>                 URL: https://issues.apache.org/jira/browse/HDDS-1700
>             Project: Hadoop Distributed Data Store
>          Issue Type: Bug
>          Components: docker, Ozone Datanode, SCM
>    Affects Versions: 0.4.0
>         Environment: datanode pod's ozone-site.xml
> {code:java}
> <configuration>
> <property><name>ozone.scm.block.client.address</name><value>ozone-managers-service:9876</value></property>
> <property><name>ozone.enabled</name><value>True</value></property>
> <property><name>ozone.scm.datanode.id</name><value>/tmp/datanode.id</value></property>
> <property><name>ozone.scm.client.address</name><value>ozone-managers-service:9876</value></property>
> <property><name>ozone.metadata.dirs</name><value>/tmp/metadata</value></property>
> <property><name>ozone.scm.names</name><value>ozone-managers-service:9876</value></property>
> <property><name>ozone.om.address</name><value>ozone-managers-service:9874</value></property>
> <property><name>ozone.handler.type</name><value>distributed</value></property>
> <property><name>ozone.scm.datanode.address</name><value>ozone-managers-service:9876</value></property>
> </configuration>
> {code}
> OM/SCM pod's ozone-site.xml
> {code:java}
> <configuration>
> <property><name>ozone.scm.block.client.address</name><value>localhost</value></property>
> <property><name>ozone.enabled</name><value>True</value></property>
> <property><name>ozone.scm.datanode.id</name><value>/tmp/datanode.id</value></property>
> <property><name>ozone.scm.client.address</name><value>localhost</value></property>
> <property><name>ozone.metadata.dirs</name><value>/tmp/metadata</value></property>
> <property><name>ozone.scm.names</name><value>localhost</value></property>
> <property><name>ozone.om.address</name><value>localhost</value></property>
> <property><name>ozone.handler.type</name><value>distributed</value></property>
> <property><name>ozone.scm.datanode.address</name><value>localhost</value></property>
> </configuration>
> {code}
>  
>  
>            Reporter: Josh Siegel
>            Priority: Minor
>
> When starting the datanode on a seperate kubernetes pod than the SCM and OM, 
> the below error appears in the datanode's {{ozone.log}}. We verified basic 
> connectivity between the datanode pod and the OM/SCM pod.
> {code:java}
> 2019-06-17 17:14:16,449 [Datanode State Machine Thread - 0] ERROR 
> (EndpointStateMachine.java:207) - Unable to communicate to SCM server at 
> ozone-managers-service:9876 for past 31800 seconds.
> java.io.IOException: Failed on local exception: 
> org.apache.hadoop.ipc.RpcException: RPC response exceeds maximum data length; 
> Host Details : local host is: "ozone-datanode/10.244.84.187"; destination 
> host is: "ozone-managers-service":9876;
> at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:816)
> at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1515)
> at org.apache.hadoop.ipc.Client.call(Client.java:1457)
> at org.apache.hadoop.ipc.Client.call(Client.java:1367)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:228)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
> at com.sun.proxy.$Proxy88.getVersion(Unknown Source)
> at 
> org.apache.hadoop.ozone.protocolPB.StorageContainerDatanodeProtocolClientSideTranslatorPB.getVersion(StorageContainerDatanodeProtocolClientSideTranslatorPB.java:112)
> at 
> org.apache.hadoop.ozone.container.common.states.endpoint.VersionEndpointTask.call(VersionEndpointTask.java:70)
> at 
> org.apache.hadoop.ozone.container.common.states.endpoint.VersionEndpointTask.call(VersionEndpointTask.java:42)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748)
> Caused by: org.apache.hadoop.ipc.RpcException: RPC response exceeds maximum 
> data length
> at org.apache.hadoop.ipc.Client$IpcStreams.readResponse(Client.java:1830)
> at 
> org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1173)
> at org.apache.hadoop.ipc.Client$Connection.run(Client.java:1069){code}
>  
> cc [~slamendola2_bloomberg]
> [~anu]
> [~elek]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org

Reply via email to