cozos commented on issue #21598:
URL: https://github.com/apache/beam/issues/21598#issuecomment-1331619714
Also running into this with the Spark RDD Runner on the Python SDK Harness:
```
22/11/30 03:35:04 INFO Executor: Finished task 505.0 in stage 8.0 (TID
1145). 18572 bytes result sent to driver
22/11/30 03:35:55 INFO Executor: Finished task 434.0 in stage 8.0 (TID
1074). 18572 bytes result sent to driver
E1130 03:36:33.423173032 2378 chttp2_transport.cc:1016]
ipv4:127.0.0.1:38291: Received a GOAWAY with error code ENHANCE_YOUR_CALM and
debug data equal to "too_many_pings". Current keepalive time (before
throttling): 20000ms
22/11/30 03:36:33 ERROR py:641: Failed to read inputs in the data plane.
Traceback (most recent call last):
File
"/databricks/python3/lib/python3.7/site-packages/apache_beam/runners/worker/data_plane.py",
line 634, in _read_inputs
for elements in elements_iterator:
File "/databricks/python3/lib/python3.7/site-packages/grpc/_channel.py",
line 426, in __next__
return self._next()
File "/databricks/python3/lib/python3.7/site-packages/grpc/_channel.py",
line 826, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC
that terminated with:
status = StatusCode.UNAVAILABLE
details = "Socket closed"
debug_error_string = "UNKNOWN:Error received from peer
ipv4:127.0.0.1:38291 {grpc_message:"Socket closed", grpc_status:14,
created_time:"2022-11-30T03:36:33.423534531+00:00"}"
> Traceback (most recent call last):
File
"/databricks/python3/lib/python3.7/site-packages/apache_beam/runners/worker/data_plane.py",
line 634, in _read_inputs
for elements in elements_iterator:
File "/databricks/python3/lib/python3.7/site-packages/grpc/_channel.py",
line 426, in __next__
return self._next()
File "/databricks/python3/lib/python3.7/site-packages/grpc/_channel.py",
line 826, in _next
raise self
grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC
that terminated with:
status = StatusCode.UNAVAILABLE
details = "Socket closed"
debug_error_string = "UNKNOWN:Error received from peer
ipv4:127.0.0.1:38291 {grpc_message:"Socket closed", grpc_status:14,
created_time:"2022-11-30T03:36:33.423534531+00:00"}"
```
I do have a shuffle/GroupBy, but I don't understand why that would cause
this. As I understand it the SDK Harness is only used to execute DoFns, and the
GroupBy/shuffle is done on the Spark side, which shouldn't affect the SDK
Harness.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]