I've noticed this too. I haven't had a chance to track down what's causing
it yet. I've seen it in Spark tests, so it looks like there may be a
problem that affects both. Probably a connection leak in the common code.

On Wed, Jan 6, 2021 at 3:44 PM Steven Wu <stevenz...@gmail.com> wrote:

> I have noticed some flakiness with Flink and Spark tests both locally and
> in CI checks. @zhangjun0x01 also reported the same problem with
> iceberg-spark3-extensions.  Below is a full stack trace from a local run
> for Flink tests.
>
> The flakiness might be recent regression, as the tests were stable for me
> until recently. Any recent hive dep change? Anyone have any ideas?
>
> org.apache.iceberg.flink.source.TestIcebergSourceReaderDeletes >
> testMixedPositionAndEqualityDeletes[fileFormat=ORC] FAILED
>
>     java.lang.RuntimeException: Failed to get table info from metastore
> default.test
>
>         at
> org.apache.iceberg.hive.HiveTableOperations.doRefresh(HiveTableOperations.java:142)
>
>         at
> org.apache.iceberg.BaseMetastoreTableOperations.refresh(BaseMetastoreTableOperations.java:86)
>
>         at
> org.apache.iceberg.BaseMetastoreTableOperations.current(BaseMetastoreTableOperations.java:69)
>
>         at
> org.apache.iceberg.BaseMetastoreCatalog.loadTable(BaseMetastoreCatalog.java:92)
>
>         at
> org.apache.iceberg.flink.TableLoader$CatalogTableLoader.loadTable(TableLoader.java:113)
>
>         at
> org.apache.iceberg.flink.source.TestIcebergSourceReaderDeletes.rowSet(TestIcebergSourceReaderDeletes.java:90)
>
>
>         Caused by:
>
>         org.apache.thrift.transport.TTransportException:
> java.net.SocketException: Broken pipe (Write failed)
>
>             at
> org.apache.thrift.transport.TIOStreamTransport.flush(TIOStreamTransport.java:161)
>
>             at
> org.apache.thrift.TServiceClient.sendBase(TServiceClient.java:73)
>
>             at
> org.apache.thrift.TServiceClient.sendBase(TServiceClient.java:62)
>
>             at
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.send_get_table_req(ThriftHiveMetastore.java:1561)
>
>             at
> org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.get_table_req(ThriftHiveMetastore.java:1553)
>
>             at
> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getTable(HiveMetaStoreClient.java:1350)
>
>             at
> org.apache.iceberg.hive.HiveTableOperations.lambda$doRefresh$0(HiveTableOperations.java:130)
>
>             at org.apache.iceberg.hive.ClientPool.run(ClientPool.java:65)
>
>             at
> org.apache.iceberg.hive.HiveTableOperations.doRefresh(HiveTableOperations.java:130)
>
>             ... 5 more
>
>
>             Caused by:
>
>             java.net.SocketException: Broken pipe (Write failed)
>
>                 at java.net.SocketOutputStream.socketWrite0(Native Method)
>
>                 at
> java.net.SocketOutputStream.socketWrite(SocketOutputStream.java:111)
>
>                 at
> java.net.SocketOutputStream.write(SocketOutputStream.java:155)
>
>                 at
> java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
>
>                 at
> java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
>
>                 at
> org.apache.thrift.transport.TIOStreamTransport.flush(TIOStreamTransport.java:159)
>
>                 ... 13 more
>


-- 
Ryan Blue
Software Engineer
Netflix

Reply via email to