Hi all, what is this daily summary intended for? Not all issues look like P1. And will a weekly summary be less noise?
<beamacti...@gmail.com>于2022年6月22日 周三23:45写道: > This is your daily summary of Beam's current P1 issues, not including > flaky tests. > > See https://beam.apache.org/contribute/issue-priorities/#p1-critical > for the meaning and expectations around P1 issues. > > > > https://api.github.com/repos/apache/beam/issues/21978: [Playground] > Implement Share Any Code feature on the frontend > https://api.github.com/repos/apache/beam/issues/21946: [Bug]: No way to > read or write to file when running Beam in Flink > https://api.github.com/repos/apache/beam/issues/21935: [Bug]: Reject > illformed GBK Coders > https://api.github.com/repos/apache/beam/issues/21897: [Feature Request]: > Flink runner savepoint backward compatibility > https://api.github.com/repos/apache/beam/issues/21893: [Bug]: BigQuery > Storage Write API implementation does not support table partitioning > https://api.github.com/repos/apache/beam/issues/21794: Dataflow runner > creates a new timer whenever the output timestamp is change > https://api.github.com/repos/apache/beam/issues/21763: [Playground Task]: > Migrate from Google Analytics to Matomo Cloud > https://api.github.com/repos/apache/beam/issues/21715: Data missing when > using CassandraIO.Read > https://api.github.com/repos/apache/beam/issues/21713: 404s in BigQueryIO > don't get output to Failed Inserts PCollection > https://api.github.com/repos/apache/beam/issues/21711: Python Streaming > job failing to drain with BigQueryIO write errors > https://api.github.com/repos/apache/beam/issues/21703: > pubsublite.ReadWriteIT failing in beam_PostCommit_Java_DataflowV1 and V2 > https://api.github.com/repos/apache/beam/issues/21702: SpannerWriteIT > failing in beam PostCommit Java V1 > https://api.github.com/repos/apache/beam/issues/21700: > --dataflowServiceOptions=use_runner_v2 is broken > https://api.github.com/repos/apache/beam/issues/21695: > DataflowPipelineResult does not raise exception for unsuccessful states. > https://api.github.com/repos/apache/beam/issues/21694: BigQuery Storage > API insert with writeResult retry and write to error table > https://api.github.com/repos/apache/beam/issues/21479: Install Python > wheel and dependencies to local venv in SDK harness > https://api.github.com/repos/apache/beam/issues/21478: > KafkaIO.read.withDynamicRead() doesn't pick up new TopicPartitions > https://api.github.com/repos/apache/beam/issues/21477: Add integration > testing for BQ Storage API write modes > https://api.github.com/repos/apache/beam/issues/21476: WriteToBigQuery > Dynamic table destinations returns wrong tableId > https://api.github.com/repos/apache/beam/issues/21475: Beam x-lang > Dataflow tests failing due to _InactiveRpcError > https://api.github.com/repos/apache/beam/issues/21473: > PVR_Spark2_Streaming perma-red > https://api.github.com/repos/apache/beam/issues/21466: Simplify version > override for Dev versions of the Go SDK. > https://api.github.com/repos/apache/beam/issues/21465: Kafka commit > offset drop data on failure for runners that have non-checkpointing shuffle > https://api.github.com/repos/apache/beam/issues/21269: Delete orphaned > files > https://api.github.com/repos/apache/beam/issues/21268: Race between > member variable being accessed due to leaking uninitialized state via > OutboundObserverFactory > https://api.github.com/repos/apache/beam/issues/21267: WriteToBigQuery > submits a duplicate BQ load job if a 503 error code is returned from > googleapi > https://api.github.com/repos/apache/beam/issues/21265: > apache_beam.runners.portability.fn_api_runner.translations_test.TranslationsTest.test_run_packable_combine_globally > 'apache_beam.coders.coder_impl._AbstractIterable' object is not reversible > https://api.github.com/repos/apache/beam/issues/21263: (Broken Pipe > induced) Bricked Dataflow Pipeline > https://api.github.com/repos/apache/beam/issues/21262: Python AfterAny, > AfterAll do not follow spec > https://api.github.com/repos/apache/beam/issues/21260: Python > DirectRunner does not emit data at GC time > https://api.github.com/repos/apache/beam/issues/21259: Consumer group > with random prefix > https://api.github.com/repos/apache/beam/issues/21258: Dataflow error in > CombinePerKey operation > https://api.github.com/repos/apache/beam/issues/21257: Either Create or > DirectRunner fails to produce all elements to the following transform > https://api.github.com/repos/apache/beam/issues/21123: Multiple jobs > running on Flink session cluster reuse the persistent Python environment. > https://api.github.com/repos/apache/beam/issues/21119: Migrate to the > next version of Python `requests` when released > https://api.github.com/repos/apache/beam/issues/21117: "Java IO IT Tests" > - missing data in grafana > https://api.github.com/repos/apache/beam/issues/21115: JdbcIO date > conversion is sensitive to OS > https://api.github.com/repos/apache/beam/issues/21112: Dataflow > SocketException (SSLException) error while trying to send message from > Cloud Pub/Sub to BigQuery > https://api.github.com/repos/apache/beam/issues/21111: Java creates an > incorrect pipeline proto when core-construction-java jar is not in the > CLASSPATH > https://api.github.com/repos/apache/beam/issues/21110: codecov/patch has > poor behavior > https://api.github.com/repos/apache/beam/issues/21109: SDF BoundedSource > seems to execute significantly slower than 'normal' BoundedSource > https://api.github.com/repos/apache/beam/issues/21108: > java.io.InvalidClassException With Flink Kafka > https://api.github.com/repos/apache/beam/issues/20979: Portable runners > should be able to issue checkpoints to Splittable DoFn > https://api.github.com/repos/apache/beam/issues/20978: > PubsubIO.readAvroGenericRecord creates SchemaCoder that fails to decode > some Avro logical types > https://api.github.com/repos/apache/beam/issues/20973: Python Beam SDK > Harness hangs when installing pip packages > https://api.github.com/repos/apache/beam/issues/20818: XmlIO.Read does > not handle XML encoding per spec > https://api.github.com/repos/apache/beam/issues/20814: JmsIO is not > acknowledging messages correctly > https://api.github.com/repos/apache/beam/issues/20813: No trigger early > repeatedly for session windows > https://api.github.com/repos/apache/beam/issues/20812: Cross-language > consistency (RequiresStableInputs) is quietly broken (at least on portable > flink runner) > https://api.github.com/repos/apache/beam/issues/20692: Timer with > dataflow runner can be set multiple times (dataflow runner) > https://api.github.com/repos/apache/beam/issues/20691: Beam metrics > should be displayed in Flink UI "Metrics" tab > https://api.github.com/repos/apache/beam/issues/20689: Kafka > commitOffsetsInFinalize OOM on Flink > https://api.github.com/repos/apache/beam/issues/20532: Support for coder > argument in WriteToBigQuery > https://api.github.com/repos/apache/beam/issues/20531: FileBasedSink: > allow setting temp directory provider per dynamic destination > https://api.github.com/repos/apache/beam/issues/20530: Make non-portable > Splittable DoFn the only option when executing Java "Read" transforms > https://api.github.com/repos/apache/beam/issues/20529: SpannerIO tests > don't actually assert anything. > https://api.github.com/repos/apache/beam/issues/20528: python > CombineGlobally().with_fanout() cause duplicate combine results for sliding > windows > https://api.github.com/repos/apache/beam/issues/20333: > beam_PerformanceTests_Kafka_IO failing due to " provided port is already > allocated" > https://api.github.com/repos/apache/beam/issues/20332: FileIO > writeDynamic with AvroIO.sink not writing all data > https://api.github.com/repos/apache/beam/issues/20330: Remove insecure > ssl options from MongoDBIO > https://api.github.com/repos/apache/beam/issues/20109: SortValues should > fail if SecondaryKey coder is not deterministic > https://api.github.com/repos/apache/beam/issues/20108: Python direct > runner doesn't emit empty pane when it should > https://api.github.com/repos/apache/beam/issues/20009: > Environment-sensitive provisioning for Dataflow > https://api.github.com/repos/apache/beam/issues/19971: [SQL] Some Hive > tests throw NullPointerException, but get marked as passing (Direct Runner) > https://api.github.com/repos/apache/beam/issues/19817: datetime and > decimal should be logical types > https://api.github.com/repos/apache/beam/issues/19815: Add support for > remaining data types in python RowCoder > https://api.github.com/repos/apache/beam/issues/19813: PubsubIO returns > empty message bodies for all messages read > https://api.github.com/repos/apache/beam/issues/19556: User reports > protobuf ClassChangeError running against 2.6.0 or above > https://api.github.com/repos/apache/beam/issues/19369: KafkaIO doesn't > commit offsets while being used as bounded source > https://api.github.com/repos/apache/beam/issues/17950: [Bug]: Java > Precommit permared >