m-trieu commented on code in PR #34148: URL: https://github.com/apache/beam/pull/34148#discussion_r2009264223
########## runners/google-cloud-dataflow-java/worker/src/main/java/org/apache/beam/runners/dataflow/worker/streaming/harness/WindmillStreamPoolSender.java: ########## @@ -0,0 +1,199 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.beam.runners.dataflow.worker.streaming.harness; + +import java.util.Optional; +import java.util.concurrent.atomic.AtomicReference; +import java.util.function.Function; +import java.util.function.Supplier; +import javax.annotation.concurrent.GuardedBy; +import javax.annotation.concurrent.ThreadSafe; +import org.apache.beam.runners.dataflow.worker.streaming.ComputationState; +import org.apache.beam.runners.dataflow.worker.streaming.Watermarks; +import org.apache.beam.runners.dataflow.worker.streaming.Work; +import org.apache.beam.runners.dataflow.worker.windmill.Windmill.GetWorkRequest; +import org.apache.beam.runners.dataflow.worker.windmill.WindmillConnection; +import org.apache.beam.runners.dataflow.worker.windmill.client.WindmillStream.GetWorkStream; +import org.apache.beam.runners.dataflow.worker.windmill.client.commits.WorkCommitter; +import org.apache.beam.runners.dataflow.worker.windmill.client.getdata.GetDataClient; +import org.apache.beam.runners.dataflow.worker.windmill.client.grpc.GrpcWindmillStreamFactory; +import org.apache.beam.runners.dataflow.worker.windmill.client.throttling.StreamingEngineThrottleTimers; +import org.apache.beam.runners.dataflow.worker.windmill.work.WorkItemReceiver; +import org.apache.beam.runners.dataflow.worker.windmill.work.budget.GetWorkBudget; +import org.apache.beam.runners.dataflow.worker.windmill.work.processing.StreamingWorkScheduler; +import org.apache.beam.runners.dataflow.worker.windmill.work.refresh.HeartbeatSender; +import org.apache.beam.sdk.annotations.Internal; +import org.apache.beam.vendor.guava.v32_1_2_jre.com.google.common.base.Preconditions; + +/** + * Owns and maintains a set of streams used to communicate with a specific Windmill worker. + * Underlying streams are "cached" in a threadsafe manner so that once {@link Supplier#get} is + * called, a stream that is already started is returned. + * + * <p>Holds references to {@link + * Supplier<org.apache.beam.runners.dataflow.worker.windmill.client.WindmillStream>} because + * initializing the streams automatically start them, and we want to do so lazily here once the + * {@link GetWorkBudget} is set. + * + * <p>Once started, the underlying streams are "alive" until they are manually closed via {@link + * #closeAllStreams()}. + * + * <p>If closed, it means that the backend endpoint is no longer in the worker set. Once closed, + * these instances are not reused. + * + * @implNote Does not manage streams for fetching {@link + * org.apache.beam.runners.dataflow.worker.windmill.Windmill.GlobalData} for side inputs. + */ +@Internal +@ThreadSafe +final class WindmillStreamPoolSender implements WindmillStreamSender, StreamSender { + // private static final java.util.logging.Logger LOG = Review Comment: remove the commented out LOG ########## runners/google-cloud-dataflow-java/worker/src/main/java/org/apache/beam/runners/dataflow/worker/windmill/WindmillEndpoints.java: ########## @@ -155,6 +163,15 @@ private static Optional<HostAndPort> tryParseDirectEndpointIntoIpV6Address( */ public abstract ImmutableSet<Endpoint> windmillEndpoints(); + public abstract EndpointType endpointType(); + + public enum EndpointType { Review Comment: Can we just use the EndpointType enum in the proto? ########## runners/google-cloud-dataflow-java/worker/src/main/java/org/apache/beam/runners/dataflow/worker/windmill/WindmillEndpoints.java: ########## @@ -46,6 +48,7 @@ public abstract class WindmillEndpoints { .setVersion(Long.MAX_VALUE) .setWindmillEndpoints(ImmutableSet.of()) .setGlobalDataEndpoints(ImmutableMap.of()) + .setEndpointType(EndpointType.UNKNOWN) Review Comment: should we default to cloud path? unless the code knows how to deal with UNKNOWN? ########## runners/google-cloud-dataflow-java/worker/src/main/java/org/apache/beam/runners/dataflow/worker/streaming/harness/WindmillStreamSender.java: ########## @@ -17,152 +17,18 @@ */ package org.apache.beam.runners.dataflow.worker.streaming.harness; -import static org.apache.beam.vendor.guava.v32_1_2_jre.com.google.common.base.Preconditions.checkState; - -import java.util.concurrent.CompletableFuture; -import java.util.concurrent.ExecutorService; -import java.util.concurrent.Executors; -import java.util.concurrent.atomic.AtomicBoolean; -import java.util.concurrent.atomic.AtomicReference; -import java.util.function.Function; -import javax.annotation.concurrent.ThreadSafe; -import org.apache.beam.runners.dataflow.worker.windmill.Windmill.GetWorkRequest; -import org.apache.beam.runners.dataflow.worker.windmill.WindmillConnection; -import org.apache.beam.runners.dataflow.worker.windmill.client.WindmillStream.CommitWorkStream; -import org.apache.beam.runners.dataflow.worker.windmill.client.WindmillStream.GetDataStream; -import org.apache.beam.runners.dataflow.worker.windmill.client.WindmillStream.GetWorkStream; -import org.apache.beam.runners.dataflow.worker.windmill.client.commits.WorkCommitter; -import org.apache.beam.runners.dataflow.worker.windmill.client.getdata.GetDataClient; -import org.apache.beam.runners.dataflow.worker.windmill.client.grpc.GrpcWindmillStreamFactory; -import org.apache.beam.runners.dataflow.worker.windmill.client.throttling.StreamingEngineThrottleTimers; -import org.apache.beam.runners.dataflow.worker.windmill.work.WorkItemScheduler; -import org.apache.beam.runners.dataflow.worker.windmill.work.budget.GetWorkBudget; import org.apache.beam.runners.dataflow.worker.windmill.work.budget.GetWorkBudgetSpender; -import org.apache.beam.runners.dataflow.worker.windmill.work.refresh.FixedStreamHeartbeatSender; -import org.apache.beam.sdk.annotations.Internal; -import org.apache.beam.vendor.guava.v32_1_2_jre.com.google.common.util.concurrent.ThreadFactoryBuilder; - -/** - * Owns and maintains a set of streams used to communicate with a specific Windmill worker. - * - * <p>Once started, the underlying streams are "alive" until they are manually closed via {@link - * #close()}. - * - * <p>If closed, it means that the backend endpoint is no longer in the worker set. Once closed, - * these instances are not reused. - * - * @implNote Does not manage streams for fetching {@link - * org.apache.beam.runners.dataflow.worker.windmill.Windmill.GlobalData} for side inputs. - */ -@Internal -@ThreadSafe -final class WindmillStreamSender implements GetWorkBudgetSpender, StreamSender { - private static final String STREAM_STARTER_THREAD_NAME = "StartWindmillStreamThread-%d"; - private final AtomicBoolean started; - private final AtomicReference<GetWorkBudget> getWorkBudget; - private final GetWorkStream getWorkStream; - private final GetDataStream getDataStream; - private final CommitWorkStream commitWorkStream; - private final WorkCommitter workCommitter; - private final StreamingEngineThrottleTimers streamingEngineThrottleTimers; - private final ExecutorService streamStarter; - private WindmillStreamSender( - WindmillConnection connection, - GetWorkRequest getWorkRequest, - AtomicReference<GetWorkBudget> getWorkBudget, - GrpcWindmillStreamFactory streamingEngineStreamFactory, - WorkItemScheduler workItemScheduler, - Function<GetDataStream, GetDataClient> getDataClientFactory, - Function<CommitWorkStream, WorkCommitter> workCommitterFactory) { - this.started = new AtomicBoolean(false); - this.getWorkBudget = getWorkBudget; - this.streamingEngineThrottleTimers = StreamingEngineThrottleTimers.create(); +/** Superclass for stream senders used to communicate with Windmill */ +public interface WindmillStreamSender extends GetWorkBudgetSpender { Review Comment: I think we can move all of this to `StreamSender` or just have `WindmillStreamSender` have the 1 method in the `StreamSender` interface. don't think we need both interfaces. ########## runners/google-cloud-dataflow-java/worker/src/main/java/org/apache/beam/runners/dataflow/worker/streaming/harness/FanOutStreamingEngineWorkerHarness.java: ########## @@ -127,7 +141,13 @@ private FanOutStreamingEngineWorkerHarness( GrpcDispatcherClient dispatcherClient, Function<WindmillStream.CommitWorkStream, WorkCommitter> workCommitterFactory, ThrottlingGetDataMetricTracker getDataMetricTracker, - ExecutorService workerMetadataConsumer) { + ExecutorService workerMetadataConsumer, Review Comment: instead of passing all of these in can the constructor take in a `Function<WindmillConnection, WindmillStreamSender> windmillStreamPoolSenderFactory` or create a WindmillStreamPoolSenderFactory.java interface ``` @FunctionalInterface interface WindmillStreamPoolSenderFactory { WindmillStreamSender create(WindmillConnection connection); } ``` then in `StreamingDataflowWorker.java` you can pass in ``` connection -> WindmillStreamPoolSender.create( connection, GetWorkRequest.newBuilder() .setClientId(jobHeader.getClientId()) .setJobId(jobHeader.getJobId()) .setProjectId(jobHeader.getProjectId()) .setWorkerId(jobHeader.getWorkerId()) .setMaxItems(totalGetWorkBudget.items()) .setMaxBytes(totalGetWorkBudget.bytes()) .build(), GetWorkBudget.noBudget(), streamFactory, streamPoolWorkCommitter, streamPoolGetDataClient, streamPoolHeartbeatSender, streamingWorkScheduler, waitForResources, computationStateFetcher) ``` and in `createAndStartWindmillStreamSender` you can call `windmillStreamPoolSenderFactory.create()` / `windmillStreamPoolSenderFactory.apply(connection)` ########## runners/google-cloud-dataflow-java/worker/src/main/java/org/apache/beam/runners/dataflow/worker/streaming/harness/WindmillStreamPoolSender.java: ########## @@ -0,0 +1,199 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.beam.runners.dataflow.worker.streaming.harness; + +import java.util.Optional; +import java.util.concurrent.atomic.AtomicReference; +import java.util.function.Function; +import java.util.function.Supplier; +import javax.annotation.concurrent.GuardedBy; +import javax.annotation.concurrent.ThreadSafe; +import org.apache.beam.runners.dataflow.worker.streaming.ComputationState; +import org.apache.beam.runners.dataflow.worker.streaming.Watermarks; +import org.apache.beam.runners.dataflow.worker.streaming.Work; +import org.apache.beam.runners.dataflow.worker.windmill.Windmill.GetWorkRequest; +import org.apache.beam.runners.dataflow.worker.windmill.WindmillConnection; +import org.apache.beam.runners.dataflow.worker.windmill.client.WindmillStream.GetWorkStream; +import org.apache.beam.runners.dataflow.worker.windmill.client.commits.WorkCommitter; +import org.apache.beam.runners.dataflow.worker.windmill.client.getdata.GetDataClient; +import org.apache.beam.runners.dataflow.worker.windmill.client.grpc.GrpcWindmillStreamFactory; +import org.apache.beam.runners.dataflow.worker.windmill.client.throttling.StreamingEngineThrottleTimers; +import org.apache.beam.runners.dataflow.worker.windmill.work.WorkItemReceiver; +import org.apache.beam.runners.dataflow.worker.windmill.work.budget.GetWorkBudget; +import org.apache.beam.runners.dataflow.worker.windmill.work.processing.StreamingWorkScheduler; +import org.apache.beam.runners.dataflow.worker.windmill.work.refresh.HeartbeatSender; +import org.apache.beam.sdk.annotations.Internal; +import org.apache.beam.vendor.guava.v32_1_2_jre.com.google.common.base.Preconditions; + +/** + * Owns and maintains a set of streams used to communicate with a specific Windmill worker. + * Underlying streams are "cached" in a threadsafe manner so that once {@link Supplier#get} is + * called, a stream that is already started is returned. + * + * <p>Holds references to {@link + * Supplier<org.apache.beam.runners.dataflow.worker.windmill.client.WindmillStream>} because + * initializing the streams automatically start them, and we want to do so lazily here once the + * {@link GetWorkBudget} is set. + * + * <p>Once started, the underlying streams are "alive" until they are manually closed via {@link + * #closeAllStreams()}. + * + * <p>If closed, it means that the backend endpoint is no longer in the worker set. Once closed, + * these instances are not reused. + * + * @implNote Does not manage streams for fetching {@link + * org.apache.beam.runners.dataflow.worker.windmill.Windmill.GlobalData} for side inputs. + */ +@Internal +@ThreadSafe +final class WindmillStreamPoolSender implements WindmillStreamSender, StreamSender { + // private static final java.util.logging.Logger LOG = + // LoggerFactory.getLogger(WindmillStreamPoolSender.class); + private final AtomicReference<GetWorkBudget> getWorkBudget; + private final GetWorkStream getWorkStream; + private final WorkCommitter workCommitter; + private final StreamingEngineThrottleTimers streamingEngineThrottleTimers; + private final GetDataClient getDataClient; + private final HeartbeatSender heartbeatSender; + private final StreamingWorkScheduler streamingWorkScheduler; + private final Runnable waitForResources; + private final Function<String, Optional<ComputationState>> computationStateFetcher; + + @GuardedBy("this") + private boolean started; + + private WindmillStreamPoolSender( + WindmillConnection connection, + GetWorkRequest getWorkRequest, + AtomicReference<GetWorkBudget> getWorkBudget, + GrpcWindmillStreamFactory streamingEngineStreamFactory, + WorkCommitter workCommitter, + GetDataClient getDataClient, + HeartbeatSender heartbeatSender, + StreamingWorkScheduler streamingWorkScheduler, + Runnable waitForResources, + Function<String, Optional<ComputationState>> computationStateFetcher) { + this.started = false; + this.getWorkBudget = getWorkBudget; + this.streamingEngineThrottleTimers = StreamingEngineThrottleTimers.create(); + this.getDataClient = getDataClient; + this.heartbeatSender = heartbeatSender; + this.streamingWorkScheduler = streamingWorkScheduler; + this.waitForResources = waitForResources; + this.computationStateFetcher = computationStateFetcher; + this.workCommitter = workCommitter; + + WorkItemReceiver processWorkItem = + (computationId, + inputDataWatermark, + synchronizedProcessingTime, + workItem, + serializedWorkItemSize, + getWorkStreamLatencies) -> + this.computationStateFetcher + .apply(computationId) + .ifPresent( + computationState -> { + this.waitForResources.run(); + this.streamingWorkScheduler.scheduleWork( + computationState, + workItem, + serializedWorkItemSize, + Watermarks.builder() + .setInputDataWatermark(Preconditions.checkNotNull(inputDataWatermark)) + .setSynchronizedProcessingTime(synchronizedProcessingTime) + .setOutputDataWatermark(workItem.getOutputDataWatermark()) + .build(), + Work.createProcessingContext( + computationId, + this.getDataClient, + workCommitter::commit, + this.heartbeatSender), + getWorkStreamLatencies); + }); + this.getWorkStream = Review Comment: we need something like https://github.com/apache/beam/blob/3595a3303df6d2e4c8e77482ca8120994d0d5881/runners/google-cloud-dataflow-java/worker/src/main/java/org/apache/beam/runners/dataflow/worker/streaming/harness/SingleSourceWorkerHarness.java#L156 that manages the GetWorkStream so we get clean stream closure from the client. ########## runners/google-cloud-dataflow-java/worker/src/main/java/org/apache/beam/runners/dataflow/worker/streaming/harness/WindmillStreamPoolSender.java: ########## @@ -0,0 +1,199 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.beam.runners.dataflow.worker.streaming.harness; + +import java.util.Optional; +import java.util.concurrent.atomic.AtomicReference; +import java.util.function.Function; +import java.util.function.Supplier; +import javax.annotation.concurrent.GuardedBy; +import javax.annotation.concurrent.ThreadSafe; +import org.apache.beam.runners.dataflow.worker.streaming.ComputationState; +import org.apache.beam.runners.dataflow.worker.streaming.Watermarks; +import org.apache.beam.runners.dataflow.worker.streaming.Work; +import org.apache.beam.runners.dataflow.worker.windmill.Windmill.GetWorkRequest; +import org.apache.beam.runners.dataflow.worker.windmill.WindmillConnection; +import org.apache.beam.runners.dataflow.worker.windmill.client.WindmillStream.GetWorkStream; +import org.apache.beam.runners.dataflow.worker.windmill.client.commits.WorkCommitter; +import org.apache.beam.runners.dataflow.worker.windmill.client.getdata.GetDataClient; +import org.apache.beam.runners.dataflow.worker.windmill.client.grpc.GrpcWindmillStreamFactory; +import org.apache.beam.runners.dataflow.worker.windmill.client.throttling.StreamingEngineThrottleTimers; +import org.apache.beam.runners.dataflow.worker.windmill.work.WorkItemReceiver; +import org.apache.beam.runners.dataflow.worker.windmill.work.budget.GetWorkBudget; +import org.apache.beam.runners.dataflow.worker.windmill.work.processing.StreamingWorkScheduler; +import org.apache.beam.runners.dataflow.worker.windmill.work.refresh.HeartbeatSender; +import org.apache.beam.sdk.annotations.Internal; +import org.apache.beam.vendor.guava.v32_1_2_jre.com.google.common.base.Preconditions; + +/** + * Owns and maintains a set of streams used to communicate with a specific Windmill worker. + * Underlying streams are "cached" in a threadsafe manner so that once {@link Supplier#get} is + * called, a stream that is already started is returned. + * + * <p>Holds references to {@link + * Supplier<org.apache.beam.runners.dataflow.worker.windmill.client.WindmillStream>} because + * initializing the streams automatically start them, and we want to do so lazily here once the + * {@link GetWorkBudget} is set. + * + * <p>Once started, the underlying streams are "alive" until they are manually closed via {@link + * #closeAllStreams()}. + * + * <p>If closed, it means that the backend endpoint is no longer in the worker set. Once closed, + * these instances are not reused. + * + * @implNote Does not manage streams for fetching {@link + * org.apache.beam.runners.dataflow.worker.windmill.Windmill.GlobalData} for side inputs. + */ +@Internal +@ThreadSafe +final class WindmillStreamPoolSender implements WindmillStreamSender, StreamSender { + // private static final java.util.logging.Logger LOG = + // LoggerFactory.getLogger(WindmillStreamPoolSender.class); + private final AtomicReference<GetWorkBudget> getWorkBudget; + private final GetWorkStream getWorkStream; + private final WorkCommitter workCommitter; + private final StreamingEngineThrottleTimers streamingEngineThrottleTimers; + private final GetDataClient getDataClient; + private final HeartbeatSender heartbeatSender; + private final StreamingWorkScheduler streamingWorkScheduler; + private final Runnable waitForResources; + private final Function<String, Optional<ComputationState>> computationStateFetcher; + + @GuardedBy("this") + private boolean started; + + private WindmillStreamPoolSender( + WindmillConnection connection, + GetWorkRequest getWorkRequest, + AtomicReference<GetWorkBudget> getWorkBudget, + GrpcWindmillStreamFactory streamingEngineStreamFactory, + WorkCommitter workCommitter, + GetDataClient getDataClient, + HeartbeatSender heartbeatSender, + StreamingWorkScheduler streamingWorkScheduler, + Runnable waitForResources, + Function<String, Optional<ComputationState>> computationStateFetcher) { + this.started = false; + this.getWorkBudget = getWorkBudget; + this.streamingEngineThrottleTimers = StreamingEngineThrottleTimers.create(); + this.getDataClient = getDataClient; + this.heartbeatSender = heartbeatSender; + this.streamingWorkScheduler = streamingWorkScheduler; + this.waitForResources = waitForResources; + this.computationStateFetcher = computationStateFetcher; + this.workCommitter = workCommitter; + + WorkItemReceiver processWorkItem = + (computationId, + inputDataWatermark, + synchronizedProcessingTime, + workItem, + serializedWorkItemSize, + getWorkStreamLatencies) -> + this.computationStateFetcher + .apply(computationId) + .ifPresent( + computationState -> { + this.waitForResources.run(); + this.streamingWorkScheduler.scheduleWork( + computationState, + workItem, + serializedWorkItemSize, + Watermarks.builder() + .setInputDataWatermark(Preconditions.checkNotNull(inputDataWatermark)) + .setSynchronizedProcessingTime(synchronizedProcessingTime) + .setOutputDataWatermark(workItem.getOutputDataWatermark()) + .build(), + Work.createProcessingContext( + computationId, + this.getDataClient, + workCommitter::commit, + this.heartbeatSender), + getWorkStreamLatencies); + }); + this.getWorkStream = + streamingEngineStreamFactory.createGetWorkStream( + connection.stub(), + getWorkRequest, + streamingEngineThrottleTimers.getWorkThrottleTimer(), + processWorkItem); + } + + static WindmillStreamPoolSender create( + WindmillConnection connection, + GetWorkRequest getWorkRequest, + GetWorkBudget getWorkBudget, + GrpcWindmillStreamFactory streamingEngineStreamFactory, + WorkCommitter workCommitter, + GetDataClient getDataClient, + HeartbeatSender heartbeatSender, + StreamingWorkScheduler streamingWorkScheduler, + Runnable waitForResources, + Function<String, Optional<ComputationState>> computationStateFetcher) { + return new WindmillStreamPoolSender( + connection, + getWorkRequest, + new AtomicReference<>(getWorkBudget), + streamingEngineStreamFactory, + workCommitter, + getDataClient, + heartbeatSender, + streamingWorkScheduler, + waitForResources, + computationStateFetcher); + } + + @SuppressWarnings("ReturnValueIgnored") + @Override + public synchronized void start() { + // LOG.info("Starting Stream Pool Sender"); + if (!started) { + getWorkStream.start(); + workCommitter.start(); + + started = true; + } + } + + @Override + public synchronized void close() { + if (started) { Review Comment: StreamPool will need to be modified to clean up any dangling streams. right now we only clean up the stream on `getStream()/releaseStream()`. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: github-unsubscr...@beam.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org