awelless commented on code in PR #10077: URL: https://github.com/apache/nifi/pull/10077#discussion_r2197703861
########## nifi-extension-bundles/nifi-aws-bundle/nifi-aws-processors/src/main/java/org/apache/nifi/processors/aws/kinesis/stream/consumev2/ConsumeKinesisStreamV2Attributes.java: ########## @@ -0,0 +1,52 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.nifi.processors.aws.kinesis.stream.consumev2; + +import software.amazon.kinesis.retrieval.KinesisClientRecord; + +import java.util.Map; + +final class ConsumeKinesisStreamV2Attributes { + private static final String PREFIX = "aws.kinesis."; + + // AWS Kinesis attributes. + static final String SHARD_ID = PREFIX + "shard.id"; + static final String SEQUENCE_NUMBER = PREFIX + "sequence.number"; + static final String SUB_SEQUENCE_NUMBER = PREFIX + "subsequence.number"; + + static final String PARTITION_KEY = PREFIX + "partition.key"; + static final String APPROXIMATE_ARRIVAL_TIMESTAMP = PREFIX + "approximate.arrival.timestamp"; + + // Record attributes. + static final String RECORD_COUNT = "record.count"; + static final String RECORD_ERROR_MESSAGE = "record.error.message"; + + static Map<String, String> fromKinesisRecord( + final String shardId, + final KinesisClientRecord record) { + return Map.of( + SHARD_ID, shardId, + SEQUENCE_NUMBER, record.sequenceNumber(), + SUB_SEQUENCE_NUMBER, String.valueOf(record.subSequenceNumber()), + PARTITION_KEY, record.partitionKey(), + APPROXIMATE_ARRIVAL_TIMESTAMP, record.approximateArrivalTimestamp().toString() Review Comment: That's strange how arrival timestamp might not be present. But let's not assume anything about 3rd party lib. Added a null check. ########## nifi-mock/src/main/java/org/apache/nifi/util/MockProcessSession.java: ########## @@ -122,6 +128,7 @@ public MockProcessSession(final SharedSessionState sharedState, final Processor this.stateManager = stateManager; this.allowSynchronousCommits = allowSynchronousCommits; this.allowRecursiveReads = allowRecursiveReads; + this.shouldRollbackSession = shouldRollbackSession; Review Comment: Named it `shouldFailCommit` and slightly reorganized the class. ########## nifi-extension-bundles/nifi-aws-bundle/nifi-aws-processors/src/test/java/org/apache/nifi/processors/aws/kinesis/stream/consumev2/RecordBufferTest.java: ########## @@ -0,0 +1,695 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.nifi.processors.aws.kinesis.stream.consumev2; + +import org.apache.nifi.documentation.init.NopComponentLog; +import org.apache.nifi.processors.aws.kinesis.stream.consumev2.RecordBuffer.ShardBufferId; +import org.apache.nifi.processors.aws.kinesis.stream.consumev2.RecordBuffer.ShardBufferLease; +import org.junit.jupiter.api.BeforeEach; +import org.junit.jupiter.api.Test; +import org.junit.jupiter.api.Timeout; +import org.junit.jupiter.params.ParameterizedTest; +import org.junit.jupiter.params.provider.ValueSource; +import software.amazon.awssdk.services.kinesis.model.Record; +import software.amazon.kinesis.exceptions.InvalidStateException; +import software.amazon.kinesis.exceptions.KinesisClientLibDependencyException; +import software.amazon.kinesis.exceptions.ShutdownException; +import software.amazon.kinesis.exceptions.ThrottlingException; +import software.amazon.kinesis.processor.Checkpointer; +import software.amazon.kinesis.processor.PreparedCheckpointer; +import software.amazon.kinesis.processor.RecordProcessorCheckpointer; +import software.amazon.kinesis.retrieval.KinesisClientRecord; + +import java.nio.ByteBuffer; +import java.nio.charset.StandardCharsets; +import java.util.Arrays; +import java.util.Collection; +import java.util.Collections; +import java.util.List; +import java.util.Optional; +import java.util.concurrent.CountDownLatch; +import java.util.concurrent.ExecutorService; +import java.util.concurrent.Executors; +import java.util.concurrent.Future; +import java.util.concurrent.atomic.AtomicBoolean; +import java.util.concurrent.atomic.AtomicInteger; +import java.util.stream.IntStream; + +import static java.util.concurrent.TimeUnit.SECONDS; +import static org.junit.jupiter.api.Assertions.assertAll; +import static org.junit.jupiter.api.Assertions.assertEquals; +import static org.junit.jupiter.api.Assertions.assertFalse; +import static org.junit.jupiter.api.Assertions.assertNotEquals; +import static org.junit.jupiter.api.Assertions.assertThrows; +import static org.junit.jupiter.api.Assertions.assertTrue; + +class RecordBufferTest { + + private static final long MAX_MEMORY_BYTES = 1024L; + private static final String SHARD_ID_1 = "shard-1"; + private static final String SHARD_ID_2 = "shard-2"; + + private RecordBuffer recordBuffer; + private TestCheckpointer checkpointer1; + private TestCheckpointer checkpointer2; + + @BeforeEach + void setUp() { + recordBuffer = new RecordBuffer(new NopComponentLog(), MAX_MEMORY_BYTES); + checkpointer1 = new TestCheckpointer(); + checkpointer2 = new TestCheckpointer(); + } + + @Test + void testCreateBuffer() { + final ShardBufferId bufferId1 = recordBuffer.createBuffer(SHARD_ID_1); + assertEquals(SHARD_ID_1, bufferId1.shardId()); + + final ShardBufferId bufferId2 = recordBuffer.createBuffer(SHARD_ID_2); + assertEquals(SHARD_ID_2, bufferId2.shardId()); + + final ShardBufferId newBufferId1 = recordBuffer.createBuffer(SHARD_ID_1); + assertEquals(SHARD_ID_1, newBufferId1.shardId()); + + assertNotEquals(bufferId1, bufferId2); + assertNotEquals(bufferId1, newBufferId1); + } + + @Test + void testAddRecordsToBuffer() { + final ShardBufferId bufferId = recordBuffer.createBuffer(SHARD_ID_1); + final List<KinesisClientRecord> records = createTestRecords(2); + + recordBuffer.addRecords(bufferId, records, checkpointer1); + + // Should be able to get buffer ID from pool since buffer has records. + final ShardBufferLease lease = recordBuffer.acquireBufferLease().orElseThrow(); + assertEquals(SHARD_ID_1, lease.shardId()); + } Review Comment: Buffers without record can't be leased. I added the assertion for that. ########## nifi-extension-bundles/nifi-aws-bundle/nifi-aws-processors/src/main/java/org/apache/nifi/processors/aws/kinesis/stream/consumev2/ConsumeKinesisStreamV2.java: ########## @@ -0,0 +1,572 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.nifi.processors.aws.kinesis.stream.consumev2; + +import org.apache.nifi.annotation.behavior.InputRequirement; +import org.apache.nifi.annotation.behavior.SystemResource; +import org.apache.nifi.annotation.behavior.SystemResourceConsideration; +import org.apache.nifi.annotation.behavior.TriggerSerially; +import org.apache.nifi.annotation.behavior.WritesAttribute; +import org.apache.nifi.annotation.behavior.WritesAttributes; +import org.apache.nifi.annotation.configuration.DefaultSettings; +import org.apache.nifi.annotation.documentation.CapabilityDescription; +import org.apache.nifi.annotation.documentation.Tags; +import org.apache.nifi.annotation.lifecycle.OnScheduled; +import org.apache.nifi.annotation.lifecycle.OnStopped; +import org.apache.nifi.components.DescribedValue; +import org.apache.nifi.components.PropertyDescriptor; +import org.apache.nifi.expression.ExpressionLanguageScope; +import org.apache.nifi.flowfile.FlowFile; +import org.apache.nifi.processor.AbstractProcessor; +import org.apache.nifi.processor.DataUnit; +import org.apache.nifi.processor.ProcessContext; +import org.apache.nifi.processor.ProcessSession; +import org.apache.nifi.processor.Relationship; +import org.apache.nifi.processor.exception.ProcessException; +import org.apache.nifi.processor.util.StandardValidators; +import org.apache.nifi.processors.aws.credentials.provider.service.AWSCredentialsProviderService; +import org.apache.nifi.processors.aws.kinesis.stream.consumev2.RecordBuffer.ShardBufferId; +import org.apache.nifi.processors.aws.kinesis.stream.consumev2.RecordBuffer.ShardBufferLease; +import org.apache.nifi.processors.aws.v2.RegionUtilV2; +import org.apache.nifi.proxy.ProxyConfiguration; +import org.apache.nifi.proxy.ProxyConfigurationService; +import org.apache.nifi.proxy.ProxySpec; +import org.apache.nifi.serialization.RecordReaderFactory; +import org.apache.nifi.serialization.RecordSetWriterFactory; +import software.amazon.awssdk.auth.credentials.AwsCredentialsProvider; +import software.amazon.awssdk.http.Protocol; +import software.amazon.awssdk.http.async.SdkAsyncHttpClient; +import software.amazon.awssdk.http.nio.netty.Http2Configuration; +import software.amazon.awssdk.http.nio.netty.NettyNioAsyncHttpClient; +import software.amazon.awssdk.regions.Region; +import software.amazon.awssdk.services.cloudwatch.CloudWatchAsyncClient; +import software.amazon.awssdk.services.dynamodb.DynamoDbAsyncClient; +import software.amazon.awssdk.services.kinesis.KinesisAsyncClient; +import software.amazon.awssdk.services.kinesis.KinesisAsyncClientBuilder; +import software.amazon.kinesis.common.ConfigsBuilder; +import software.amazon.kinesis.common.InitialPositionInStream; +import software.amazon.kinesis.common.InitialPositionInStreamExtended; +import software.amazon.kinesis.coordinator.Scheduler; +import software.amazon.kinesis.lifecycle.events.InitializationInput; +import software.amazon.kinesis.lifecycle.events.LeaseLostInput; +import software.amazon.kinesis.lifecycle.events.ProcessRecordsInput; +import software.amazon.kinesis.lifecycle.events.ShardEndedInput; +import software.amazon.kinesis.lifecycle.events.ShutdownRequestedInput; +import software.amazon.kinesis.processor.ShardRecordProcessor; +import software.amazon.kinesis.processor.ShardRecordProcessorFactory; +import software.amazon.kinesis.processor.SingleStreamTracker; +import software.amazon.kinesis.retrieval.KinesisClientRecord; + +import java.net.URI; +import java.nio.channels.Channels; +import java.time.Duration; +import java.time.Instant; +import java.util.Date; +import java.util.List; +import java.util.Optional; +import java.util.Set; +import java.util.UUID; + +import static org.apache.nifi.processors.aws.kinesis.stream.consumev2.ConsumeKinesisStreamV2Attributes.APPROXIMATE_ARRIVAL_TIMESTAMP; +import static org.apache.nifi.processors.aws.kinesis.stream.consumev2.ConsumeKinesisStreamV2Attributes.PARTITION_KEY; +import static org.apache.nifi.processors.aws.kinesis.stream.consumev2.ConsumeKinesisStreamV2Attributes.RECORD_COUNT; +import static org.apache.nifi.processors.aws.kinesis.stream.consumev2.ConsumeKinesisStreamV2Attributes.RECORD_ERROR_MESSAGE; +import static org.apache.nifi.processors.aws.kinesis.stream.consumev2.ConsumeKinesisStreamV2Attributes.SEQUENCE_NUMBER; +import static org.apache.nifi.processors.aws.kinesis.stream.consumev2.ConsumeKinesisStreamV2Attributes.SHARD_ID; +import static org.apache.nifi.processors.aws.kinesis.stream.consumev2.ConsumeKinesisStreamV2Attributes.SUB_SEQUENCE_NUMBER; + +@InputRequirement(InputRequirement.Requirement.INPUT_FORBIDDEN) +@TriggerSerially +@Tags({"amazon", "aws", "kinesis", "consume", "stream", "record"}) +@CapabilityDescription("Consumes data from the specified AWS Kinesis stream and outputs a FlowFile for every processed Record (raw) " + + " or a FlowFile for a batch of processed records if a Record Reader and Record Writer are configured. " + + "AWS Kinesis Client Library can take several seconds to initialise before starting to fetch data. " + + "Uses DynamoDB for check pointing and coordination, and CloudWatch (optional) for metrics. " + + "Ensure that the credentials provided have access to DynamoDB and CloudWatch (optional) along with Kinesis.") +@WritesAttributes({ + @WritesAttribute(attribute = SHARD_ID, + description = "Shard ID from which all Kinesis Records in the Flow File were read"), + @WritesAttribute(attribute = PARTITION_KEY, + description = "Partition key of the last Kinesis Record in the Flow File"), + @WritesAttribute(attribute = SEQUENCE_NUMBER, + description = "A Sequence Number of the last Kinesis Record in the Flow File"), + @WritesAttribute(attribute = SUB_SEQUENCE_NUMBER, + description = "A SubSequence Number of the last Kinesis Record in the Flow File. Generated by KPL when aggregating records into a single Kinesis Record"), + @WritesAttribute(attribute = APPROXIMATE_ARRIVAL_TIMESTAMP, + description = "Approximate arrival timestamp of the last Kinesis Record in the Flow File"), + @WritesAttribute(attribute = "mime.type", + description = "Sets the mime.type attribute to the MIME Type specified by the Record Writer (if configured)"), + @WritesAttribute(attribute = RECORD_COUNT, + description = "Number of records written to the FlowFiles by the Record Writer (if configured)"), + @WritesAttribute(attribute = RECORD_ERROR_MESSAGE, + description = "This attribute provides on failure the error message encountered by the Record Reader or Record Writer (if configured)") +}) +@DefaultSettings(yieldDuration = "100 millis") +@SystemResourceConsideration(resource = SystemResource.CPU, description = "Kinesis Client Library is used to create a Worker thread for consumption of Kinesis Records. " + + "The Worker is initialised and started when this Processor has been scheduled. It runs continually, spawning Kinesis Record Processors as required " + + "to fetch Kinesis Records. The Worker Thread (and any child Record Processor threads) are not controlled by the normal NiFi scheduler as part of the " + + "Concurrent Thread pool and are not released until this processor is stopped.") +@SystemResourceConsideration(resource = SystemResource.NETWORK, description = "Kinesis Client Library will continually poll for new Records, " + + "requesting up to a maximum number of Records/bytes per call. This can result in sustained network usage.") +public class ConsumeKinesisStreamV2 extends AbstractProcessor { + + public enum InitialPosition implements DescribedValue { + TRIM_HORIZON("Trim Horizon", "Start reading at the last untrimmed record in the shard in the system, which is the oldest data record in the shard."), + LATEST("Latest", "Start reading just after the most recent record in the shard, so that you always read the most recent data in the shard."), + AT_TIMESTAMP("At Timestamp", "Start reading at the record with the specified timestamp."); + + private final String displayName; + private final String description; + + InitialPosition(final String displayName, final String description) { + this.displayName = displayName; + this.description = description; + } + + @Override + public String getValue() { + return name(); + } + + @Override + public String getDisplayName() { + return displayName; + } + + @Override + public String getDescription() { + return description; + } + } + + static final PropertyDescriptor KINESIS_STREAM_NAME = new PropertyDescriptor.Builder() + .name("Amazon Kinesis Stream Name") + .description("The name of the Kinesis stream to consume from.") + .required(true) + .addValidator(StandardValidators.NON_EMPTY_VALIDATOR) + .build(); + + static final PropertyDescriptor APPLICATION_NAME = new PropertyDescriptor.Builder() + .name("Application Name") + .description("The name of the Kinesis application. This is used for DynamoDB table naming and worker coordination.") + .required(true) + .addValidator(StandardValidators.NON_EMPTY_VALIDATOR) + .build(); + + static final PropertyDescriptor AWS_CREDENTIALS_PROVIDER_SERVICE = new PropertyDescriptor.Builder() + .name("AWS Credentials Provider service") + .displayName("AWS Credentials Provider Service") + .description("The Controller Service that is used to obtain AWS credentials provider.") + .required(true) + .identifiesControllerService(AWSCredentialsProviderService.class) + .build(); + + static final PropertyDescriptor REGION = new PropertyDescriptor.Builder() + .name("Region") + .description("AWS Region in which the Kinesis stream is located.") + .required(true) + .allowableValues(RegionUtilV2.getAvailableRegions()) + .defaultValue(RegionUtilV2.createAllowableValue(Region.US_WEST_2).getValue()) + .build(); + + static final PropertyDescriptor RECORD_READER = new PropertyDescriptor.Builder() + .name("Record Reader") + .description(""" + The Record Reader to use for parsing the data received from Kinesis. + If not set, the records are written one per FlowFile without parsing.""") + .required(false) + .identifiesControllerService(RecordReaderFactory.class) + .build(); + + static final PropertyDescriptor RECORD_WRITER = new PropertyDescriptor.Builder() + .name("Record Writer") + .description("The Record Writer to use for serializing records.") + .required(true) + .dependsOn(RECORD_READER) + .identifiesControllerService(RecordSetWriterFactory.class) + .build(); + + static final PropertyDescriptor INITIAL_STREAM_POSITION = new PropertyDescriptor.Builder() + .name("Initial Stream Position") + .description("The position in the stream where the processor should start reading.") + .required(true) + .allowableValues(InitialPosition.class) + .defaultValue(InitialPosition.TRIM_HORIZON) + .build(); + + static final PropertyDescriptor STREAM_POSITION_TIMESTAMP = new PropertyDescriptor.Builder() + .name("Stream Position Timestamp") + .description("Timestamp position in stream from which to start reading Kinesis Records. The timestamp must be in ISO 8601 format.") + .addValidator(StandardValidators.ISO8601_INSTANT_VALIDATOR) + .dependsOn(INITIAL_STREAM_POSITION, InitialPosition.AT_TIMESTAMP) + .required(true) + .build(); + + static final PropertyDescriptor MAX_BYTES_TO_BUFFER = new PropertyDescriptor.Builder() + .name("Max Bytes to Buffer") + .description(""" + The maximum size of Kinesis Records that can be buffered in memory when receiving from Kinesis is faster than processing records. " + Using a larger value may improve throughput, but will do so at the expense of using additional heap.""") + .required(true) + .addValidator(StandardValidators.DATA_SIZE_VALIDATOR) + .defaultValue("100 MB") + .build(); + + static final PropertyDescriptor TIMEOUT = new PropertyDescriptor.Builder() + .name("Communications Timeout") + .description(""" + Timeout for communication with AWS Kinesis, DynamoDB, and CloudWatch. + This timeout is applied to all requests made by this processor to AWS services.""") + .required(true) + .addValidator(StandardValidators.TIME_PERIOD_VALIDATOR) + .defaultValue("30 secs") + .build(); + + static final PropertyDescriptor ENDPOINT_OVERRIDE = new PropertyDescriptor.Builder() + .name("Kinesis Endpoint Override URL") + .description(""" + Endpoint URL to use instead of the AWS default including scheme, host, port, and path. + The AWS libraries select an endpoint URL based on the AWS region, but this property overrides the selected endpoint URL, allowing use with other S3-compatible endpoints.""") + .expressionLanguageSupported(ExpressionLanguageScope.ENVIRONMENT) + .required(false) + .addValidator(StandardValidators.URL_VALIDATOR) + .build(); + + static final PropertyDescriptor DYNAMODB_ENDPOINT_OVERRIDE = new PropertyDescriptor.Builder() + .name("Dynamo DB Endpoint Override") + .description("An optional endpoint URL to use for DynamoDB. If not specified, the default AWS endpoint for the region will be used.") + .addValidator(StandardValidators.URL_VALIDATOR) + .expressionLanguageSupported(ExpressionLanguageScope.ENVIRONMENT) + .required(false) + .build(); + + static final PropertyDescriptor REPORT_CLOUDWATCH_METRICS = new PropertyDescriptor.Builder() + .name("Report Metrics to CloudWatch") + .description("Whether to report Kinesis usage metrics to CloudWatch.") + .addValidator(StandardValidators.BOOLEAN_VALIDATOR) + .allowableValues("true", "false") + .defaultValue("false") + .required(true) + .build(); + + static final PropertyDescriptor CLOUDWATCH_ENDPOINT_OVERRIDE = new PropertyDescriptor.Builder() + .name("CloudWatch Endpoint Override") + .description("An optional endpoint URL to use for CloudWatch. If not specified, the default AWS endpoint for the region will be used.") + .addValidator(StandardValidators.URL_VALIDATOR) + .expressionLanguageSupported(ExpressionLanguageScope.ENVIRONMENT) + .required(false) + .build(); + + static final PropertyDescriptor PROXY_CONFIGURATION_SERVICE = ProxyConfiguration.createProxyConfigPropertyDescriptor(ProxySpec.HTTP, ProxySpec.HTTP_AUTH); + + private static final List<PropertyDescriptor> PROPERTY_DESCRIPTORS = List.of( + KINESIS_STREAM_NAME, + APPLICATION_NAME, + AWS_CREDENTIALS_PROVIDER_SERVICE, + REGION, + RECORD_READER, + RECORD_WRITER, + INITIAL_STREAM_POSITION, + STREAM_POSITION_TIMESTAMP, + MAX_BYTES_TO_BUFFER, + TIMEOUT, + ENDPOINT_OVERRIDE, + DYNAMODB_ENDPOINT_OVERRIDE, + CLOUDWATCH_ENDPOINT_OVERRIDE, + PROXY_CONFIGURATION_SERVICE, + REPORT_CLOUDWATCH_METRICS + ); + + static final Relationship REL_SUCCESS = new Relationship.Builder() + .name("success") + .description("FlowFiles that are created when records are successfully read from Kinesis and parsed.") + .build(); + + static final Relationship REL_PARSE_FAILURE = new Relationship.Builder() + .name("parse.failure") + .description("FlowFiles that failed to parse using the configured Record Reader.") + .build(); + + private static final Set<Relationship> RAW_FILE_PROPERTIES = Set.of(REL_SUCCESS); + private static final Set<Relationship> RECORD_FILE_PROPERTIES = Set.of(REL_SUCCESS, REL_PARSE_FAILURE); + + private volatile DynamoDbAsyncClient dynamoDbClient; + private volatile CloudWatchAsyncClient cloudWatchClient; + private volatile KinesisAsyncClient kinesisClient; + private volatile Scheduler kinesisScheduler; + + private volatile RecordBuffer recordBuffer; + + @Override + protected List<PropertyDescriptor> getSupportedPropertyDescriptors() { + return PROPERTY_DESCRIPTORS; + } + + @Override + public Set<Relationship> getRelationships() { + return RAW_FILE_PROPERTIES; + } + + @OnScheduled + public void setup(final ProcessContext context) { +// TODO: support Record Reader and Writer. +// final RecordReaderFactory recordReaderFactory = context.getProperty(RECORD_READER).asControllerService(RecordReaderFactory.class); +// if (recordReaderFactory != null) { +// final RecordSetWriterFactory recordWriterFactory = context.getProperty(RECORD_WRITER).asControllerService(RecordSetWriterFactory.class); +// } + + final Region region = Region.of(context.getProperty(REGION).getValue()); + final AwsCredentialsProvider credentialsProvider = context.getProperty(AWS_CREDENTIALS_PROVIDER_SERVICE) + .asControllerService(AWSCredentialsProviderService.class).getAwsCredentialsProvider(); + + final String kinesisEndpointOverride = context.getProperty(ENDPOINT_OVERRIDE).evaluateAttributeExpressions().getValue(); + final URI kinesisEndpoint = kinesisEndpointOverride == null ? null : URI.create(kinesisEndpointOverride); + kinesisClient = KinesisAsyncClient.builder() + .region(region) + .credentialsProvider(credentialsProvider) + .endpointOverride(kinesisEndpoint) + .httpClient(createKinesisHttpClient(context)) + .build(); + + final String dynamoDbEndpointOverride = context.getProperty(DYNAMODB_ENDPOINT_OVERRIDE).evaluateAttributeExpressions().getValue(); + final URI dynamoDbEndpoint = dynamoDbEndpointOverride == null ? null : URI.create(dynamoDbEndpointOverride); + dynamoDbClient = DynamoDbAsyncClient.builder() + .region(region) + .credentialsProvider(credentialsProvider) + .endpointOverride(dynamoDbEndpoint) + .httpClient(createHttpClientBuilder(context).build()) + .build(); + + final String cloudwatchEndpointOverride = context.getProperty(CLOUDWATCH_ENDPOINT_OVERRIDE).evaluateAttributeExpressions().getValue(); + final URI cloudWatchEndpoint = cloudwatchEndpointOverride == null ? null : URI.create(cloudwatchEndpointOverride); + cloudWatchClient = CloudWatchAsyncClient.builder() + .region(region) + .credentialsProvider(credentialsProvider) + .endpointOverride(cloudWatchEndpoint) + .httpClient(createHttpClientBuilder(context).build()) + .build(); + + final String streamName = context.getProperty(KINESIS_STREAM_NAME).getValue(); + final InitialPositionInStreamExtended initialPositionExtended = getInitialPosition(context); + final SingleStreamTracker streamTracker = new SingleStreamTracker(streamName, initialPositionExtended); + + recordBuffer = new RecordBuffer(getLogger(), context.getProperty(MAX_BYTES_TO_BUFFER).asDataSize(DataUnit.B).longValue()); + final ShardRecordProcessorFactory recordProcessorFactory = () -> new ConsumeKinesisRecordProcessor(recordBuffer); + + final String applicationName = context.getProperty(APPLICATION_NAME).getValue(); + final String workerId = UUID.randomUUID().toString(); + final ConfigsBuilder configsBuilder = new ConfigsBuilder(streamTracker, applicationName, kinesisClient, dynamoDbClient, cloudWatchClient, workerId, recordProcessorFactory); + + if (!context.getProperty(REPORT_CLOUDWATCH_METRICS).asBoolean()) { + configsBuilder.metricsConfig().metricsFactory(null); + } + + kinesisScheduler = new Scheduler( + configsBuilder.checkpointConfig(), + configsBuilder.coordinatorConfig(), + configsBuilder.leaseManagementConfig(), + configsBuilder.lifecycleConfig(), + configsBuilder.metricsConfig(), + configsBuilder.processorConfig(), + configsBuilder.retrievalConfig() + ); + + final Thread schedulerThread = new Thread(kinesisScheduler); + schedulerThread.setDaemon(true); + schedulerThread.start(); + } + + /** + * Creating Kinesis HTTP client, as per + * {@link software.amazon.kinesis.common.KinesisClientUtil#adjustKinesisClientBuilder(KinesisAsyncClientBuilder)}. + */ + private static SdkAsyncHttpClient createKinesisHttpClient(final ProcessContext context) { + return createHttpClientBuilder(context) + .maxConcurrency(Integer.MAX_VALUE) // todo - is it a good idea? do perf testing + .http2Configuration(Http2Configuration.builder() + .initialWindowSize(512 * 1024) // 512 KB + .healthCheckPingPeriod(Duration.ofMinutes(1)) + .build()) + .protocol(Protocol.HTTP2) + .build(); + } + + private static NettyNioAsyncHttpClient.Builder createHttpClientBuilder(final ProcessContext context) { + final Duration timeout = context.getProperty(TIMEOUT).asDuration(); + + final NettyNioAsyncHttpClient.Builder builder = NettyNioAsyncHttpClient.builder() + .connectionTimeout(timeout) + .readTimeout(timeout); + + final ProxyConfigurationService proxyConfigService = context.getProperty(PROXY_CONFIGURATION_SERVICE).asControllerService(ProxyConfigurationService.class); + if (proxyConfigService != null) { + final ProxyConfiguration proxyConfig = proxyConfigService.getConfiguration(); + + final software.amazon.awssdk.http.nio.netty.ProxyConfiguration.Builder proxyConfigBuilder = software.amazon.awssdk.http.nio.netty.ProxyConfiguration.builder() + .host(proxyConfig.getProxyServerHost()) + .port(proxyConfig.getProxyServerPort()); + + if (proxyConfig.hasCredential()) { + proxyConfigBuilder.username(proxyConfig.getProxyUserName()); + proxyConfigBuilder.password(proxyConfig.getProxyUserPassword()); + } + + builder.proxyConfiguration(proxyConfigBuilder.build()); + } + + return builder; + } + + private static InitialPositionInStreamExtended getInitialPosition(final ProcessContext context) { + final InitialPosition initialPosition = context.getProperty(INITIAL_STREAM_POSITION).asAllowableValue(InitialPosition.class); + return switch (initialPosition) { + case TRIM_HORIZON -> InitialPositionInStreamExtended.newInitialPosition(InitialPositionInStream.TRIM_HORIZON); + case LATEST -> InitialPositionInStreamExtended.newInitialPosition(InitialPositionInStream.LATEST); + case AT_TIMESTAMP -> { + final String timestampValue = context.getProperty(STREAM_POSITION_TIMESTAMP).getValue(); + final Instant timestamp = Instant.parse(timestampValue); + yield InitialPositionInStreamExtended.newInitialPositionAtTimestamp(Date.from(timestamp)); + } + }; + } + + @OnStopped + public void onStopped() { + if (kinesisScheduler != null) { + kinesisScheduler.shutdown(); + } + if (kinesisClient != null) { + kinesisClient.close(); + } + if (dynamoDbClient != null) { + dynamoDbClient.close(); + } + if (cloudWatchClient != null) { + cloudWatchClient.close(); + } + + recordBuffer = null; + } + + @Override + public void onTrigger(final ProcessContext context, final ProcessSession session) throws ProcessException { + final Optional<ShardBufferLease> maybeLease = recordBuffer.acquireBufferLease(); + + maybeLease.ifPresentOrElse( + lease -> processRecordsFromBuffer(session, lease), + context::yield + ); + } + + private void processRecordsFromBuffer(final ProcessSession session, final ShardBufferLease lease) { + try { + final List<KinesisClientRecord> records = recordBuffer.consumeRecords(lease); + + if (records.isEmpty()) { + recordBuffer.returnBufferLease(lease); + return; + } + + final String shardId = lease.shardId(); + + processRecordsAsRaw(session, shardId, records); + + // TODO support Record Reader and Writer. +// if (recordProcessor == null) { +// processRecordsAsRaw(session, shardId, records); +// } else { +// processRecordsWithRecordProcessor(recordProcessor, session, shardId, records); +// } + + session.adjustCounter("Records Processed", records.size(), false); + + session.commitAsync( + () -> { + recordBuffer.commitConsumedRecords(lease); + recordBuffer.returnBufferLease(lease); + }, + __ -> { + recordBuffer.rollbackConsumedRecords(lease); + recordBuffer.returnBufferLease(lease); + } + ); + } catch (final RuntimeException e) { + getLogger().error("Failed to process records from Kinesis stream: {}", e.getMessage()); + recordBuffer.rollbackConsumedRecords(lease); + recordBuffer.returnBufferLease(lease); + throw e; + } + } + + private static void processRecordsAsRaw(final ProcessSession session, final String shardId, final List<KinesisClientRecord> records) { + for (final KinesisClientRecord record : records) { + final FlowFile flowFile = session.create(); + session.putAllAttributes(flowFile, ConsumeKinesisStreamV2Attributes.fromKinesisRecord(shardId, record)); + + session.write(flowFile, out -> Channels.newChannel(out).write(record.data())); + + session.transfer(flowFile, REL_SUCCESS); + } + } + + /** + * An adapter between Kinesis Consumer Library and {@link RecordBuffer}. + */ + private static class ConsumeKinesisRecordProcessor implements ShardRecordProcessor { + + private final RecordBuffer recordBuffer; + private volatile ShardBufferId bufferId; + + ConsumeKinesisRecordProcessor(final RecordBuffer recordBuffer) { + this.recordBuffer = recordBuffer; + } + + @Override + public void initialize(final InitializationInput initializationInput) { + bufferId = recordBuffer.createBuffer(initializationInput.shardId()); Review Comment: I tend to avoid excessive null checks in 1st party code. Instead I'm convinced `null`s should never be returned and `Optional` should be used, or at least, the return types should be marked as `@Nullable`. Though, if you think the null-check is necessary I don't mind adding it. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected]
