exceptionfactory commented on code in PR #10077:
URL: https://github.com/apache/nifi/pull/10077#discussion_r2300928007


##########
nifi-extension-bundles/nifi-aws-bundle/nifi-aws-kinesis/src/main/java/org/apache/nifi/processors/aws/kinesis/ConsumeKinesis.java:
##########
@@ -0,0 +1,616 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.nifi.processors.aws.kinesis;
+
+import jakarta.annotation.Nullable;
+import org.apache.nifi.annotation.behavior.InputRequirement;
+import org.apache.nifi.annotation.behavior.SystemResource;
+import org.apache.nifi.annotation.behavior.SystemResourceConsideration;
+import org.apache.nifi.annotation.behavior.WritesAttribute;
+import org.apache.nifi.annotation.behavior.WritesAttributes;
+import org.apache.nifi.annotation.configuration.DefaultSettings;
+import org.apache.nifi.annotation.documentation.CapabilityDescription;
+import org.apache.nifi.annotation.documentation.Tags;
+import org.apache.nifi.annotation.lifecycle.OnScheduled;
+import org.apache.nifi.annotation.lifecycle.OnStopped;
+import org.apache.nifi.components.DescribedValue;
+import org.apache.nifi.components.PropertyDescriptor;
+import org.apache.nifi.expression.ExpressionLanguageScope;
+import org.apache.nifi.flowfile.FlowFile;
+import org.apache.nifi.processor.AbstractProcessor;
+import org.apache.nifi.processor.DataUnit;
+import org.apache.nifi.processor.ProcessContext;
+import org.apache.nifi.processor.ProcessSession;
+import org.apache.nifi.processor.Relationship;
+import org.apache.nifi.processor.exception.ProcessException;
+import org.apache.nifi.processor.util.StandardValidators;
+import 
org.apache.nifi.processors.aws.credentials.provider.service.AWSCredentialsProviderService;
+import 
org.apache.nifi.processors.aws.kinesis.ReaderRecordProcessor.ProcessingResult;
+import org.apache.nifi.processors.aws.kinesis.RecordBuffer.ShardBufferId;
+import org.apache.nifi.processors.aws.kinesis.RecordBuffer.ShardBufferLease;
+import org.apache.nifi.processors.aws.region.RegionUtilV2;
+import org.apache.nifi.proxy.ProxyConfiguration;
+import org.apache.nifi.proxy.ProxyConfigurationService;
+import org.apache.nifi.proxy.ProxySpec;
+import org.apache.nifi.serialization.RecordReaderFactory;
+import org.apache.nifi.serialization.RecordSetWriterFactory;
+import software.amazon.awssdk.auth.credentials.AwsCredentialsProvider;
+import software.amazon.awssdk.http.Protocol;
+import software.amazon.awssdk.http.async.SdkAsyncHttpClient;
+import software.amazon.awssdk.http.nio.netty.Http2Configuration;
+import software.amazon.awssdk.http.nio.netty.NettyNioAsyncHttpClient;
+import software.amazon.awssdk.regions.Region;
+import software.amazon.awssdk.services.cloudwatch.CloudWatchAsyncClient;
+import software.amazon.awssdk.services.dynamodb.DynamoDbAsyncClient;
+import software.amazon.awssdk.services.kinesis.KinesisAsyncClient;
+import software.amazon.awssdk.services.kinesis.KinesisAsyncClientBuilder;
+import software.amazon.kinesis.common.ConfigsBuilder;
+import software.amazon.kinesis.common.InitialPositionInStream;
+import software.amazon.kinesis.common.InitialPositionInStreamExtended;
+import software.amazon.kinesis.coordinator.Scheduler;
+import software.amazon.kinesis.lifecycle.events.InitializationInput;
+import software.amazon.kinesis.lifecycle.events.LeaseLostInput;
+import software.amazon.kinesis.lifecycle.events.ProcessRecordsInput;
+import software.amazon.kinesis.lifecycle.events.ShardEndedInput;
+import software.amazon.kinesis.lifecycle.events.ShutdownRequestedInput;
+import software.amazon.kinesis.processor.ShardRecordProcessor;
+import software.amazon.kinesis.processor.ShardRecordProcessorFactory;
+import software.amazon.kinesis.processor.SingleStreamTracker;
+import software.amazon.kinesis.retrieval.KinesisClientRecord;
+
+import java.net.URI;
+import java.nio.channels.Channels;
+import java.time.Duration;
+import java.time.Instant;
+import java.util.Date;
+import java.util.List;
+import java.util.Optional;
+import java.util.Set;
+import java.util.UUID;
+
+import static 
org.apache.nifi.processors.aws.kinesis.ConsumeKinesisAttributes.APPROXIMATE_ARRIVAL_TIMESTAMP;
+import static 
org.apache.nifi.processors.aws.kinesis.ConsumeKinesisAttributes.PARTITION_KEY;
+import static 
org.apache.nifi.processors.aws.kinesis.ConsumeKinesisAttributes.RECORD_COUNT;
+import static 
org.apache.nifi.processors.aws.kinesis.ConsumeKinesisAttributes.RECORD_ERROR_MESSAGE;
+import static 
org.apache.nifi.processors.aws.kinesis.ConsumeKinesisAttributes.SEQUENCE_NUMBER;
+import static 
org.apache.nifi.processors.aws.kinesis.ConsumeKinesisAttributes.SHARD_ID;
+import static 
org.apache.nifi.processors.aws.kinesis.ConsumeKinesisAttributes.SUB_SEQUENCE_NUMBER;
+
+@InputRequirement(InputRequirement.Requirement.INPUT_FORBIDDEN)
+@Tags({"amazon", "aws", "kinesis", "consume", "stream", "record"})
+@CapabilityDescription("Consumes data from the specified AWS Kinesis stream 
and outputs a FlowFile for every processed Record (raw) " +
+        "or a FlowFile for a batch of processed records if a Record Reader and 
Record Writer are configured. " +
+        "AWS Kinesis Client Library can take several seconds to initialise 
before starting to fetch data. " +
+        "Uses DynamoDB for check pointing and coordination, and CloudWatch 
(optional) for metrics. " +
+        "Ensure that the credentials provided have access to DynamoDB and 
CloudWatch (optional) along with Kinesis.")
+@WritesAttributes({
+        @WritesAttribute(attribute = SHARD_ID,
+                description = "Shard ID from which all Kinesis Records in the 
Flow File were read"),
+        @WritesAttribute(attribute = PARTITION_KEY,
+                description = "Partition key of the last Kinesis Record in the 
Flow File"),
+        @WritesAttribute(attribute = SEQUENCE_NUMBER,
+                description = "A Sequence Number of the last Kinesis Record in 
the Flow File"),
+        @WritesAttribute(attribute = SUB_SEQUENCE_NUMBER,
+                description = "A SubSequence Number of the last Kinesis Record 
in the Flow File. Generated by KPL when aggregating records into a single 
Kinesis Record"),
+        @WritesAttribute(attribute = APPROXIMATE_ARRIVAL_TIMESTAMP,
+                description = "Approximate arrival timestamp of the last 
Kinesis Record in the Flow File"),
+        @WritesAttribute(attribute = "mime.type",
+                description = "Sets the mime.type attribute to the MIME Type 
specified by the Record Writer (if configured)"),
+        @WritesAttribute(attribute = RECORD_COUNT,
+                description = "Number of records written to the FlowFiles by 
the Record Writer (if configured)"),
+        @WritesAttribute(attribute = RECORD_ERROR_MESSAGE,
+                description = "This attribute provides on failure the error 
message encountered by the Record Reader or Record Writer (if configured)")
+})
+@DefaultSettings(yieldDuration = "100 millis")
+@SystemResourceConsideration(resource = SystemResource.CPU, description = 
"Kinesis Client Library is used to create a Worker thread for consumption of 
Kinesis Records. " +
+        "The Worker is initialised and started when this Processor has been 
scheduled. It runs continually, spawning Kinesis Record Processors as required 
" +
+        "to fetch Kinesis Records. The Worker Thread (and any child Record 
Processor threads) are not controlled by the normal NiFi scheduler as part of 
the " +
+        "Concurrent Thread pool and are not released until this processor is 
stopped.")
+@SystemResourceConsideration(resource = SystemResource.NETWORK, description = 
"Kinesis Client Library will continually poll for new Records, " +
+        "requesting up to a maximum number of Records/bytes per call. This can 
result in sustained network usage.")
+@SystemResourceConsideration(resource = SystemResource.MEMORY, description = 
"ConsumeKinesis buffers Kinesis Records in memory until they can be processed. 
" +
+        "The maximum size of the buffer is controlled by the 'Max Bytes to 
Buffer' property.")
+public class ConsumeKinesis extends AbstractProcessor {
+
+    public enum InitialPosition implements DescribedValue {
+        TRIM_HORIZON("Trim Horizon", "Start reading at the last untrimmed 
record in the shard in the system, which is the oldest data record in the 
shard."),
+        LATEST("Latest", "Start reading just after the most recent record in 
the shard, so that you always read the most recent data in the shard."),
+        AT_TIMESTAMP("At Timestamp", "Start reading at the record with the 
specified timestamp.");
+
+        private final String displayName;
+        private final String description;
+
+        InitialPosition(final String displayName, final String description) {
+            this.displayName = displayName;
+            this.description = description;
+        }
+
+        @Override
+        public String getValue() {
+            return name();
+        }
+
+        @Override
+        public String getDisplayName() {
+            return displayName;
+        }
+
+        @Override
+        public String getDescription() {
+            return description;
+        }
+    }
+
+    static final PropertyDescriptor KINESIS_STREAM_NAME = new 
PropertyDescriptor.Builder()
+            .name("Amazon Kinesis Stream Name")
+            .description("The name of the Kinesis stream to consume from.")
+            .required(true)
+            .addValidator(StandardValidators.NON_EMPTY_VALIDATOR)
+            .build();
+
+    static final PropertyDescriptor APPLICATION_NAME = new 
PropertyDescriptor.Builder()
+            .name("Application Name")
+            .description("The name of the Kinesis application. This is used 
for DynamoDB table naming and worker coordination.")
+            .required(true)
+            .addValidator(StandardValidators.NON_EMPTY_VALIDATOR)
+            .build();
+
+    static final PropertyDescriptor AWS_CREDENTIALS_PROVIDER_SERVICE = new 
PropertyDescriptor.Builder()
+            .name("AWS Credentials Provider service")
+            .displayName("AWS Credentials Provider Service")
+            .description("The Controller Service that is used to obtain AWS 
credentials provider.")
+            .required(true)
+            .identifiesControllerService(AWSCredentialsProviderService.class)
+            .build();
+
+    static final PropertyDescriptor REGION = new PropertyDescriptor.Builder()
+            .name("Region")
+            .description("AWS Region in which the Kinesis stream is located.")
+            .required(true)
+            .allowableValues(RegionUtilV2.getAvailableRegions())
+            
.defaultValue(RegionUtilV2.createAllowableValue(Region.US_WEST_2).getValue())
+            .build();
+
+    static final PropertyDescriptor RECORD_READER = new 
PropertyDescriptor.Builder()
+            .name("Record Reader")
+            .description("""
+                    The Record Reader to use for parsing the data received 
from Kinesis.
+                    If not set, the records are written one per FlowFile 
without parsing.""")
+            .required(false)
+            .identifiesControllerService(RecordReaderFactory.class)
+            .build();
+
+    static final PropertyDescriptor RECORD_WRITER = new 
PropertyDescriptor.Builder()
+            .name("Record Writer")
+            .description("The Record Writer to use for serializing records.")
+            .required(true)
+            .dependsOn(RECORD_READER)
+            .identifiesControllerService(RecordSetWriterFactory.class)
+            .build();
+
+    static final PropertyDescriptor INITIAL_STREAM_POSITION = new 
PropertyDescriptor.Builder()
+            .name("Initial Stream Position")
+            .description("The position in the stream where the processor 
should start reading.")
+            .required(true)
+            .allowableValues(InitialPosition.class)
+            .defaultValue(InitialPosition.TRIM_HORIZON)
+            .build();
+
+    static final PropertyDescriptor STREAM_POSITION_TIMESTAMP = new 
PropertyDescriptor.Builder()
+            .name("Stream Position Timestamp")
+            .description("Timestamp position in stream from which to start 
reading Kinesis Records. The timestamp must be in ISO 8601 format.")
+            .addValidator(StandardValidators.ISO8601_INSTANT_VALIDATOR)
+            .dependsOn(INITIAL_STREAM_POSITION, InitialPosition.AT_TIMESTAMP)
+            .required(true)
+            .build();
+
+    static final PropertyDescriptor MAX_BYTES_TO_BUFFER = new 
PropertyDescriptor.Builder()
+            .name("Max Bytes to Buffer")
+            .description("""
+                    The maximum size of Kinesis Records that can be buffered 
in memory before being processed by NiFi.
+                    If the buffer size exceeds the limit, the KCL will stop 
consuming new records until free space is available.
+
+                    Using a larger value may improve throughput, but will do 
so at the expense of using additional heap.
+                    Using a smaller value may back off the Kinesis Client 
Library (KCL) from consuming records if the buffer is full, which may result in 
lower throughput.
+                    """)
+            .required(true)
+            .addValidator(StandardValidators.DATA_SIZE_VALIDATOR)
+            .defaultValue("100 MB")
+            .build();
+
+    static final PropertyDescriptor CHECKPOINT_INTERVAL = new 
PropertyDescriptor.Builder()
+            .name("Checkpoint Interval")
+            .description("""
+                    Interval between checkpointing consumed Kinesis records. 
To checkpoint records on each NiFi session commit, set this value to 0 seconds.
+
+                    Checkpointing too frequently may result in performance 
degradation and higher DynamoDB costs.
+                    Checkpointing too rarely may result in duplicated records 
whenever a Shard lease is lost or NiFi server restarts.
+                    """)
+            .addValidator(StandardValidators.TIME_PERIOD_VALIDATOR)
+            .defaultValue("5 sec")
+            .required(true)
+            .build();
+
+    static final PropertyDescriptor TIMEOUT = new PropertyDescriptor.Builder()
+            .name("Communications Timeout")
+            .description("""
+                    Timeout for communication with AWS Kinesis, DynamoDB, and 
CloudWatch.
+                    This timeout is applied to all requests made by this 
processor to AWS services.""")
+            .required(true)
+            .addValidator(StandardValidators.TIME_PERIOD_VALIDATOR)
+            .defaultValue("30 secs")
+            .build();
+
+    static final PropertyDescriptor ENDPOINT_OVERRIDE = new 
PropertyDescriptor.Builder()
+            .name("Kinesis Endpoint Override URL")
+            .description("""
+                    Endpoint URL to use instead of the AWS default including 
scheme, host, port, and path.
+                    The AWS libraries select an endpoint URL based on the AWS 
region, but this property overrides the selected endpoint URL, allowing use 
with other S3-compatible endpoints.""")
+            .expressionLanguageSupported(ExpressionLanguageScope.ENVIRONMENT)
+            .required(false)
+            .addValidator(StandardValidators.URL_VALIDATOR)
+            .build();
+
+    static final PropertyDescriptor DYNAMODB_ENDPOINT_OVERRIDE = new 
PropertyDescriptor.Builder()
+            .name("Dynamo DB Endpoint Override")
+            .description("An optional endpoint URL to use for DynamoDB. If not 
specified, the default AWS endpoint for the region will be used.")
+            .addValidator(StandardValidators.URL_VALIDATOR)
+            .expressionLanguageSupported(ExpressionLanguageScope.ENVIRONMENT)
+            .required(false)
+            .build();
+
+    static final PropertyDescriptor REPORT_CLOUDWATCH_METRICS = new 
PropertyDescriptor.Builder()
+            .name("Report Metrics to CloudWatch")
+            .description("Whether to report Kinesis usage metrics to 
CloudWatch.")
+            .addValidator(StandardValidators.BOOLEAN_VALIDATOR)
+            .allowableValues("true", "false")
+            .defaultValue("false")
+            .required(true)
+            .build();
+
+    static final PropertyDescriptor CLOUDWATCH_ENDPOINT_OVERRIDE = new 
PropertyDescriptor.Builder()
+            .name("CloudWatch Endpoint Override")
+            .description("An optional endpoint URL to use for CloudWatch. If 
not specified, the default AWS endpoint for the region will be used.")
+            .addValidator(StandardValidators.URL_VALIDATOR)
+            .expressionLanguageSupported(ExpressionLanguageScope.ENVIRONMENT)
+            .required(false)
+            .build();
+
+    static final PropertyDescriptor PROXY_CONFIGURATION_SERVICE = 
ProxyConfiguration.createProxyConfigPropertyDescriptor(ProxySpec.HTTP, 
ProxySpec.HTTP_AUTH);
+
+    private static final List<PropertyDescriptor> PROPERTY_DESCRIPTORS = 
List.of(
+            KINESIS_STREAM_NAME,
+            APPLICATION_NAME,
+            AWS_CREDENTIALS_PROVIDER_SERVICE,
+            REGION,
+            RECORD_READER,
+            RECORD_WRITER,
+            INITIAL_STREAM_POSITION,
+            STREAM_POSITION_TIMESTAMP,
+            MAX_BYTES_TO_BUFFER,
+            CHECKPOINT_INTERVAL,
+            TIMEOUT,
+            ENDPOINT_OVERRIDE,
+            DYNAMODB_ENDPOINT_OVERRIDE,
+            CLOUDWATCH_ENDPOINT_OVERRIDE,
+            PROXY_CONFIGURATION_SERVICE,
+            REPORT_CLOUDWATCH_METRICS
+    );
+
+    static final Relationship REL_SUCCESS = new Relationship.Builder()
+            .name("success")
+            .description("FlowFiles that are created when records are 
successfully read from Kinesis and parsed.")
+            .build();
+
+    static final Relationship REL_PARSE_FAILURE = new Relationship.Builder()
+            .name("parse.failure")
+            .description("FlowFiles that failed to parse using the configured 
Record Reader.")
+            .build();
+
+    private static final Set<Relationship> RAW_FILE_RELATIONSHIPS = 
Set.of(REL_SUCCESS);
+    private static final Set<Relationship> RECORD_FILE_RELATIONSHIPS = 
Set.of(REL_SUCCESS, REL_PARSE_FAILURE);
+
+    private volatile DynamoDbAsyncClient dynamoDbClient;
+    private volatile CloudWatchAsyncClient cloudWatchClient;
+    private volatile KinesisAsyncClient kinesisClient;
+    private volatile Scheduler kinesisScheduler;
+
+    private volatile RecordBuffer recordBuffer;
+    private volatile boolean useReader = false;
+    private volatile Optional<ReaderRecordProcessor> readerRecordProcessor = 
Optional.empty();
+
+    @Override
+    protected List<PropertyDescriptor> getSupportedPropertyDescriptors() {
+        return PROPERTY_DESCRIPTORS;
+    }
+
+    @Override
+    public Set<Relationship> getRelationships() {
+        return useReader ? RECORD_FILE_RELATIONSHIPS : RAW_FILE_RELATIONSHIPS;
+    }
+
+    @Override
+    public void onPropertyModified(final PropertyDescriptor descriptor, final 
String oldValue, final String newValue) {
+        if (descriptor.equals(RECORD_READER)) {
+            useReader = newValue != null;
+        }
+    }
+
+    @OnScheduled
+    public void setup(final ProcessContext context) {
+        final RecordReaderFactory recordReaderFactory = 
context.getProperty(RECORD_READER).asControllerService(RecordReaderFactory.class);
+        if (recordReaderFactory != null) {
+            final RecordSetWriterFactory recordWriterFactory = 
context.getProperty(RECORD_WRITER).asControllerService(RecordSetWriterFactory.class);
+            readerRecordProcessor = Optional.of(new 
ReaderRecordProcessor(recordReaderFactory, recordWriterFactory, getLogger()));
+        }
+
+        final Region region = 
Region.of(context.getProperty(REGION).getValue());
+        final AwsCredentialsProvider credentialsProvider = 
context.getProperty(AWS_CREDENTIALS_PROVIDER_SERVICE)
+                
.asControllerService(AWSCredentialsProviderService.class).getAwsCredentialsProvider();
+
+        final String kinesisEndpointOverride = 
context.getProperty(ENDPOINT_OVERRIDE).evaluateAttributeExpressions().getValue();
+        final URI kinesisEndpoint = kinesisEndpointOverride == null ? null : 
URI.create(kinesisEndpointOverride);
+        kinesisClient = KinesisAsyncClient.builder()
+                .region(region)
+                .credentialsProvider(credentialsProvider)
+                .endpointOverride(kinesisEndpoint)
+                .httpClient(createKinesisHttpClient(context))
+                .build();
+
+        final String dynamoDbEndpointOverride = 
context.getProperty(DYNAMODB_ENDPOINT_OVERRIDE).evaluateAttributeExpressions().getValue();
+        final URI dynamoDbEndpoint = dynamoDbEndpointOverride == null ? null : 
URI.create(dynamoDbEndpointOverride);
+        dynamoDbClient = DynamoDbAsyncClient.builder()
+                .region(region)
+                .credentialsProvider(credentialsProvider)
+                .endpointOverride(dynamoDbEndpoint)
+                .httpClient(createHttpClientBuilder(context).build())
+                .build();
+
+        final String cloudwatchEndpointOverride = 
context.getProperty(CLOUDWATCH_ENDPOINT_OVERRIDE).evaluateAttributeExpressions().getValue();
+        final URI cloudWatchEndpoint = cloudwatchEndpointOverride == null ? 
null : URI.create(cloudwatchEndpointOverride);
+        cloudWatchClient = CloudWatchAsyncClient.builder()
+                .region(region)
+                .credentialsProvider(credentialsProvider)
+                .endpointOverride(cloudWatchEndpoint)
+                .httpClient(createHttpClientBuilder(context).build())
+                .build();
+
+        final String streamName = 
context.getProperty(KINESIS_STREAM_NAME).getValue();
+        final InitialPositionInStreamExtended initialPositionExtended = 
getInitialPosition(context);
+        final SingleStreamTracker streamTracker = new 
SingleStreamTracker(streamName, initialPositionExtended);
+
+        final long maxBytesToBuffer = 
context.getProperty(MAX_BYTES_TO_BUFFER).asDataSize(DataUnit.B).longValue();
+        final Duration checkpointInterval = 
context.getProperty(CHECKPOINT_INTERVAL).asDuration();
+        recordBuffer = new RecordBuffer(getLogger(), maxBytesToBuffer, 
checkpointInterval);
+        final ShardRecordProcessorFactory recordProcessorFactory = () -> new 
ConsumeKinesisRecordProcessor(recordBuffer);
+
+        final String applicationName = 
context.getProperty(APPLICATION_NAME).getValue();
+        final String workerId = UUID.randomUUID().toString();
+        final ConfigsBuilder configsBuilder = new 
ConfigsBuilder(streamTracker, applicationName, kinesisClient, dynamoDbClient, 
cloudWatchClient, workerId, recordProcessorFactory);
+
+        if (!context.getProperty(REPORT_CLOUDWATCH_METRICS).asBoolean()) {
+            configsBuilder.metricsConfig().metricsFactory(null);
+        }
+
+        kinesisScheduler = new Scheduler(
+                configsBuilder.checkpointConfig(),
+                configsBuilder.coordinatorConfig(),
+                configsBuilder.leaseManagementConfig(),
+                configsBuilder.lifecycleConfig(),
+                configsBuilder.metricsConfig(),
+                configsBuilder.processorConfig(),
+                configsBuilder.retrievalConfig()
+        );
+
+        final Thread schedulerThread = new Thread(kinesisScheduler);
+        schedulerThread.setDaemon(true);
+        schedulerThread.start();
+    }
+
+    /**
+     * Creating Kinesis HTTP client, as per
+     * {@link 
software.amazon.kinesis.common.KinesisClientUtil#adjustKinesisClientBuilder(KinesisAsyncClientBuilder)}.
+     */
+    private static SdkAsyncHttpClient createKinesisHttpClient(final 
ProcessContext context) {
+        return createHttpClientBuilder(context)
+                .maxConcurrency(Integer.MAX_VALUE)

Review Comment:
   The optimal setting may vary on a number of factors, including the number of 
CPU cores, but more particularly based on other things being processed in the 
framework. I agree that avoiding a configuration property is optimal, although 
not out of the question. Did you consider making it a multiple of the 
concurrent tasks for the Processor?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to