ChrisSamo632 commented on a change in pull request #4822: URL: https://github.com/apache/nifi/pull/4822#discussion_r627180647
########## File path: nifi-nar-bundles/nifi-aws-bundle/nifi-aws-processors/src/main/java/org/apache/nifi/processors/aws/kinesis/stream/ConsumeKinesisStream.java ########## @@ -0,0 +1,680 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.nifi.processors.aws.kinesis.stream; + +import com.amazonaws.ClientConfiguration; +import com.amazonaws.auth.AWSCredentials; +import com.amazonaws.auth.AWSCredentialsProvider; +import com.amazonaws.auth.AWSStaticCredentialsProvider; +import com.amazonaws.services.kinesis.AmazonKinesisClient; +import com.amazonaws.services.kinesis.clientlibrary.interfaces.v2.IRecordProcessorFactory; +import com.amazonaws.services.kinesis.clientlibrary.lib.worker.InitialPositionInStream; +import com.amazonaws.services.kinesis.clientlibrary.lib.worker.KinesisClientLibConfiguration; +import com.amazonaws.services.kinesis.clientlibrary.lib.worker.Worker; +import com.amazonaws.services.kinesis.metrics.impl.NullMetricsFactory; +import org.apache.commons.beanutils.BeanUtilsBean; +import org.apache.commons.beanutils.ConvertUtilsBean2; +import org.apache.commons.beanutils.FluentPropertyBeanIntrospector; +import org.apache.commons.beanutils.PropertyUtilsBean; +import org.apache.commons.lang3.StringUtils; +import org.apache.nifi.annotation.behavior.DynamicProperties; +import org.apache.nifi.annotation.behavior.DynamicProperty; +import org.apache.nifi.annotation.behavior.InputRequirement; +import org.apache.nifi.annotation.behavior.SystemResource; +import org.apache.nifi.annotation.behavior.SystemResourceConsideration; +import org.apache.nifi.annotation.behavior.WritesAttribute; +import org.apache.nifi.annotation.behavior.WritesAttributes; +import org.apache.nifi.annotation.documentation.CapabilityDescription; +import org.apache.nifi.annotation.documentation.SeeAlso; +import org.apache.nifi.annotation.documentation.Tags; +import org.apache.nifi.annotation.lifecycle.OnScheduled; +import org.apache.nifi.annotation.lifecycle.OnStopped; +import org.apache.nifi.components.AllowableValue; +import org.apache.nifi.components.PropertyDescriptor; +import org.apache.nifi.components.ValidationContext; +import org.apache.nifi.components.ValidationResult; +import org.apache.nifi.controller.ControllerService; +import org.apache.nifi.expression.ExpressionLanguageScope; +import org.apache.nifi.processor.ProcessContext; +import org.apache.nifi.processor.ProcessSession; +import org.apache.nifi.processor.ProcessSessionFactory; +import org.apache.nifi.processor.Relationship; +import org.apache.nifi.processor.exception.ProcessException; +import org.apache.nifi.processor.util.StandardValidators; +import org.apache.nifi.processors.aws.kinesis.stream.record.AbstractKinesisRecordProcessor; +import org.apache.nifi.processors.aws.kinesis.stream.record.KinesisRecordProcessorRaw; +import org.apache.nifi.processors.aws.kinesis.stream.record.KinesisRecordProcessorRecord; +import org.apache.nifi.serialization.RecordReaderFactory; +import org.apache.nifi.serialization.RecordSetWriterFactory; +import org.apache.nifi.serialization.record.RecordFieldType; + +import java.lang.reflect.InvocationTargetException; +import java.net.InetAddress; +import java.net.UnknownHostException; +import java.time.LocalDateTime; +import java.time.ZoneOffset; +import java.time.format.DateTimeFormatter; +import java.util.Arrays; +import java.util.Collection; +import java.util.Collections; +import java.util.Date; +import java.util.HashMap; +import java.util.HashSet; +import java.util.List; +import java.util.Map; +import java.util.Optional; +import java.util.Set; +import java.util.UUID; +import java.util.concurrent.ExecutionException; +import java.util.concurrent.Future; +import java.util.concurrent.TimeUnit; +import java.util.concurrent.TimeoutException; +import java.util.concurrent.atomic.AtomicBoolean; +import java.util.stream.Collectors; + +@InputRequirement(InputRequirement.Requirement.INPUT_FORBIDDEN) +@Tags({"amazon", "aws", "kinesis", "get", "stream"}) +@CapabilityDescription("Reads data from the specified AWS Kinesis stream and outputs a FlowFile for every processed Record (raw) " + + " or a FlowFile for a batch of processed records if a Record Reader and Record Writer are configured. " + + "At-least-once delivery of all Kinesis Records within the Stream while the processor is running. " + + "AWS Kinesis Client Library can take several seconds to initialise before starting to fetch data. " + + "Uses DynamoDB for check pointing and CloudWatch (optional) for metrics. " + + "Ensure that the credentials provided have access to DynamoDB and CloudWatch (if used) along with Kinesis.") +@WritesAttributes({ + @WritesAttribute(attribute = AbstractKinesisRecordProcessor.AWS_KINESIS_PARTITION_KEY, + description = "Partition from which the (last) Kinesis Record was read"), + @WritesAttribute(attribute = AbstractKinesisRecordProcessor.AWS_KINESIS_SEQUENCE_NUMBER, + description = "The unique identifier of the (last) Kinesis Record within its shard"), + @WritesAttribute(attribute = AbstractKinesisRecordProcessor.AWS_KINESIS_APPROXIMATE_ARRIVAL_TIMESTAMP, + description = "Approximate arrival timestamp of the (last) Kinesis Record read from the stream"), + @WritesAttribute(attribute = AbstractKinesisRecordProcessor.AWS_KINESIS_SHARD_ID, + description = "Shard ID from which the Kinesis Record was read"), + @WritesAttribute(attribute = "mime.type", + description = "Sets the mime.type attribute to the MIME Type specified by the Record Writer (if configured)"), + @WritesAttribute(attribute = "record.count", + description = "Number of records written to the FlowFiles by the Record Writer (if configured)"), + @WritesAttribute(attribute = "record.error.message", + description = "This attribute provides on failure the error message encountered by the Record Reader or Record Writer (if configured)") +}) +@DynamicProperties({ + @DynamicProperty(name="Kinesis Client Library (KCL) Configuration property name", + description="Override default KCL Configuration properties with required values. Supports setting of values via the \"with\" " + + "methods on the KCL Configuration class. Specify the property to be set without the leading prefix, e.g. \"maxInitialisationAttempts\" " + + "will call \"withMaxInitialisationAttempts\" and set the provided value. Only supports setting of simple property values, e.g. String, " + + "int, long and boolean. Does not allow override of KCL Configuration settings handled by non-dynamic processor properties.", + expressionLanguageScope = ExpressionLanguageScope.NONE, value="Value to set in the KCL Configuration property") +}) +@SystemResourceConsideration(resource = SystemResource.CPU, description = "Kinesis Client Library is used to create a Worker thread for consumption of Kinesis Records. " + + "The Worker is initialised and started when this Processor has been triggered. It runs continually, spawning Kinesis Record Processors as required " + + "to fetch Kinesis Records. The Worker Thread (and any child Record Processor threads) is not released until this processor is stopped. " + + "This means a NiFi Concurrent Thread is permanently assigned to this Processor while it is running and other threads will be created within the JVM " + + "that are not controlled by the normal NiFi scheduler.") +@SystemResourceConsideration(resource = SystemResource.NETWORK, description = "Kinesis Client Library will continually poll for new Records, " + + "requesting up to a maximum number of Records/bytes per call. This can result in sustained network usage.") +@SeeAlso(PutKinesisStream.class) +@SuppressWarnings({"java:S110", "java:S2160", "java:S3077"}) +public class ConsumeKinesisStream extends AbstractKinesisStreamProcessor { + static final AllowableValue TRIM_HORIZON = new AllowableValue( + InitialPositionInStream.TRIM_HORIZON.toString(), + InitialPositionInStream.TRIM_HORIZON.toString(), + "Start reading at the last untrimmed record in the shard in the system, which is the oldest data record in the shard." + ); + static final AllowableValue LATEST = new AllowableValue( + InitialPositionInStream.LATEST.toString(), + InitialPositionInStream.LATEST.toString(), + "Start reading just after the most recent record in the shard, so that you always read the most recent data in the shard." + ); + static final AllowableValue AT_TIMESTAMP = new AllowableValue( + InitialPositionInStream.AT_TIMESTAMP.toString(), + InitialPositionInStream.AT_TIMESTAMP.toString(), "Start reading from the position denoted by a specific time stamp, provided in the value Timestamp." + ); + + public static final PropertyDescriptor APPLICATION_NAME = new PropertyDescriptor.Builder() + .displayName("Application Name") + .name("amazon-kinesis-stream-application-name") + .description("The Kinesis stream reader application name.") + .addValidator(StandardValidators.NON_BLANK_VALIDATOR) + .required(true).build(); + + public static final PropertyDescriptor INITIAL_STREAM_POSITION = new PropertyDescriptor.Builder() + .displayName("Initial Stream Position") + .name("amazon-kinesis-stream-initial-position") + .description("Initial position to read Kinesis streams.") + .allowableValues(LATEST, TRIM_HORIZON, AT_TIMESTAMP) + .defaultValue(LATEST.getValue()) + .required(true).build(); + + public static final PropertyDescriptor STREAM_POSITION_TIMESTAMP = new PropertyDescriptor.Builder() + .displayName("Stream Position Timestamp") + .name("amazon-kinesis-stream-position-timestamp") + .description("Timestamp position in stream from which to start reading Kinesis Records. " + + "Required if " + INITIAL_STREAM_POSITION.getDescription() + " is " + AT_TIMESTAMP.getDisplayName() + ". " + + "Uses the Timestamp Format to parse value into a Date.") + .addValidator(StandardValidators.NON_BLANK_VALIDATOR) // customValidate checks the value against TIMESTAMP_FORMAT + .dependsOn(INITIAL_STREAM_POSITION, AT_TIMESTAMP) + .required(false).build(); + + public static final PropertyDescriptor TIMESTAMP_FORMAT = new PropertyDescriptor.Builder() + .displayName("Timestamp Format") + .name("amazon-kinesis-stream-timestamp-format") + .description("Format to use for parsing the " + STREAM_POSITION_TIMESTAMP.getDisplayName() + " into a Date " + + "and converting the Kinesis Record's Approximate Arrival Timestamp into a FlowFile attribute.") + .addValidator((subject, input, context) -> { + if (StringUtils.isNotBlank(input)) { + try { + DateTimeFormatter.ofPattern(input); + } catch (Exception e) { + return new ValidationResult.Builder().valid(false).subject(subject).input(input) + .explanation("Must be a valid java.time.DateTimeFormatter pattern, e.g. " + RecordFieldType.TIMESTAMP.getDefaultFormat()) + .build(); + } + } + return new ValidationResult.Builder().valid(true).subject(subject).build(); + }) + .defaultValue(RecordFieldType.TIMESTAMP.getDefaultFormat()) + .expressionLanguageSupported(ExpressionLanguageScope.VARIABLE_REGISTRY) + .required(true).build(); + + public static final PropertyDescriptor CHECKPOINT_INTERVAL_MILLIS = new PropertyDescriptor.Builder() + .displayName("Checkpoint Interval") + .name("amazon-kinesis-stream-checkpoint-interval") + .description("Interval (milliseconds) between Kinesis checkpoints") + .addValidator(StandardValidators.LONG_VALIDATOR) + .defaultValue("60000") + .required(true).build(); + + public static final PropertyDescriptor NUM_RETRIES = new PropertyDescriptor.Builder() + .displayName("Retry Count") + .name("amazon-kinesis-stream-retry-count") + .description("Number of times to retry a Kinesis operation (process record, checkpoint, shutdown)") + .addValidator(StandardValidators.INTEGER_VALIDATOR) + .defaultValue("10") + .required(true).build(); + + public static final PropertyDescriptor RETRY_WAIT_MILLIS = new PropertyDescriptor.Builder() + .displayName("Retry Wait") + .name("amazon-kinesis-stream-retry-wait") + .description("Interval (milliseconds) between Kinesis operation retries (get records, checkpoint, shutdown)") + .addValidator(StandardValidators.LONG_VALIDATOR) + .defaultValue("1000") + .required(true).build(); + + public static final PropertyDescriptor DYNAMODB_ENDPOINT_OVERRIDE = new PropertyDescriptor.Builder() + .displayName("DynamoDB Override") + .name("amazon-kinesis-stream-dynamodb-override") + .description("DynamoDB override to use non-AWS deployments") + .addValidator(StandardValidators.URL_VALIDATOR) + .expressionLanguageSupported(ExpressionLanguageScope.VARIABLE_REGISTRY) + .required(false).build(); + + public static final PropertyDescriptor REPORT_CLOUDWATCH_METRICS = new PropertyDescriptor.Builder() + .displayName("Report Metrics to CloudWatch") + .name("amazon-kinesis-stream-cloudwatch-flag") + .description("Whether to report Kinesis usage metrics to CloudWatch.") + .addValidator(StandardValidators.BOOLEAN_VALIDATOR) + .allowableValues("true", "false") + .defaultValue("false") + .required(true).build(); + + public static final PropertyDescriptor RECORD_READER = new PropertyDescriptor.Builder() + .name("amazon-kinesis-stream-record-reader") + .displayName("Record Reader") + .description("The Record Reader to use for reading received messages." + + " The Kinesis Stream name can be referred to by Expression Language '${" + + AbstractKinesisRecordProcessor.KINESIS_RECORD_SCHEMA_KEY + "}' to access a schema." + + " If Record Reader/Writer are not specified, each Kinesis Record will create a FlowFile.") + .identifiesControllerService(RecordReaderFactory.class) + .expressionLanguageSupported(ExpressionLanguageScope.NONE) + .required(false) + .build(); + + public static final PropertyDescriptor RECORD_WRITER = new PropertyDescriptor.Builder() + .name("amazon-kinesis-stream-record-writer") + .displayName("Record Writer") + .description("The Record Writer to use for serializing Records to an output FlowFile." + + " The Kinesis Stream name can be referred to by Expression Language '${" + + AbstractKinesisRecordProcessor.KINESIS_RECORD_SCHEMA_KEY + "}' to access a schema." + + " If Record Reader/Writer are not specified, each Kinesis Record will create a FlowFile.") + .identifiesControllerService(RecordSetWriterFactory.class) + .expressionLanguageSupported(ExpressionLanguageScope.NONE) + .required(false) + .build(); + + public static final Relationship REL_PARSE_FAILURE = new Relationship.Builder() + .name("parse.failure") + .description("If a message from Kinesis cannot be parsed using the configured Record Reader" + + " or failed to be written by the configured Record Writer," + + " the contents of the message will be routed to this Relationship as its own individual FlowFile.") + .build(); + + public static final List<PropertyDescriptor> properties = Collections.unmodifiableList( + Arrays.asList( + // Kinesis Stream specific properties + KINESIS_STREAM_NAME, APPLICATION_NAME, RECORD_READER, RECORD_WRITER, REGION, ENDPOINT_OVERRIDE, + DYNAMODB_ENDPOINT_OVERRIDE, INITIAL_STREAM_POSITION, STREAM_POSITION_TIMESTAMP, TIMESTAMP_FORMAT, + CHECKPOINT_INTERVAL_MILLIS, NUM_RETRIES, RETRY_WAIT_MILLIS, REPORT_CLOUDWATCH_METRICS, + // generic AWS processor properties + TIMEOUT, AWS_CREDENTIALS_PROVIDER_SERVICE, ACCESS_KEY, SECRET_KEY, CREDENTIALS_FILE, + PROXY_CONFIGURATION_SERVICE, PROXY_HOST, PROXY_HOST_PORT, PROXY_USERNAME, PROXY_PASSWORD + ) + ); + + @SuppressWarnings({"java:S3599", "java:S1171"}) + private static final Map<String, PropertyDescriptor> DISALLOWED_DYNAMIC_KCL_PROPERTIES = new HashMap<String, PropertyDescriptor>(){{ + put("regionName", REGION); + put("timestampAtInitialPositionInStream", STREAM_POSITION_TIMESTAMP); + put("initialPositionInStream", INITIAL_STREAM_POSITION); + put("dynamoDBEndpoint", DYNAMODB_ENDPOINT_OVERRIDE); + put("kinesisEndpoint", KINESIS_STREAM_NAME); + }}; + + private static final Set<Relationship> RELATIONSHIPS = Collections.singleton(REL_SUCCESS); + private static final Set<Relationship> RECORD_RELATIONSHIPS = Collections.unmodifiableSet(new HashSet<>(Arrays.asList(REL_SUCCESS, REL_PARSE_FAILURE))); + + private final PropertyUtilsBean propertyUtilsBean; + private final BeanUtilsBean beanUtilsBean; + + private volatile long retryWaitMillis; + private volatile int numRetries; + private volatile InitialPositionInStream initialPositionInStream; + private volatile DateTimeFormatter dateTimeFormatter; + private volatile Date startStreamPositionTimestamp; + private volatile AWSCredentials awsCredentials; + + private volatile RecordReaderFactory readerFactory; + private volatile RecordSetWriterFactory writerFactory; + private volatile boolean isRecordReaderSet = false; + private volatile boolean isRecordWriterSet = false; + + private volatile Worker worker; + + public ConsumeKinesisStream() { + propertyUtilsBean = new PropertyUtilsBean(); + propertyUtilsBean.addBeanIntrospector(new FluentPropertyBeanIntrospector("with")); + + final ConvertUtilsBean2 convertUtilsBean2 = new ConvertUtilsBean2() { + @SuppressWarnings("unchecked") // generic Enum conversion from String property values + @Override + public Object convert(final String value, final Class clazz) { + if (clazz.isEnum()) { + return Enum.valueOf(clazz, value); + }else{ + return super.convert(value, clazz); + } + } + }; + + beanUtilsBean = new BeanUtilsBean(convertUtilsBean2, propertyUtilsBean); + } + + @Override + protected List<PropertyDescriptor> getSupportedPropertyDescriptors() { + return properties; + } + + @Override + public Set<Relationship> getRelationships() { + return isRecordReaderSet && isRecordWriterSet ? RECORD_RELATIONSHIPS : RELATIONSHIPS; + } + + @Override + public void onPropertyModified(PropertyDescriptor descriptor, String oldValue, String newValue) { + if (RECORD_READER.equals(descriptor)) { + isRecordReaderSet = StringUtils.isNotEmpty(newValue); + } else if (RECORD_WRITER.equals(descriptor)) { + isRecordWriterSet = StringUtils.isNotEmpty(newValue); + } + } + + @Override + protected PropertyDescriptor getSupportedDynamicPropertyDescriptor(final String propertyDescriptorName) { + final PropertyDescriptor.Builder builder = new PropertyDescriptor.Builder() + .name(propertyDescriptorName) + .required(false) + .dynamic(true) + .addValidator(StandardValidators.ATTRIBUTE_KEY_PROPERTY_NAME_VALIDATOR) + .addValidator(this::validateDynamicKCLConfigProperty) + .expressionLanguageSupported(ExpressionLanguageScope.NONE); + + return builder.build(); + } + + private ValidationResult validateDynamicKCLConfigProperty(final String subject, final String input, final ValidationContext context) { + final ValidationResult.Builder validationResult = new ValidationResult.Builder().subject(subject).input(input); + + if (!subject.matches("^(?!with)[a-z]\\w*$")) { + return validationResult + .explanation("Property name must not have a prefix of \"with\", must start with a lowercase letter and contain only letters, numbers or underscores") + .valid(false).build(); + } + + if (DISALLOWED_DYNAMIC_KCL_PROPERTIES.keySet().stream().anyMatch(k -> k.equalsIgnoreCase(subject))) { + return validationResult + .explanation(String.format("Use \"%s\" instead of a dynamic property", DISALLOWED_DYNAMIC_KCL_PROPERTIES.get(subject).getDisplayName())) + .valid(false).build(); + } + + @SuppressWarnings("java:S1192") + final KinesisClientLibConfiguration kclTemp = new KinesisClientLibConfiguration("validate", "validate", null, "validate"); + try { + if (!propertyUtilsBean.isWriteable(kclTemp, subject)) { + return validationResult + .explanation(String.format("Kinesis Client Library Configuration property with name with%s does not exist or is not writable", StringUtils.capitalize(subject))) + .valid(false).build(); + } + beanUtilsBean.setProperty(kclTemp, subject, input); + } catch (IllegalAccessException e) { + return validationResult + .explanation(String.format("Kinesis Client Library Configuration property with name with%s is not accessible", StringUtils.capitalize(subject))) + .valid(false).build(); + } catch (InvocationTargetException e) { + return buildDynamicPropertyBeanValidationResult(validationResult, subject, input, e.getTargetException().getLocalizedMessage()); + } catch (IllegalArgumentException e) { + return buildDynamicPropertyBeanValidationResult(validationResult, subject, input, e.getLocalizedMessage()); + } + + return validationResult.valid(true).build(); + } + + private ValidationResult buildDynamicPropertyBeanValidationResult(final ValidationResult.Builder validationResult, + final String subject, final String input, final String message) { + return validationResult + .explanation( + String.format("Kinesis Client Library Configuration property with name with%s cannot be used with value \"%s\" : %s", + StringUtils.capitalize(subject), input, message) + ) + .valid(false).build(); + } + + @Override + protected Collection<ValidationResult> customValidate(final ValidationContext validationContext) { + final Set<ValidationResult> validationResults = new HashSet<>(super.customValidate(validationContext)); + + initialPositionInStream = InitialPositionInStream.valueOf(StringUtils.trimToEmpty(validationContext.getProperty(INITIAL_STREAM_POSITION).getValue())); + try { + dateTimeFormatter = DateTimeFormatter.ofPattern(validationContext.getProperty(TIMESTAMP_FORMAT).evaluateAttributeExpressions().getValue()); + } catch (IllegalArgumentException iae) { + validationResults.add(new ValidationResult.Builder().valid(false) + .subject(TIMESTAMP_FORMAT.getName()) + .explanation(String.format("%s must be a valid java.time.DateTimeFormatter format", TIMESTAMP_FORMAT.getDisplayName())) + .build() + ); + } + + if (InitialPositionInStream.AT_TIMESTAMP == initialPositionInStream) { + if (!validationContext.getProperty(STREAM_POSITION_TIMESTAMP).isSet()) { + validationResults.add(new ValidationResult.Builder().valid(false) + .subject(STREAM_POSITION_TIMESTAMP.getName()) + .explanation(String.format("%s must be provided when %s is %s", STREAM_POSITION_TIMESTAMP.getDisplayName(), + INITIAL_STREAM_POSITION.getDisplayName(), AT_TIMESTAMP.getDisplayName())) + .build() + ); + } else if (dateTimeFormatter != null) { + final String streamTimestamp = validationContext.getProperty(STREAM_POSITION_TIMESTAMP).getValue(); + try { + dateTimeFormatter.parse(streamTimestamp); + startStreamPositionTimestamp = new Date( + LocalDateTime.parse(validationContext.getProperty(STREAM_POSITION_TIMESTAMP).getValue(), dateTimeFormatter) + .toInstant(ZoneOffset.UTC).toEpochMilli() + ); + } catch (Exception e) { + validationResults.add(new ValidationResult.Builder().valid(false) + .subject(STREAM_POSITION_TIMESTAMP.getName()) + .explanation(String.format("%s must be parsable by %s", STREAM_POSITION_TIMESTAMP.getDisplayName(), Review comment: The oh-so-reliable [Wikipedia](https://en.wiktionary.org/wiki/parsable) suggests both versions of the word are correct and "without the e" generally prefered (in formal language at least) - IntelliJ complains at me if I add the `e` so I'll leave as-is unless you really object :slightly_smiling_face: -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org