Github user trixpan commented on a diff in the pull request:

    https://github.com/apache/nifi/pull/1692#discussion_r113917927
  
    --- Diff: 
nifi-nar-bundles/nifi-cybersecurity-bundle/nifi-cybersecurity-processors/src/main/java/org/apache/nifi/processors/cybersecurity/CompareFuzzyHash.java
 ---
    @@ -0,0 +1,289 @@
    +/*
    + * Licensed to the Apache Software Foundation (ASF) under one or more
    + * contributor license agreements.  See the NOTICE file distributed with
    + * this work for additional information regarding copyright ownership.
    + * The ASF licenses this file to You under the Apache License, Version 2.0
    + * (the "License"); you may not use this file except in compliance with
    + * the License.  You may obtain a copy of the License at
    + *
    + *     http://www.apache.org/licenses/LICENSE-2.0
    + *
    + * Unless required by applicable law or agreed to in writing, software
    + * distributed under the License is distributed on an "AS IS" BASIS,
    + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    + * See the License for the specific language governing permissions and
    + * limitations under the License.
    + */
    +
    +package org.apache.nifi.processors.cybersecurity;
    +
    +import org.apache.nifi.annotation.behavior.EventDriven;
    +import org.apache.nifi.annotation.behavior.InputRequirement;
    +import org.apache.nifi.annotation.behavior.SideEffectFree;
    +import org.apache.nifi.annotation.behavior.SupportsBatching;
    +import org.apache.nifi.annotation.behavior.WritesAttribute;
    +import org.apache.nifi.annotation.behavior.WritesAttributes;
    +import org.apache.nifi.annotation.documentation.CapabilityDescription;
    +import org.apache.nifi.annotation.documentation.SeeAlso;
    +import org.apache.nifi.annotation.documentation.Tags;
    +import org.apache.nifi.annotation.lifecycle.OnScheduled;
    +import org.apache.nifi.components.AllowableValue;
    +import org.apache.nifi.components.PropertyDescriptor;
    +import org.apache.nifi.flowfile.FlowFile;
    +import org.apache.nifi.logging.ComponentLog;
    +import org.apache.nifi.processor.ProcessContext;
    +import org.apache.nifi.processor.ProcessSession;
    +import org.apache.nifi.processor.ProcessorInitializationContext;
    +import org.apache.nifi.processor.Relationship;
    +import org.apache.nifi.processor.exception.ProcessException;
    +import org.apache.nifi.processor.util.StandardValidators;
    +import org.apache.nifi.processors.cybersecurity.matchers.FuzzyHashMatcher;
    +import org.apache.nifi.processors.cybersecurity.matchers.SSDeepHashMatcher;
    +import org.apache.nifi.processors.cybersecurity.matchers.TLSHHashMatcher;
    +
    +import java.io.BufferedReader;
    +import java.io.File;
    +import java.io.FileInputStream;
    +import java.io.FileNotFoundException;
    +import java.io.IOException;
    +import java.io.InputStreamReader;
    +import java.util.ArrayList;
    +import java.util.Collections;
    +import java.util.HashSet;
    +import java.util.List;
    +import java.util.Map;
    +import java.util.Set;
    +import java.util.concurrent.ConcurrentHashMap;
    +
    +
    +@EventDriven
    +@SideEffectFree
    +@SupportsBatching
    +@InputRequirement(InputRequirement.Requirement.INPUT_REQUIRED)
    +@SeeAlso({FuzzyHashContent.class})
    +@Tags({"hashing", "fuzzy-hashing", "cyber-security"})
    +@CapabilityDescription("Compares an attribute containing a Fuzzy Hash 
against a file containing a list of fuzzy hashes, " +
    +        "appending an attribute to the FlowFile in case of a successful 
match.")
    +
    +@WritesAttributes({
    +        @WritesAttribute(attribute = "XXXX.N.match", description = "The 
match that resembles the attribute specified " +
    +                "by the <Hash Attribute Name> property. Note that: 'XXX' 
gets replaced with the <Hash Attribute Name>"),
    +        @WritesAttribute(attribute = "XXXX.N.similarity", description = 
"The similarity score between this flowfile" +
    +                "and its match of the same number N. Note that: 'XXX' gets 
replaced with the <Hash Attribute Name>")})
    +
    +public class CompareFuzzyHash extends AbstractFuzzyHashProcessor {
    +    public static final AllowableValue singleMatch = new AllowableValue(
    +            "single",
    +            "single",
    +            "Send FlowFile to matched after the first match above 
threshold");
    +    public static final AllowableValue multiMatch = new AllowableValue(
    +            "multi-match",
    +            "multi-match",
    +            "Iterate full list of hashes before deciding to send FlowFile 
to matched or unmatched");
    +
    +    public static final PropertyDescriptor HASH_LIST_FILE = new 
PropertyDescriptor.Builder()
    +            .name("HASH_LIST_FILE")
    +            .displayName("Hash List source file")
    +            .description("Path to the file containing hashes to be 
validated against")
    +            .required(true)
    +            .addValidator(StandardValidators.FILE_EXISTS_VALIDATOR)
    +            .build();
    +
    +    // Note we add a PropertyDescriptor HASH_ALGORITHM and ATTRIBUTE_NAME 
from parent class
    +
    +    public static final PropertyDescriptor MATCH_THRESHOLD = new 
PropertyDescriptor.Builder()
    +            // Note that while both TLSH and SSDeep seems to return int, 
we treat them as double in code.
    +            // The rationale behind being the expectation that other 
algorithms thatmay return double values
    +            // may be added to the processor later on.
    +            .name("MATCH_THRESHOLD")
    +            .displayName("Match threshold")
    +            .description("The similarity score must exceed or be equal to 
in order for" +
    +                    "match to be considered true. Refer to Additional 
Information for differences between TLSH " +
    +                    "and SSDEEP scores and how they relate to this 
property.")
    +            .required(true)
    +            .addValidator(StandardValidators.NUMBER_VALIDATOR)
    +            .build();
    +
    +    public static final PropertyDescriptor MATCHING_MODE = new 
PropertyDescriptor.Builder()
    +            .name("MATCHING_MODE")
    +            .displayName("Matching mode")
    +            .description("Defines if the Processor should try to match as 
many entries as possible (" + multiMatch.getDisplayName() +
    +                    ") or if it should stio after the first match (" + 
singleMatch.getDisplayName() + ")")
    +            .required(true)
    +            .allowableValues(singleMatch,multiMatch)
    +            .defaultValue(singleMatch.getValue())
    +            .build();
    +
    +    public static final Relationship REL_FOUND = new Relationship.Builder()
    +            .name("found")
    +            .description("Any FlowFile that is successfully matched to an 
existing hash will be sent to this Relationship.")
    +            .build();
    +
    +    public static final Relationship REL_NOT_FOUND = new 
Relationship.Builder()
    +            .name("not found")
    +            .description("Any FlowFile that cannot be matched to an 
existing hash will be sent to this Relationship.")
    +            .build();
    +
    +    public static final Relationship REL_FAILURE = new 
Relationship.Builder()
    +            .name("failure")
    +            .description("Any FlowFile that cannot be matched, e.g. (lacks 
the attribute) will be sent to this Relationship.")
    +            .build();
    +
    +    @Override
    +    protected void init(final ProcessorInitializationContext context) {
    +        final List<PropertyDescriptor> descriptors = new 
ArrayList<PropertyDescriptor>();
    +        descriptors.add(HASH_LIST_FILE);
    +        // As mentioned above, add the PropertyDescriptor HASH_ALGORITHM 
and ATTRIBUTE_NAME from parent class
    +        descriptors.add(HASH_ALGORITHM);
    +        descriptors.add(ATTRIBUTE_NAME);
    +        descriptors.add(MATCH_THRESHOLD);
    +        descriptors.add(MATCHING_MODE);
    +        this.descriptors = Collections.unmodifiableList(descriptors);
    +
    +        final Set<Relationship> relationships = new 
HashSet<Relationship>();
    +        relationships.add(REL_FOUND);
    +        relationships.add(REL_NOT_FOUND);
    +        relationships.add(REL_FAILURE);
    +        this.relationships = Collections.unmodifiableSet(relationships);
    +    }
    +
    +    @Override
    +    public Set<Relationship> getRelationships() {
    +        return this.relationships;
    +    }
    +
    +    @Override
    +    public final List<PropertyDescriptor> 
getSupportedPropertyDescriptors() {
    +        return descriptors;
    +    }
    +
    +    @OnScheduled
    +    public void onScheduled(final ProcessContext context) {
    +    }
    +
    +    @Override
    +    public void onTrigger(ProcessContext context, ProcessSession session) 
throws ProcessException {
    +
    +        FlowFile flowFile = session.get();
    +        if (flowFile == null) {
    +            return;
    +        }
    +
    +        final ComponentLog logger = getLogger();
    +        String algorithm = context.getProperty(HASH_ALGORITHM).getValue();
    +
    +        String inputHash = 
flowFile.getAttribute(context.getProperty(ATTRIBUTE_NAME).getValue());
    +
    +        if (inputHash == null) {
    +            getLogger().info("FlowFile {} lacks the required '{}' 
attribute, routing to failure.",
    +                    new Object[]{flowFile, 
context.getProperty(ATTRIBUTE_NAME).getValue() });
    +            session.transfer(flowFile, REL_FAILURE);
    +            return;
    +        }
    +
    +        FuzzyHashMatcher fuzzyHashMatcher = null;
    +
    +        switch (algorithm) {
    +            case tlsh:
    +                fuzzyHashMatcher = new TLSHHashMatcher(getLogger());
    +                break;
    +            case ssdeep:
    +                fuzzyHashMatcher = new SSDeepHashMatcher(getLogger());
    +                break;
    +        }
    +
    +        if (fuzzyHashMatcher.isValidHash(inputHash) == false) {
    +            // and if that is the case we log
    +            logger.error("Invalid hash provided. Sending to failure");
    +            //  and send to failure
    +            session.transfer(flowFile, REL_FAILURE);
    +            session.commit();
    +            return;
    +        }
    +
    +        File file = new 
File(context.getProperty(HASH_LIST_FILE).getValue());
    +
    +        double similarity = 0;
    +        double matchThreshold = 
context.getProperty(MATCH_THRESHOLD).asDouble();
    +
    +        try {
    +            Map<String, Double> matched = new ConcurrentHashMap<String, 
Double>();
    +            FileInputStream fileInputStream = new FileInputStream(file);
    +            BufferedReader reader = new BufferedReader(new 
InputStreamReader(fileInputStream));
    +
    +            // If SSdeep skip the first line (as the usual format used by 
other tools add a header line
    +            // to a file list
    +            if (algorithm == ssdeep) {
    +                reader.readLine();
    +            }
    +
    +            String line = null;
    +            String[] hashToCompare = null;
    +
    +            iterateFile: while ((line = reader.readLine()) != null) {
    +                switch (context.getProperty(HASH_ALGORITHM).getValue()) {
    +                    case tlsh:
    +                        hashToCompare = line.split("\t", 2);
    +                        break;
    +                    case ssdeep:
    +                        hashToCompare = line.split(",", 2);
    +                        break;
    +                }
    +
    +                if (hashToCompare != null) {
    +                    similarity = fuzzyHashMatcher.getSimilarity(inputHash, 
hashToCompare[0]);
    +
    +                    if (fuzzyHashMatcher.matchExceedsThreshold(similarity, 
matchThreshold)) {
    +                        //
    +                        matched.put(hashToCompare[1], similarity);
    +                    }
    +                }
    +
    +                // Check if single match is desired and if a match has 
been made
    +                if (context.getProperty(MATCHING_MODE).getValue() == 
singleMatch.getValue() && (matched.size() > 0)) {
    +                    // and save time by breaking the outer loop
    +                    break iterateFile;
    +                }
    +            }
    +            // no matter if the break was called or not, Continue 
processing
    +            // First by creating a new map to hold attributes
    +            Map<String, String> attributes = new ConcurrentHashMap<String, 
String>();
    +
    +            // Then by iterating over the hashmap of matches
    +            if (matched.size() > 0) {
    +                int x = 0;
    +                for (Map.Entry<String, Double> entry : matched.entrySet()) 
{
    +                    // defining attributes accordingly
    +                    attributes.put(
    +                            context.getProperty(ATTRIBUTE_NAME).getValue() 
+ "." + x + ".match",
    --- End diff --
    
    wow... wasn't aware of the refactoring hotkey for this. Super thanks!


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

Reply via email to