[ 
https://issues.apache.org/jira/browse/NIFI-589?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14575359#comment-14575359
 ] 

ASF GitHub Bot commented on NIFI-589:
-------------------------------------

Github user rdblue commented on a diff in the pull request:

    https://github.com/apache/incubator-nifi/pull/51#discussion_r31857427
  
    --- Diff: 
nifi/nifi-nar-bundles/nifi-flume-bundle/nifi-flume-processors/src/main/java/org/apache/nifi/processors/flume/FlumeSinkProcessor.java
 ---
    @@ -0,0 +1,193 @@
    +/*
    + * Licensed to the Apache Software Foundation (ASF) under one or more
    + * contributor license agreements.  See the NOTICE file distributed with
    + * this work for additional information regarding copyright ownership.
    + * The ASF licenses this file to You under the Apache License, Version 2.0
    + * (the "License"); you may not use this file except in compliance with
    + * the License.  You may obtain a copy of the License at
    + *
    + *     http://www.apache.org/licenses/LICENSE-2.0
    + *
    + * Unless required by applicable law or agreed to in writing, software
    + * distributed under the License is distributed on an "AS IS" BASIS,
    + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    + * See the License for the specific language governing permissions and
    + * limitations under the License.
    + */
    +package org.apache.nifi.processors.flume;
    +
    +import com.google.common.base.Throwables;
    +import com.google.common.collect.ImmutableList;
    +import com.google.common.collect.ImmutableSet;
    +import com.google.common.collect.Lists;
    +import java.util.List;
    +import java.util.Set;
    +import org.apache.flume.Context;
    +import org.apache.flume.EventDeliveryException;
    +import org.apache.flume.Sink;
    +import org.apache.flume.Transaction;
    +import org.apache.flume.channel.MemoryChannel;
    +import org.apache.flume.conf.Configurables;
    +import org.apache.jasper.compiler.JspUtil;
    +import org.apache.nifi.annotation.documentation.CapabilityDescription;
    +import org.apache.nifi.annotation.documentation.Tags;
    +import org.apache.nifi.annotation.lifecycle.OnScheduled;
    +import org.apache.nifi.annotation.lifecycle.OnUnscheduled;
    +
    +import org.apache.nifi.components.PropertyDescriptor;
    +import org.apache.nifi.components.Validator;
    +import org.apache.nifi.flowfile.FlowFile;
    +import org.apache.nifi.processor.ProcessContext;
    +import org.apache.nifi.processor.ProcessSession;
    +import org.apache.nifi.processor.ProcessorInitializationContext;
    +import org.apache.nifi.processor.Relationship;
    +import org.apache.nifi.processor.SchedulingContext;
    +import org.apache.nifi.processor.exception.ProcessException;
    +import org.apache.nifi.processor.util.StandardValidators;
    +import org.apache.nifi.processors.flume.util.FlowFileEvent;
    +
    +/**
    + * This processor runs a Flume sink
    + */
    +@Tags({"flume", "hadoop", "get", "sink"})
    +@CapabilityDescription("Generate FlowFile data from a Flume sink")
    +public class FlumeSinkProcessor extends AbstractFlumeProcessor {
    +
    +    private Sink sink;
    +    private MemoryChannel channel;
    +
    +    public static final PropertyDescriptor SINK_TYPE = new 
PropertyDescriptor.Builder()
    +            .name("Sink Type")
    +            .description("The fully-qualified name of the Sink class")
    +            .required(true)
    +            .addValidator(createSinkValidator())
    +            .build();
    +    public static final PropertyDescriptor AGENT_NAME = new 
PropertyDescriptor.Builder()
    +            .name("Agent Name")
    +            .description("The name of the agent used in the Flume sink 
configuration")
    +            .required(true)
    +            .defaultValue("tier1")
    +            .addValidator(StandardValidators.NON_EMPTY_VALIDATOR)
    +            .build();
    +    public static final PropertyDescriptor SOURCE_NAME = new 
PropertyDescriptor.Builder()
    +            .name("Sink Name")
    +            .description("The name of the sink used in the Flume sink 
configuration")
    +            .required(true)
    +            .defaultValue("sink-1")
    +            .addValidator(StandardValidators.NON_EMPTY_VALIDATOR)
    +            .build();
    +    public static final PropertyDescriptor FLUME_CONFIG = new 
PropertyDescriptor.Builder()
    +            .name("Flume Configuration")
    +            .description("The Flume configuration for the sink copied from 
the flume.properties file")
    +            .required(true)
    +            .defaultValue("")
    +            .addValidator(Validator.VALID)
    +            .build();
    +    public static final PropertyDescriptor BATCH_SIZE = new 
PropertyDescriptor.Builder()
    +            .name("Batch Size")
    +            .description("The number of FlowFiles to process in a single 
batch")
    +            .required(true)
    +            .defaultValue("100")
    +            .addValidator(StandardValidators.INTEGER_VALIDATOR)
    +            .build();
    +
    +    public static final Relationship SUCCESS = new 
Relationship.Builder().name("success").build();
    +    public static final Relationship FAILURE = new 
Relationship.Builder().name("failure").build();
    +
    +    private List<PropertyDescriptor> descriptors;
    +    private Set<Relationship> relationships;
    +    private int batchSize;
    +
    +    @Override
    +    protected void init(final ProcessorInitializationContext context) {
    +        this.descriptors = ImmutableList.of(SINK_TYPE, AGENT_NAME, 
SOURCE_NAME, FLUME_CONFIG, BATCH_SIZE);
    +        this.relationships = ImmutableSet.of(SUCCESS, FAILURE);
    +    }
    +
    +    @Override
    +    protected List<PropertyDescriptor> getSupportedPropertyDescriptors() {
    +        return descriptors;
    +    }
    +
    +    @Override
    +    public Set<Relationship> getRelationships() {
    +        return relationships;
    +    }
    +
    +    @OnScheduled
    +    public void onScheduled(final SchedulingContext context) {
    +        batchSize = context.getProperty(BATCH_SIZE).asInteger();
    +
    +        try {
    +            channel = new MemoryChannel();
    +            Context memoryChannelContext = new Context();
    +            memoryChannelContext.put("capacity", 
String.valueOf(batchSize*10));
    +            memoryChannelContext.put("transactionCapacity", 
String.valueOf(batchSize*10));
    +            Configurables.configure(channel, memoryChannelContext);
    +            channel.start();
    +
    +            sink = 
SINK_FACTORY.create(context.getProperty(SOURCE_NAME).getValue(),
    +                    context.getProperty(SINK_TYPE).getValue());
    +            sink.setChannel(channel);
    +
    +            String flumeConfig = 
context.getProperty(FLUME_CONFIG).getValue();
    +            String agentName = context.getProperty(AGENT_NAME).getValue();
    +            String sinkName = context.getProperty(SOURCE_NAME).getValue();
    +            Configurables.configure(sink,
    +                    getFlumeSinkContext(flumeConfig, agentName, sinkName));
    +
    +            sink.start();
    +        } catch (Throwable th) {
    +            getLogger().error("Error creating sink", th);
    +            throw Throwables.propagate(th);
    +        }
    +    }
    +
    +    @OnUnscheduled
    +    public void unScheduled() {
    +        sink.stop();
    +        channel.stop();
    +    }
    +
    +    @Override
    +    public void onTrigger(final ProcessContext context,
    +            final ProcessSession session) throws ProcessException {
    +        List<FlowFile> flowFiles = 
Lists.newArrayListWithExpectedSize(batchSize);
    +        for (int i = 0; i < batchSize; i++) {
    +            FlowFile flowFile = session.get();
    +            if (flowFile == null) {
    +              break;
    +            }
    +
    +            flowFiles.add(flowFile);
    +        }
    +
    +        Transaction transaction = channel.getTransaction();
    --- End diff --
    
    Why not wrap the current session in a Flume transaction like the polling 
source does?


> Add processors that can run Apache Flume sources/sinks
> ------------------------------------------------------
>
>                 Key: NIFI-589
>                 URL: https://issues.apache.org/jira/browse/NIFI-589
>             Project: Apache NiFi
>          Issue Type: New Feature
>          Components: Extensions
>    Affects Versions: 0.1.0
>            Reporter: Joey Echeverria
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to