[ 
https://issues.apache.org/jira/browse/FLUME-2816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ralph Goers resolved FLUME-2816.
--------------------------------
    Resolution: Won't Fix

Elasticsearch 2.x has been EOL for 4 years. Support for a more recent version 
of Elasticsearch is planned for Flume 2.0.

> Unable to get apache-flume-1.6.0 to work with ElasticSearch 2.0 beta 2 - 
> Works with Elasticsearch 1.7.3
> -------------------------------------------------------------------------------------------------------
>
>                 Key: FLUME-2816
>                 URL: https://issues.apache.org/jira/browse/FLUME-2816
>             Project: Flume
>          Issue Type: Improvement
>          Components: Sinks+Sources
>    Affects Versions: 1.6.0
>         Environment: Centos 6
>            Reporter: David Waugh
>            Priority: Major
>         Attachments: flume-1.6-elastic-2.0.patch
>
>
> Hello
> I am unable to get apache-flume-1.6.0 to work with ElasticSearch 2.0 beta 2.
> The same configuration works fine with Elasticsearch 1.7.3
> My flume-en.sh contains
> {noformat}
> Give Flume more memory and pre-allocate, enable remote monitoring via JMX
> export JAVA_OPTS="-Xms100m -Xmx2000m -Dcom.sun.management.jmxremote"
> Note that the Flume conf directory is always included in the classpath.
> FLUME_CLASSPATH="/usr/share/elasticsearch/lib/elasticsearch-2.0.0-beta2.jar"
> export JAVA_HOME=/usr/lib/jvm/java-1.7.0-openjdk-1.7.0.85.x86_64/jre
> {noformat}
> I have copied the lucene-core-5.2.1.jar from /usr/share/elasticsearch/lib to 
> /opt/flume/lib
> my elastic.conf file is:
> elastic.conf
> {noformat}
> #elastic.conf: A single-node Flume configuration
> #Reading avro files via spool dir and use the elastic sink
> #Name the components on this agent
> a1.sources = avrofld
> a1.sinks = k1
> a1.channels = c1
> #Describe/configure the folder with the avro files from Warehouse Connector
> a1.sources.avrofld.type = spooldir
> a1.sources.avrofld.channels = c1
> a1.sources.avrofld.spoolDir = /var/export
> a1.sources.avrofld.fileHeader = true
> a1.sources.avrofld.deserializer = avro
> Describe the elasticsearch sink
> a1.sinks.k1.type = org.apache.flume.sink.elasticsearch.ElasticSearchSink
> a1.sinks.k1.hostNames = 127.0.0.1:9300
> a1.sinks.k1.indexName = sa
> a1.sinks.k1.indexType = data
> a1.sinks.k1.clusterName = elasticsearch
> a1.sinks.k1.batchSize = 1000
> a1.sinks.k1.ttl = 2d
> a1.sinks.k1.serializer = 
> com.rsa.flume.serialization.FlumeAvroEventDeserializer
> #Use a channel which buffers events in memory
> a1.channels.c1.type = memory
> a1.channels.c1.capacity = 1000000
> a1.channels.c1.transactionCapacity = 100000
> #Bind the source and sink to the channel
> a1.sources.r1.channels = c1
> a1.sinks.k1.channel = c1
> {noformat}
> Elasticsearch 2.0 is successfully started and listening on the port
> {noformat}
> [root@CentosB lib]# netstat -na |grep 900
> tcp 0 0 127.0.0.1:9200 0.0.0.0: LISTEN
> tcp 0 0 127.0.0.1:9300 0.0.0.0:* LISTEN
> {noformat}
> However when starting my flume I get the following error
> {noformat}
> [root@CentosB flume]# bin/flume-ng agent --conf conf --conf-file 
> conf/elastic.conf --name a1 -Dflume.root.logger=DEBUG,console
> 2015-10-19 10:24:31,152 (lifecycleSupervisor-1-1) [ERROR - 
> org.apache.flume.lifecycle.LifecycleSupervisor$MonitorRunnable.run(LifecycleSupervisor.java:253)]
>  Unable to start SinkRunner: { 
> policy:org.apache.flume.sink.DefaultSinkProcessor@5c5dc0a5 counterGroup:{ 
> name:null counters:{} } } - Exception follows.
> java.lang.NoSuchMethodError: 
> org.elasticsearch.common.transport.InetSocketTransportAddress.(Ljava/lang/String;I)V
> at 
> org.apache.flume.sink.elasticsearch.client.ElasticSearchTransportClient.configureHostnames(ElasticSearchTransportClient.java:143)
> at 
> org.apache.flume.sink.elasticsearch.client.ElasticSearchTransportClient.(ElasticSearchTransportClient.java:77)
> at 
> org.apache.flume.sink.elasticsearch.client.ElasticSearchClientFactory.getClient(ElasticSearchClientFactory.java:48)
> at 
> org.apache.flume.sink.elasticsearch.ElasticSearchSink.start(ElasticSearchSink.java:357)
> at 
> org.apache.flume.sink.DefaultSinkProcessor.start(DefaultSinkProcessor.java:46)
> at org.apache.flume.SinkRunner.start(SinkRunner.java:79)
> at 
> org.apache.flume.lifecycle.LifecycleSupervisor$MonitorRunnable.run(LifecycleSupervisor.java:251)
> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
> at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:304)
> at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:178)
> at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
> at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> at java.lang.Thread.run(Thread.java:745)
> {noformat}
> Full Log Below
> {noformat}
> Info: Sourcing environment configuration script /opt/flume/conf/flume-env.sh
> Info: Including Hive libraries found via () for Hive access
> exec /usr/lib/jvm/java-1.7.0-openjdk-1.7.0.85.x86_64/jre/bin/java -Xms100m 
> -Xmx2000m -Dcom.sun.management.jmxremote -Dflume.root.logger=DEBUG,console 
> -cp 
> '/opt/flume/conf:/opt/flume/lib/:/usr/share/elasticsearch/lib/elasticsearch-2.0.0-beta2.jar:/lib/'
>  -Djava.library.path= org.apache.flume.node.Application --conf-file 
> conf/elastic.conf --name a1 2015-10-19 10:24:31,017 (lifecycleSupervisor-1-0) 
> [INFO - 
> org.apache.flume.node.PollingPropertiesFileConfigurationProvider.start(PollingPropertiesFileConfigurationProvider.java:61)]
>  Configuration provider starting 2015-10-19 10:24:31,020 
> (lifecycleSupervisor-1-0) [DEBUG - 
> org.apache.flume.node.PollingPropertiesFileConfigurationProvider.start(PollingPropertiesFileConfigurationProvider.java:78)]
>  Configuration provider started 2015-10-19 10:24:31,023 (conf-file-poller-0) 
> [DEBUG - 
> org.apache.flume.node.PollingPropertiesFileConfigurationProvider$FileWatcherRunnable.run(PollingPropertiesFileConfigurationProvider.java:126)]
>  Checking file:conf/elastic.conf for changes 2015-10-19 10:24:31,023 
> (conf-file-poller-0) [INFO - 
> org.apache.flume.node.PollingPropertiesFileConfigurationProvider$FileWatcherRunnable.run(PollingPropertiesFileConfigurationProvider.java:133)]
>  Reloading configuration file:conf/elastic.conf 2015-10-19 10:24:31,028 
> (conf-file-poller-0) [INFO - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:1017)]
>  Processing:k1 2015-10-19 10:24:31,028 (conf-file-poller-0) [DEBUG - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:1021)]
>  Created context for k1: serializer 2015-10-19 10:24:31,029 
> (conf-file-poller-0) [INFO - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:1017)]
>  Processing:k1 2015-10-19 10:24:31,029 (conf-file-poller-0) [INFO - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:931)]
>  Added sinks: k1 Agent: a1 2015-10-19 10:24:31,029 (conf-file-poller-0) [INFO 
> - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:1017)]
>  Processing:k1 2015-10-19 10:24:31,029 (conf-file-poller-0) [INFO - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:1017)]
>  Processing:k1 2015-10-19 10:24:31,029 (conf-file-poller-0) [INFO - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:1017)]
>  Processing:k1 2015-10-19 10:24:31,029 (conf-file-poller-0) [INFO - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:1017)]
>  Processing:k1 2015-10-19 10:24:31,030 (conf-file-poller-0) [INFO - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:1017)]
>  Processing:k1 2015-10-19 10:24:31,030 (conf-file-poller-0) [INFO - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:1017)]
>  Processing:k1 2015-10-19 10:24:31,030 (conf-file-poller-0) [INFO - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:1017)]
>  Processing:k1 2015-10-19 10:24:31,030 (conf-file-poller-0) [DEBUG - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.isValid(FlumeConfiguration.java:314)]
>  Starting validation of configuration for agent: a1, initial-configuration: 
> AgentConfiguration[a1] SOURCES: {avrofld={ parameters:{fileHeader=true, 
> channels=c1, spoolDir=/var/export, type=spooldir, deserializer=avro} }, r1={ 
> parameters:{channels=c1} }} CHANNELS: {c1={ 
> parameters:{transactionCapacity=100000, capacity=1000000, type=memory} }} 
> SINKS: {k1={ parameters:{clusterName=elasticsearch, indexType=data, 
> serializer=com.rsa.flume.serialization.FlumeAvroEventDeserializer, 
> indexName=sa, batchSize=1000, hostNames=127.0.0.1:9300, 
> type=org.apache.flume.sink.elasticsearch.ElasticSearchSink, ttl=2d, 
> channel=c1} }} 2015-10-19 10:24:31,039 (conf-file-poller-0) [DEBUG - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.validateChannels(FlumeConfiguration.java:469)]
>  Created channel c1 2015-10-19 10:24:31,046 (conf-file-poller-0) [DEBUG - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.validateSinks(FlumeConfiguration.java:675)]
>  Creating sink: k1 using ELASTICSEARCH 2015-10-19 10:24:31,047 
> (conf-file-poller-0) [DEBUG - 
> org.apache.flume.conf.FlumeConfiguration$AgentConfiguration.isValid(FlumeConfiguration.java:372)]
>  Post validation configuration for a1 AgentConfiguration created without 
> Configuration stubs for which only basic syntactical validation was 
> performed[a1] SOURCES: {avrofld={ parameters:{fileHeader=true, channels=c1, 
> spoolDir=/var/export, type=spooldir, deserializer=avro} }} CHANNELS: {c1={ 
> parameters:{transactionCapacity=100000, capacity=1000000, type=memory} }} 
> SINKS: {k1={ parameters:{clusterName=elasticsearch, indexType=data, 
> serializer=com.rsa.flume.serialization.FlumeAvroEventDeserializer, 
> indexName=sa, batchSize=1000, hostNames=127.0.0.1:9300, 
> type=org.apache.flume.sink.elasticsearch.ElasticSearchSink, ttl=2d, 
> channel=c1} }} 2015-10-19 10:24:31,047 (conf-file-poller-0) [DEBUG - 
> org.apache.flume.conf.FlumeConfiguration.validateConfiguration(FlumeConfiguration.java:136)]
>  Channels:c1 2015-10-19 10:24:31,048 (conf-file-poller-0) [DEBUG - 
> org.apache.flume.conf.FlumeConfiguration.validateConfiguration(FlumeConfiguration.java:137)]
>  Sinks k1 2015-10-19 10:24:31,048 (conf-file-poller-0) [DEBUG - 
> org.apache.flume.conf.FlumeConfiguration.validateConfiguration(FlumeConfiguration.java:138)]
>  Sources avrofld 2015-10-19 10:24:31,048 (conf-file-poller-0) [INFO - 
> org.apache.flume.conf.FlumeConfiguration.validateConfiguration(FlumeConfiguration.java:141)]
>  Post-validation flume configuration contains configuration for agents: [a1] 
> 2015-10-19 10:24:31,048 (conf-file-poller-0) [INFO - 
> org.apache.flume.node.AbstractConfigurationProvider.loadChannels(AbstractConfigurationProvider.java:145)]
>  Creating channels 2015-10-19 10:24:31,056 (conf-file-poller-0) [INFO - 
> org.apache.flume.channel.DefaultChannelFactory.create(DefaultChannelFactory.java:42)]
>  Creating instance of channel c1 type memory 2015-10-19 10:24:31,063 
> (conf-file-poller-0) [INFO - 
> org.apache.flume.node.AbstractConfigurationProvider.loadChannels(AbstractConfigurationProvider.java:200)]
>  Created channel c1 2015-10-19 10:24:31,063 (conf-file-poller-0) [INFO - 
> org.apache.flume.source.DefaultSourceFactory.create(DefaultSourceFactory.java:41)]
>  Creating instance of source avrofld, type spooldir 2015-10-19 10:24:31,086 
> (conf-file-poller-0) [INFO - 
> org.apache.flume.sink.DefaultSinkFactory.create(DefaultSinkFactory.java:42)] 
> Creating instance of sink: k1, type: 
> org.apache.flume.sink.elasticsearch.ElasticSearchSink 2015-10-19 10:24:31,086 
> (conf-file-poller-0) [DEBUG - 
> org.apache.flume.sink.DefaultSinkFactory.getClass(DefaultSinkFactory.java:63)]
>  Sink type org.apache.flume.sink.elasticsearch.ElasticSearchSink is a custom 
> type 2015-10-19 10:24:31,115 (conf-file-poller-0) [INFO - 
> org.apache.flume.node.AbstractConfigurationProvider.getConfiguration(AbstractConfigurationProvider.java:114)]
>  Channel c1 connected to [avrofld, k1] 2015-10-19 10:24:31,124 
> (conf-file-poller-0) [INFO - 
> org.apache.flume.node.Application.startAllComponents(Application.java:138)] 
> Starting new configuration:{ sourceRunners:{avrofld=EventDrivenSourceRunner: 
> { source:Spool Directory source avrofld: { spoolDir: /var/export } }} 
> sinkRunners:{k1=SinkRunner: { 
> policy:org.apache.flume.sink.DefaultSinkProcessor@5c5dc0a5 counterGroup:{ 
> name:null counters:{} } }} 
> channels:{c1=org.apache.flume.channel.MemoryChannel{name: c1}} } 2015-10-19 
> 10:24:31,134 (conf-file-poller-0) [INFO - 
> org.apache.flume.node.Application.startAllComponents(Application.java:145)] 
> Starting Channel c1 2015-10-19 10:24:31,136 (lifecycleSupervisor-1-0) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.register(MonitoredCounterGroup.java:120)]
>  Monitored counter group for type: CHANNEL, name: c1: Successfully registered 
> new MBean. 2015-10-19 10:24:31,136 (lifecycleSupervisor-1-0) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.start(MonitoredCounterGroup.java:96)]
>  Component type: CHANNEL, name: c1 started 2015-10-19 10:24:31,137 
> (conf-file-poller-0) [INFO - 
> org.apache.flume.node.Application.startAllComponents(Application.java:173)] 
> Starting Sink k1 2015-10-19 10:24:31,138 (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.sink.elasticsearch.ElasticSearchSink.start(ElasticSearchSink.java:350)]
>  ElasticSearch sink {} started 2015-10-19 10:24:31,139 
> (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.register(MonitoredCounterGroup.java:120)]
>  Monitored counter group for type: SINK, name: k1: Successfully registered 
> new MBean. 2015-10-19 10:24:31,139 (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.start(MonitoredCounterGroup.java:96)]
>  Component type: SINK, name: k1 started 2015-10-19 10:24:31,142 
> (conf-file-poller-0) [INFO - 
> org.apache.flume.node.Application.startAllComponents(Application.java:184)] 
> Starting Source avrofld 2015-10-19 10:24:31,142 (lifecycleSupervisor-1-2) 
> [INFO - 
> org.apache.flume.source.SpoolDirectorySource.start(SpoolDirectorySource.java:78)]
>  SpoolDirectorySource source starting with directory: /var/export 2015-10-19 
> 10:24:31,147 (lifecycleSupervisor-1-1) [WARN - 
> org.apache.flume.sink.elasticsearch.client.ElasticSearchTransportClient.configureHostnames(ElasticSearchTransportClient.java:136)]
>  [127.0.0.1:9300] 2015-10-19 10:24:31,152 (lifecycleSupervisor-1-1) [ERROR - 
> org.apache.flume.lifecycle.LifecycleSupervisor$MonitorRunnable.run(LifecycleSupervisor.java:253)]
>  Unable to start SinkRunner: { 
> policy:org.apache.flume.sink.DefaultSinkProcessor@5c5dc0a5 counterGroup:{ 
> name:null counters:{} } } - Exception follows. java.lang.NoSuchMethodError: 
> org.elasticsearch.common.transport.InetSocketTransportAddress.(Ljava/lang/String;I)V
>  at 
> org.apache.flume.sink.elasticsearch.client.ElasticSearchTransportClient.configureHostnames(ElasticSearchTransportClient.java:143)
>  at 
> org.apache.flume.sink.elasticsearch.client.ElasticSearchTransportClient.(ElasticSearchTransportClient.java:77)
>  at 
> org.apache.flume.sink.elasticsearch.client.ElasticSearchClientFactory.getClient(ElasticSearchClientFactory.java:48)
>  at 
> org.apache.flume.sink.elasticsearch.ElasticSearchSink.start(ElasticSearchSink.java:357)
>  at 
> org.apache.flume.sink.DefaultSinkProcessor.start(DefaultSinkProcessor.java:46)
>  at org.apache.flume.SinkRunner.start(SinkRunner.java:79) at 
> org.apache.flume.lifecycle.LifecycleSupervisor$MonitorRunnable.run(LifecycleSupervisor.java:251)
>  at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) 
> at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:304) at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:178)
>  at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>  at java.lang.Thread.run(Thread.java:745) 2015-10-19 10:24:31,160 
> (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.sink.elasticsearch.ElasticSearchSink.stop(ElasticSearchSink.java:376)]
>  ElasticSearch sink {} stopping 2015-10-19 10:24:31,160 
> (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:150)]
>  Component type: SINK, name: k1 stopped 2015-10-19 10:24:31,160 
> (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:156)]
>  Shutdown Metric for type: SINK, name: k1. sink.start.time == 1445246671139 
> 2015-10-19 10:24:31,160 (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:162)]
>  Shutdown Metric for type: SINK, name: k1. sink.stop.time == 1445246671160 
> 2015-10-19 10:24:31,160 (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:178)]
>  Shutdown Metric for type: SINK, name: k1. sink.batch.complete == 0 
> 2015-10-19 10:24:31,161 (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:178)]
>  Shutdown Metric for type: SINK, name: k1. sink.batch.empty == 0 2015-10-19 
> 10:24:31,161 (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:178)]
>  Shutdown Metric for type: SINK, name: k1. sink.batch.underflow == 0 
> 2015-10-19 10:24:31,161 (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:178)]
>  Shutdown Metric for type: SINK, name: k1. sink.connection.closed.count == 1 
> 2015-10-19 10:24:31,161 (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:178)]
>  Shutdown Metric for type: SINK, name: k1. sink.connection.creation.count == 
> 0 2015-10-19 10:24:31,161 (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:178)]
>  Shutdown Metric for type: SINK, name: k1. sink.connection.failed.count == 0 
> 2015-10-19 10:24:31,161 (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:178)]
>  Shutdown Metric for type: SINK, name: k1. sink.event.drain.attempt == 0 
> 2015-10-19 10:24:31,161 (lifecycleSupervisor-1-1) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.stop(MonitoredCounterGroup.java:178)]
>  Shutdown Metric for type: SINK, name: k1. sink.event.drain.sucess == 0 
> 2015-10-19 10:24:31,161 (lifecycleSupervisor-1-1) [WARN - 
> org.apache.flume.lifecycle.LifecycleSupervisor$MonitorRunnable.run(LifecycleSupervisor.java:260)]
>  Component SinkRunner: { 
> policy:org.apache.flume.sink.DefaultSinkProcessor@5c5dc0a5 counterGroup:{ 
> name:null counters:{} } } stopped, since it could not besuccessfully started 
> due to missing dependencies 2015-10-19 10:24:31,155 (lifecycleSupervisor-1-2) 
> [DEBUG - 
> org.apache.flume.client.avro.ReliableSpoolingFileEventReader.(ReliableSpoolingFileEventReader.java:138)]
>  Initializing ReliableSpoolingFileEventReader with directory=/var/export, 
> metaDir=.flumespool, deserializer=avro 2015-10-19 10:24:31,175 
> (lifecycleSupervisor-1-2) [DEBUG - 
> org.apache.flume.client.avro.ReliableSpoolingFileEventReader.(ReliableSpoolingFileEventReader.java:160)]
>  Successfully created and deleted canary file: 
> /var/export/flume-spooldir-perm-check-8040653594475164987.canary 2015-10-19 
> 10:24:31,176 (lifecycleSupervisor-1-2) [DEBUG - 
> org.apache.flume.source.SpoolDirectorySource.start(SpoolDirectorySource.java:111)]
>  SpoolDirectorySource source started 2015-10-19 10:24:31,177 
> (lifecycleSupervisor-1-2) [INFO - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.register(MonitoredCounterGroup.java:120)]
>  Monitored counter group for type: SOURCE, name: avrofld: Successfully 
> registered new MBean. 2015-10-19 10:24:31,177 (lifecycleSupervisor-1-2) [INFO 
> - 
> org.apache.flume.instrumentation.MonitoredCounterGroup.start(MonitoredCounterGroup.java:96)]
>  Component type: SOURCE, name: avrofld started
> {noformat}
> I opened a gihub issue
> https://github.com/elastic/elasticsearch/issues/14187
> Looks like it is due to this change 
> https://www.elastic.co/guide/en/elasticsearch/reference/2.0/_java_api_changes.html#_inetsockettransportaddress_removed
> And this looks to be because the ElasticSerach API has changed.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@flume.apache.org
For additional commands, e-mail: issues-h...@flume.apache.org

Reply via email to