Hi Jeff,

Thanks for the reply. Your suggestion worked.

I've got one more question. The logs generated at the HDFS, using 
exec/spooling,  are in the following format :
-rw-r--r--   3 vikas         66 2013-05-01 05:43 /tmp/FlumeData.1367412193479
-rw-r--r--   3 vikas         67 2013-05-01 05:43 /tmp/FlumeData.1367412193480
-rw-r--r--   3 vikas         61 2013-05-01 05:43 /tmp/FlumeData.1367412193481
-rw-r--r--   3 vikas         44 2013-05-01 05:43 /tmp/FlumeData.1367412193482

Is there any way where I can copy the files (eg. com/org/test/flm/Sample.txt) 
from source to HDFS in the same folder structure/size/name?
If not, what is the alternative?

Thanks,
Vikas

________________________________
 From: Jeff Lord <[email protected]>
To: [email protected]; Vikas Kanth <[email protected]> 
Sent: Wednesday, 1 May 2013 7:05 AM
Subject: Re: Getting "Checking file:conf/flume.conf for changes" message in loop
 


Vikas,

This message is normal and harmless.

2013-04-29 08:26:11,868 (conf-file-poller-0) [DEBUG - 
org.apache.flume.conf.file.AbstractFileConfigurationProvider$FileWatcherRunnable.run(AbstractFileConfigurationProvider.java:188)]
 Checking file:conf/flume.conf for changes


If you change your log settings to INFO level it will also not show up.

Regarding the reason for which you do not see the contents of your file in hdfs.
One thing with the exec source and tail is that the events are buffered until 
20 events have been written to the cache. One way to work around this is to 
change the default from 20 -> 1

batchSize 20The max number of lines to read and send to the channel at a time


Alternatively there was a recent patch that sets the batchTimeout for exec 
source that will let you flush the cache based on elapsed time. That fix is 
available on the latest version of trunk.

-Jeff




On Mon, Apr 29, 2013 at 8:31 AM, Vikas Kanth <[email protected]> wrote:

Hi,
>
>
>I am getting following message in loop. The source file hasn't moved to the 
>destination.
>
>
>2013-04-29 08:24:41,346 (lifecycleSupervisor-1-0) [INFO - 
>org.apache.flume.instrumentation.MonitoredCounterGroup.start(MonitoredCounterGroup.java:73)]
> Component type: CHANNEL, name: Channel-2 started
>2013-04-29 08:24:41,846 (conf-file-poller-0) [INFO - 
>org.apache.flume.node.nodemanager.DefaultLogicalNodeManager.startAllComponents(DefaultLogicalNodeManager.java:141)]
> Starting Sink HDFS
>2013-04-29 08:24:41,847 (conf-file-poller-0) [INFO - 
>org.apache.flume.node.nodemanager.DefaultLogicalNodeManager.startAllComponents(DefaultLogicalNodeManager.java:152)]
> Starting Source tail
>2013-04-29 08:24:41,847 (lifecycleSupervisor-1-3) [INFO - 
>org.apache.flume.source.ExecSource.start(ExecSource.java:155)] Exec source 
>starting with command:tail -F /home/vkanth/temp/Sample2.txt
>2013-04-29 08:24:41,850 (lifecycleSupervisor-1-3) [DEBUG - 
>org.apache.flume.source.ExecSource.start(ExecSource.java:173)] Exec source 
>started
>2013-04-29 08:24:41,850 (lifecycleSupervisor-1-0) [INFO - 
>org.apache.flume.instrumentation.MonitoredCounterGroup.register(MonitoredCounterGroup.java:89)]
> Monitoried counter group for type: SINK, name: HDFS, registered successfully.
>2013-04-29 08:24:41,851 (lifecycleSupervisor-1-0) [INFO - 
>org.apache.flume.instrumentation.MonitoredCounterGroup.start(MonitoredCounterGroup.java:73)]
> Component type: SINK, name: HDFS started
>2013-04-29 08:24:41,852 (SinkRunner-PollingRunner-DefaultSinkProcessor) [DEBUG 
>- org.apache.flume.SinkRunner$PollingRunner.run(SinkRunner.java:143)] Polling 
>sink runner starting
>2013-04-29 08:25:11,855(conf-file-poller-0) [DEBUG - 
>org.apache.flume.conf.file.AbstractFileConfigurationProvider$FileWatcherRunnable.run(AbstractFileConfigurationProvider.java:188)]
> Checking file:conf/flume.conf for changes
>2013-04-29 08:25:41,861 (conf-file-poller-0) [DEBUG - 
>org.apache.flume.conf.file.AbstractFileConfigurationProvider$FileWatcherRunnable.run(AbstractFileConfigurationProvider.java:188)]
> Checking file:conf/flume.conf for changes
>2013-04-29 08:26:11,868(conf-file-poller-0) [DEBUG - 
>org.apache.flume.conf.file.AbstractFileConfigurationProvider$FileWatcherRunnable.run(AbstractFileConfigurationProvider.java:188)]
> Checking file:conf/flume.conf for changes
>.......
>.......
>
>
>
>
>Flume.conf:
>agent1.sources = tail
>agent1.channels = Channel-2
>agent1.sinks = HDFS
>
>
>agent1.sources.tail.type = exec
>agent1.sources.tail.command = tail -F /home/vikas/temp/Sample2.txt
>agent1.sources.tail.channels = Channel-2
>
>
>agent1.sinks.HDFS.channel = Channel-2
>agent1.sinks.HDFS.type = hdfs
>agent1.sinks.HDFS.hdfs.path = hdfs://dev-pub01.xyz.abc.com:8020/tmp
>agent1.sinks.HDFS.hdfs.file.fileType = DataStream
>
>
>agent1.channels.Channel-2.type = memory
>agent1.channels.Channel-2.capacity = 1000
>agent1.channels.Channel-2.transactionCapacity=10
>
>
>Command:
>bin/flume-ng agent --conf ./conf/ -f conf/flume.conf 
>-Dflume.root.logger=DEBUG,console -n agent1
>
>
>Please let me know if I am missing something.
>
>
>Thanks,
>Vikas
>
>

Reply via email to