Hi Thanks, Still the problem persist.
My flume-env.sh file as after changes: export JAVA_HOME= /usr/java/jdk1.7.0_01 export HADOOP_HOME=/home/mukhtaj/hadoop #JAVA_OPTS="-Xms100m -Xmx200m -Dcom.sun.management.jmxremote" export FLUME_CLASSPATH=/home/mukhtaj/apache-flume/lib Please suggest me any solution. Thanks and regards On Wed, Aug 1, 2012 at 3:15 PM, alo alt <[email protected]> wrote: > Or add the directory which include the libs into the classpath like the > flume-libs. > > cheers > Alex > > > On Aug 1, 2012, at 4:13 PM, Brock Noland <[email protected]> wrote: > > > export PATH=$PATH:/path/to/hadoop > > > > This could be done in the flume-env.sh file. > > > > Brock > > > > On Wed, Aug 1, 2012 at 9:12 AM, mardan Khan <[email protected]> > wrote: > > > >> Thanks Brock > >> > >> How to add the hadoop command to path. Can you please give me an > example. > >> > >> Thanks > >> > >> On Wed, Aug 1, 2012 at 3:07 PM, Brock Noland <[email protected]> > wrote: > >> > >>> Hi, > >>> > >>> I think this is because you need more than just the hadoop-core.jar > file. > >>> If you add the "hadoop" command to your path, the flume-ng script > should > >>> pickup these dependencies automatically. > >>> > >>> Brock > >>> > >>> > >>> On Tue, Jul 31, 2012 at 12:25 PM, mardan Khan <[email protected] > >wrote: > >>> > >>>> HI, > >>>> > >>>> I am posting again as I am still struggling for the solution. I have > >>>> simple configuration file which upload the data into hadoop but give > me > >>>> error message: Agent Failed because dependencies were not found. I am > using > >>>> the following: > >>>> > >>>> 1). Flume1.2.0 > >>>> 2) Hadoop-1.0.3 > >>>> 3). Window 7 > >>>> 4). Cygwin. > >>>> > >>>> > >>>> Configuration File: > >>>> > >>>> agent1.sources = source1 > >>>> agent1.sinks = sink1 > >>>> agent1.channels = channel1 > >>>> > >>>> agent1.sources.source1.type = netcat > >>>> agent1.sources.source1.bind = localhost > >>>> agent1.sources.source1.port = 23 > >>>> > >>>> agent1.sinks.sink1.type = logger > >>>> agent1.sinks.sink1.type = hdfs > >>>> agent1.sinks.sink1.hdfs.path = > >>>> hdfs://localhost:9000/user/cyg_server/flume > >>>> agent1.channels.channel1.type = memory > >>>> agent1.channels.channel1.capacity = 1000 > >>>> agent1.channels.channel1.transactionCapactiy = 100 > >>>> > >>>> agent1.sources.source1.channels = channel1 > >>>> agent1.sinks.sink1.channel = channel1 > >>>> > >>>> *ERROR MESSAGE* > >>>> > >>>> mukhtaj@mukhtaj-PC ~/apache-flume > >>>> $ bin/flume-ng agent -n agent1 -c conf -f > >>>> conf/flume-conf.properties.template > >>>> cygpath: can't convert empty path > >>>> + /cygdrive/c/java/jdk1.7.0_01/bin/java -Xmx20m -cp > >>>> 'C:\cygwin\home\mukhtaj\apac > >>>> he-flume\conf;C:\cygwin\home\mukhtaj\apache-flume\lib\*' > >>>> -Djava.library.path= or > >>>> g.apache.flume.node.Application -n agent1 -f > >>>> conf/flume-conf.properties.template > >>>> > >>>> 2012-07-31 18:17:56,120 (main) [INFO - > >>>> org.apache.flume.lifecycle.LifecycleSuper > >>>> visor.start(LifecycleSupervisor.java:67)] Starting lifecycle > supervisor 1 > >>>> 2012-07-31 18:17:56,124 (main) [INFO - > >>>> org.apache.flume.node.FlumeNode.start(Flu > >>>> meNode.java:54)] Flume node starting - agent1 > >>>> 2012-07-31 18:17:56,128 (lifecycleSupervisor-1-0) [INFO - > >>>> org.apache.flume.node. > >>>> > >>>> > nodemanager.DefaultLogicalNodeManager.start(DefaultLogicalNodeManager.java:187)] > >>>> Node manager starting > >>>> 2012-07-31 18:17:56,128 (lifecycleSupervisor-1-1) [INFO - > >>>> org.apache.flume.conf. > >>>> > >>>> > file.AbstractFileConfigurationProvider.start(AbstractFileConfigurationProvider.j > >>>> ava:67)] Configuration provider starting > >>>> 2012-07-31 18:17:56,130 (lifecycleSupervisor-1-0) [INFO - > >>>> org.apache.flume.lifec > >>>> ycle.LifecycleSupervisor.start(LifecycleSupervisor.java:67)] Starting > >>>> lifecycle > >>>> supervisor 9 > >>>> 2012-07-31 18:17:56,131 (lifecycleSupervisor-1-0) [DEBUG - > >>>> org.apache.flume.node > >>>> > >>>> > .nodemanager.DefaultLogicalNodeManager.start(DefaultLogicalNodeManager.java:191) > >>>> ] Node manager started > >>>> 2012-07-31 18:17:56,132 (lifecycleSupervisor-1-1) [DEBUG - > >>>> org.apache.flume.conf > >>>> > >>>> > .file.AbstractFileConfigurationProvider.start(AbstractFileConfigurationProvider. > >>>> java:86)] Configuration provider started > >>>> 2012-07-31 18:17:56,132 (conf-file-poller-0) [DEBUG - > >>>> org.apache.flume.conf.file > >>>> > >>>> > .AbstractFileConfigurationProvider$FileWatcherRunnable.run(AbstractFileConfigura > >>>> tionProvider.java:188)] Checking > >>>> file:conf\flume-conf.properties.template for ch > >>>> anges > >>>> 2012-07-31 18:17:56,134 (conf-file-poller-0) [INFO - > >>>> org.apache.flume.conf.file. > >>>> > >>>> > AbstractFileConfigurationProvider$FileWatcherRunnable.run(AbstractFileConfigurat > >>>> ionProvider.java:195)] Reloading configuration > >>>> file:conf\flume-conf.properties.t > >>>> emplate > >>>> 2012-07-31 18:17:56,140 (conf-file-poller-0) [INFO - > >>>> org.apache.flume.conf.Flume > >>>> > Configuration$AgentConfiguration.addProperty(FlumeConfiguration.java:988)] > >>>> Proce > >>>> ssing:sink1 > >>>> 2012-07-31 18:17:56,141 (conf-file-poller-0) [DEBUG - > >>>> org.apache.flume.conf.Flum > >>>> > eConfiguration$AgentConfiguration.addProperty(FlumeConfiguration.java:992)] > >>>> Crea > >>>> ted context for sink1: hdfs.path > >>>> 2012-07-31 18:17:56,142 (conf-file-poller-0) [INFO - > >>>> org.apache.flume.conf.Flume > >>>> > Configuration$AgentConfiguration.addProperty(FlumeConfiguration.java:988)] > >>>> Proce > >>>> ssing:sink1 > >>>> 2012-07-31 18:17:56,143 (conf-file-poller-0) [INFO - > >>>> org.apache.flume.conf.Flume > >>>> > Configuration$AgentConfiguration.addProperty(FlumeConfiguration.java:988)] > >>>> Proce > >>>> ssing:sink1 > >>>> 2012-07-31 18:17:56,145 (conf-file-poller-0) [INFO - > >>>> org.apache.flume.conf.Flume > >>>> > Configuration$AgentConfiguration.addProperty(FlumeConfiguration.java:902)] > >>>> Added > >>>> sinks: sink1 Agent: agent1 > >>>> 2012-07-31 18:17:56,146 (conf-file-poller-0) [DEBUG - > >>>> org.apache.flume.conf.Flum > >>>> > eConfiguration$AgentConfiguration.isValid(FlumeConfiguration.java:295)] > >>>> Starting > >>>> validation of configuration for agent: agent1, initial-configuration: > >>>> AgentConf > >>>> iguration[agent1] > >>>> SOURCES: {source1={ parameters:{port=23, channels=channel1, > type=netcat, > >>>> bind=lo > >>>> calhost} }} > >>>> CHANNELS: {channel1={ parameters:{transactionCapactiy=100, > >>>> capacity=1000, type=m > >>>> emory} }} > >>>> SINKS: {sink1={ > >>>> parameters:{hdfs.path=hdfs://localhost:9000/user/cyg_server/flum > >>>> e, type=hdfs, channel=channel1} }} > >>>> > >>>> 2012-07-31 18:17:56,156 (conf-file-poller-0) [DEBUG - > >>>> org.apache.flume.conf.Flum > >>>> > >>>> > eConfiguration$AgentConfiguration.validateChannels(FlumeConfiguration.java:450)] > >>>> Created channel channel1 > >>>> 2012-07-31 18:17:56,174 (conf-file-poller-0) [DEBUG - > >>>> org.apache.flume.conf.Flum > >>>> > eConfiguration$AgentConfiguration.validateSinks(FlumeConfiguration.java:649)] > >>>> Cr > >>>> eating sink: sink1 using HDFS > >>>> 2012-07-31 18:17:56,176 (conf-file-poller-0) [DEBUG - > >>>> org.apache.flume.conf.Flum > >>>> > eConfiguration$AgentConfiguration.isValid(FlumeConfiguration.java:353)] > >>>> Post val > >>>> idation configuration for agent1 > >>>> AgentConfiguration created without Configuration stubs for which only > >>>> basic synt > >>>> actical validation was performed[agent1] > >>>> SOURCES: {source1={ parameters:{port=23, channels=channel1, > type=netcat, > >>>> bind=lo > >>>> calhost} }} > >>>> CHANNELS: {channel1={ parameters:{transactionCapactiy=100, > >>>> capacity=1000, type=m > >>>> emory} }} > >>>> SINKS: {sink1={ > >>>> parameters:{hdfs.path=hdfs://localhost:9000/user/cyg_server/flum > >>>> e, type=hdfs, channel=channel1} }} > >>>> > >>>> 2012-07-31 18:17:56,177 (conf-file-poller-0) [DEBUG - > >>>> org.apache.flume.conf.Flum > >>>> eConfiguration.validateConfiguration(FlumeConfiguration.java:117)] > >>>> Channels:chan > >>>> nel1 > >>>> > >>>> 2012-07-31 18:17:56,177 (conf-file-poller-0) [DEBUG - > >>>> org.apache.flume.conf.Flum > >>>> eConfiguration.validateConfiguration(FlumeConfiguration.java:118)] > Sinks > >>>> sink1 > >>>> > >>>> 2012-07-31 18:17:56,178 (conf-file-poller-0) [DEBUG - > >>>> org.apache.flume.conf.Flum > >>>> eConfiguration.validateConfiguration(FlumeConfiguration.java:119)] > >>>> Sources sourc > >>>> e1 > >>>> > >>>> 2012-07-31 18:17:56,178 (conf-file-poller-0) [INFO - > >>>> org.apache.flume.conf.Flume > >>>> Configuration.validateConfiguration(FlumeConfiguration.java:122)] > >>>> Post-validatio > >>>> n flume configuration contains configuration for agents: [agent1] > >>>> 2012-07-31 18:17:56,178 (conf-file-poller-0) [INFO - > >>>> org.apache.flume.conf.prope > >>>> > >>>> > rties.PropertiesFileConfigurationProvider.loadChannels(PropertiesFileConfigurati > >>>> onProvider.java:249)] Creating channels > >>>> 2012-07-31 18:17:56,179 (conf-file-poller-0) [DEBUG - > >>>> org.apache.flume.channel.D > >>>> efaultChannelFactory.create(DefaultChannelFactory.java:68)] Creating > >>>> instance of > >>>> channel channel1 type memory > >>>> 2012-07-31 18:17:56,238 (conf-file-poller-0) [INFO - > >>>> org.apache.flume.instrument > >>>> ation.MonitoredCounterGroup.<init>(MonitoredCounterGroup.java:68)] > >>>> Monitoried co > >>>> unter group for type: CHANNEL, name: channel1, registered > successfully. > >>>> 2012-07-31 18:17:56,239 (conf-file-poller-0) [INFO - > >>>> org.apache.flume.conf.prope > >>>> > >>>> > rties.PropertiesFileConfigurationProvider.loadChannels(PropertiesFileConfigurati > >>>> onProvider.java:273)] created channel channel1 > >>>> 2012-07-31 18:17:56,239 (conf-file-poller-0) [DEBUG - > >>>> org.apache.flume.source.De > >>>> faultSourceFactory.create(DefaultSourceFactory.java:74)] Creating > >>>> instance of so > >>>> urce source1, type netcat > >>>> 2012-07-31 18:17:56,316 (conf-file-poller-0) [INFO - > >>>> org.apache.flume.sink.Defau > >>>> ltSinkFactory.create(DefaultSinkFactory.java:70)] Creating instance of > >>>> sink: sin > >>>> k1, type: hdfs > >>>> 2012-07-31 18:17:56,392 (conf-file-poller-0) [DEBUG - > >>>> org.apache.hadoop.conf.Con > >>>> figuration.<init>(Configuration.java:227)] java.io.IOException: > config() > >>>> at > >>>> org.apache.hadoop.conf.Configuration.<init>(Configuration.java:227) > >>>> at > >>>> org.apache.hadoop.conf.Configuration.<init>(Configuration.java:214) > >>>> at > >>>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(Use > >>>> rGroupInformation.java:184) > >>>> at > >>>> org.apache.hadoop.security.UserGroupInformation.isSecurityEnabled(Use > >>>> rGroupInformation.java:236) > >>>> at > >>>> org.apache.flume.sink.hdfs.HDFSEventSink.authenticate(HDFSEventSink.j > >>>> ava:516) > >>>> at > >>>> org.apache.flume.sink.hdfs.HDFSEventSink.configure(HDFSEventSink.java > >>>> :238) > >>>> at > >>>> org.apache.flume.conf.Configurables.configure(Configurables.java:41) > >>>> at > >>>> org.apache.flume.conf.properties.PropertiesFileConfigurationProvider. > >>>> loadSinks(PropertiesFileConfigurationProvider.java:373) > >>>> at > >>>> org.apache.flume.conf.properties.PropertiesFileConfigurationProvider. > >>>> load(PropertiesFileConfigurationProvider.java:223) > >>>> at > >>>> org.apache.flume.conf.file.AbstractFileConfigurationProvider.doLoad(A > >>>> bstractFileConfigurationProvider.java:123) > >>>> at > >>>> org.apache.flume.conf.file.AbstractFileConfigurationProvider.access$3 > >>>> 00(AbstractFileConfigurationProvider.java:38) > >>>> at > >>>> org.apache.flume.conf.file.AbstractFileConfigurationProvider$FileWatc > >>>> herRunnable.run(AbstractFileConfigurationProvider.java:202) > >>>> at > >>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:47 > >>>> 1) > >>>> at > >>>> java.util.concurrent.FutureTask$Sync.innerRunAndReset(FutureTask.java > >>>> :351) > >>>> at > >>>> java.util.concurrent.FutureTask.runAndReset(FutureTask.java:178) > >>>> at > >>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask. > >>>> access$301(ScheduledThreadPoolExecutor.java:178) > >>>> at > >>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask. > >>>> run(ScheduledThreadPoolExecutor.java:293) > >>>> at > >>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor. > >>>> java:1110) > >>>> at > >>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor > >>>> .java:603) > >>>> at java.lang.Thread.run(Thread.java:722) > >>>> > >>>> 2012-07-31 18:17:56,464 (conf-file-poller-0) [DEBUG - > >>>> org.apache.hadoop.security > >>>> .Groups.getUserToGroupsMappingService(Groups.java:139)] Creating new > >>>> Groups obj > >>>> ect > >>>> 2012-07-31 18:17:56,508 (conf-file-poller-0) [DEBUG - > >>>> org.apache.hadoop.security > >>>> .Groups.<init>(Groups.java:59)] Group mapping impl= > >>>> org.apache.hadoop.security.Sh > >>>> ellBasedUnixGroupsMapping; cacheTimeout=300000 > >>>> 2012-07-31 18:17:56,551 (conf-file-poller-0) [DEBUG - > >>>> org.apache.hadoop.conf.Con > >>>> figuration.<init>(Configuration.java:227)] java.io.IOException: > config() > >>>> at > >>>> org.apache.hadoop.conf.Configuration.<init>(Configuration.java:227) > >>>> at > >>>> org.apache.hadoop.conf.Configuration.<init>(Configuration.java:214) > >>>> at > >>>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(Use > >>>> rGroupInformation.java:184) > >>>> at > >>>> org.apache.hadoop.security.UserGroupInformation.isSecurityEnabled(Use > >>>> rGroupInformation.java:236) > >>>> at > >>>> org.apache.hadoop.security.KerberosName.<clinit>(KerberosName.java:79 > >>>> ) > >>>> at > >>>> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupI > >>>> nformation.java:209) > >>>> at > >>>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(Use > >>>> rGroupInformation.java:184) > >>>> at > >>>> org.apache.hadoop.security.UserGroupInformation.isSecurityEnabled(Use > >>>> rGroupInformation.java:236) > >>>> at > >>>> org.apache.flume.sink.hdfs.HDFSEventSink.authenticate(HDFSEventSink.j > >>>> ava:516) > >>>> at > >>>> org.apache.flume.sink.hdfs.HDFSEventSink.configure(HDFSEventSink.java > >>>> :238) > >>>> at > >>>> org.apache.flume.conf.Configurables.configure(Configurables.java:41) > >>>> at > >>>> org.apache.flume.conf.properties.PropertiesFileConfigurationProvider. > >>>> loadSinks(PropertiesFileConfigurationProvider.java:373) > >>>> at > >>>> org.apache.flume.conf.properties.PropertiesFileConfigurationProvider. > >>>> load(PropertiesFileConfigurationProvider.java:223) > >>>> at > >>>> org.apache.flume.conf.file.AbstractFileConfigurationProvider.doLoad(A > >>>> bstractFileConfigurationProvider.java:123) > >>>> at > >>>> org.apache.flume.conf.file.AbstractFileConfigurationProvider.access$3 > >>>> 00(AbstractFileConfigurationProvider.java:38) > >>>> at > >>>> org.apache.flume.conf.file.AbstractFileConfigurationProvider$FileWatc > >>>> herRunnable.run(AbstractFileConfigurationProvider.java:202) > >>>> at > >>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:47 > >>>> 1) > >>>> at > >>>> java.util.concurrent.FutureTask$Sync.innerRunAndReset(FutureTask.java > >>>> :351) > >>>> at > >>>> java.util.concurrent.FutureTask.runAndReset(FutureTask.java:178) > >>>> at > >>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask. > >>>> access$301(ScheduledThreadPoolExecutor.java:178) > >>>> at > >>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask. > >>>> run(ScheduledThreadPoolExecutor.java:293) > >>>> at > >>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor. > >>>> java:1110) > >>>> at > >>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor > >>>> .java:603) > >>>> at java.lang.Thread.run(Thread.java:722) > >>>> > >>>> 2012-07-31 18:17:56,592 (conf-file-poller-0) > >>>> Runnable.run(AbstractFileConfigura > >>>> tionProvider.java:207)] Failed to start agent because [ERROR - > >>>> org.apache.flume.conf.file > >>>> .AbstractFileConfigurationProvider$FileWatchuse dependencies were not > >>>> foun > >>>> d in classpath. Error follows. > >>>> java.lang.NoClassDefFoundError: > >>>> org/apache/commons/configuration/Configuration > >>>> at > >>>> org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.<init>(DefaultMet > >>>> ricsSystem.java:37) > >>>> at > >>>> org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.<clinit>(DefaultM > >>>> etricsSystem.java:34) > >>>> at > >>>> org.apache.hadoop.security.UgiInstrumentation.create(UgiInstrumentati > >>>> on.java:51) > >>>> at > >>>> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupI > >>>> nformation.java:216) > >>>> at > >>>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(Use > >>>> rGroupInformation.java:184) > >>>> at > >>>> org.apache.hadoop.security.UserGroupInformation.isSecurityEnabled(Use > >>>> rGroupInformation.java:236) > >>>> at > >>>> org.apache.hadoop.security.KerberosName.<clinit>(KerberosName.java:79 > >>>> ) > >>>> at > >>>> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupI > >>>> nformation.java:209) > >>>> at > >>>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(Use > >>>> rGroupInformation.java:184) > >>>> at > >>>> org.apache.hadoop.security.UserGroupInformation.isSecurityEnabled(Use > >>>> rGroupInformation.java:236) > >>>> at > >>>> org.apache.flume.sink.hdfs.HDFSEventSink.authenticate(HDFSEventSink.j > >>>> ava:516) > >>>> at > >>>> org.apache.flume.sink.hdfs.HDFSEventSink.configure(HDFSEventSink.java > >>>> :238) > >>>> at > >>>> org.apache.flume.conf.Configurables.configure(Configurables.java:41) > >>>> at > >>>> org.apache.flume.conf.properties.PropertiesFileConfigurationProvider. > >>>> loadSinks(PropertiesFileConfigurationProvider.java:373) > >>>> at > >>>> org.apache.flume.conf.properties.PropertiesFileConfigurationProvider. > >>>> load(PropertiesFileConfigurationProvider.java:223) > >>>> at > >>>> org.apache.flume.conf.file.AbstractFileConfigurationProvider.doLoad(A > >>>> bstractFileConfigurationProvider.java:123) > >>>> at > >>>> org.apache.flume.conf.file.AbstractFileConfigurationProvider.access$3 > >>>> 00(AbstractFileConfigurationProvider.java:38) > >>>> at > >>>> org.apache.flume.conf.file.AbstractFileConfigurationProvider$FileWatc > >>>> herRunnable.run(AbstractFileConfigurationProvider.java:202) > >>>> at > >>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:47 > >>>> 1) > >>>> at > >>>> java.util.concurrent.FutureTask$Sync.innerRunAndReset(FutureTask.java > >>>> :351) > >>>> at > >>>> java.util.concurrent.FutureTask.runAndReset(FutureTask.java:178) > >>>> at > >>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask. > >>>> access$301(ScheduledThreadPoolExecutor.java:178) > >>>> at > >>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask. > >>>> run(ScheduledThreadPoolExecutor.java:293) > >>>> at > >>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor. > >>>> java:1110) > >>>> at > >>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor > >>>> .java:603) > >>>> at java.lang.Thread.run(Thread.java:722) > >>>> Caused by: java.lang.ClassNotFoundException: > >>>> org.apache.commons.configuration.Co > >>>> nfiguration > >>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:366) > >>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:355) > >>>> at java.security.AccessController.doPrivileged(Native Method) > >>>> at java.net.URLClassLoader.findClass(URLClassLoader.java:354) > >>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:423) > >>>> at > sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) > >>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:356) > >>>> ... 26 more > >>>> 2012-07-31 18:18:26,601 (conf-file-poller-0) [DEBUG - > >>>> org.apache.flume.conf.file > >>>> > >>>> > .AbstractFileConfigurationProvider$FileWatcherRunnable.run(AbstractFileConfigura > >>>> tionProvider.java:188)] Checking > >>>> file:conf\flume-conf.properties.template for ch > >>>> anges > >>>> > >>>> > >>>> Please give me any solution if someone have. I dont understand what is > >>>> mean by dependencies were not found. > >>>> > >>>> Thanks > >>>> > >>>> > >>>> > >>> > >>> > >>> -- > >>> Apache MRUnit - Unit testing MapReduce - > >>> http://incubator.apache.org/mrunit/ > >>> > >> > >> > > > > > > -- > > Apache MRUnit - Unit testing MapReduce - > http://incubator.apache.org/mrunit/ > > > -- > Alexander Alten-Lorenz > http://mapredit.blogspot.com > German Hadoop LinkedIn Group: http://goo.gl/N8pCF > >
