I've been using Flume with CDH 4.5.0 and it has been working without any problems. Today all of a sudden I'm getting what I think is some sort of classpath error when I try to start my agent from the command line. Running the default agent from Cloudera Manager seems to work fine.
Info: Including Hadoop libraries found via (/usr/bin/hadoop) for HDFS access > Exception in thread "main" java.lang.NoClassDefFoundError: > org/apache/flume/tools/GetJavaProperty > Caused by: java.lang.ClassNotFoundException: > org.apache.flume.tools.GetJavaProperty > at java.net.URLClassLoader$1.run(URLClassLoader.java:202) > at java.security.AccessController.doPrivileged(Native Method) > at java.net.URLClassLoader.findClass(URLClassLoader.java:190) > at java.lang.ClassLoader.loadClass(ClassLoader.java:306) > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301) > at java.lang.ClassLoader.loadClass(ClassLoader.java:247) I don't recall changing anything in regards to classpaths. I'm trying to run the agent directly on the master node (namenode, secondarynamenode, hdfs gateway, mapreduce gateway) but I'm running it from the command line (versus from Cloudera Manager). which hadoop /usr/bin/hadoop which java /usr/bin/java hadoop classpath /etc/hadoop/conf:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/hadoop/libexec/../../hadoop/lib/*:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/hadoop/libexec/../../hadoop/.//*:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/hadoop/libexec/../../hadoop-hdfs/./:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/hadoop/libexec/../../hadoop-hdfs/lib/*:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/hadoop/libexec/../../hadoop-hdfs/.//*:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/hadoop/libexec/../../hadoop-yarn/lib/*:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/hadoop/libexec/../../hadoop-yarn/.//*:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/hadoop/libexec/../../hadoop-0.20-mapreduce/./:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/hadoop/libexec/../../hadoop-0.20-mapreduce/lib/*:/opt/cloudera/parcels/CDH-4.5.0-1.cdh4.5.0.p0.30/lib/hadoop/libexec/../../hadoop-0.20-mapreduce/.//* Running normal hadoop HDFS commands like "hadoop fs -ls" works just fine without having to specify the HDFS host. I tried redeploying all the client configs and even restarting all the services to no avail. Previously I had no issues running fume from the command line so I know I must have messed up a setting somewhere. Does anyone have any ideas as to what would be causing this issue? Thank you! Best Regards, Ed
