Isn't that what the exception from SLF4J is?

On Aug 14, 2011, at 11:03 PM, Lance Norskog wrote:

> It might help to add a startup hack that complains somewhere if SLF4J
> is not hooked up right.
> 
> On Sat, Aug 13, 2011 at 3:51 PM, Ted Dunning <[email protected]> wrote:
>> Also, SLF4J is not a logging package, but a logging facade.  I think that
>> you need slf4j *and* a specific logging package.  You have slf4j-log4j, but
>> it isn't clear you have have log4j.  You may prefer to put the static slf4j
>> binding in the class path to avoid the need for log4j.
>> 
>> Logging packages are generally a pain in the butt since they tend to be kind
>> of fancy in what they require on the classpath.
>> 
>> On Sat, Aug 13, 2011 at 3:18 PM, Sean Owen <[email protected]> wrote:
>> 
>>> The unhelpful comment is that, well, the problem is exactly what you've
>>> already said: there is no SLF4J binding in the classpath where it needs to
>>> be. I suspect you're perhaps mixing up classpaths. Which actual Java task
>>> isn't printing -- the Hadoop worker? are you sure the bindings are in its
>>> classpath, versus say the classpath of the runner program?
>>> 
>>> On Sat, Aug 13, 2011 at 9:39 PM, Dhruv Kumar <[email protected]> wrote:
>>> 
>>>> I am running into a strange problem which has started happening only
>>> after
>>>> I
>>>> did a svn update on Mahout's trunk this morning.
>>>> 
>>>> For testing out MAHOUT-627, I have created some client code using
>>> IntelliJ
>>>> IDEA. In the jar artifact, I have added Mahout and Hadoop-0.20.203.0
>>>> libraries as dependencies.
>>>> 
>>>> If I run my code using:
>>>> 
>>>> dhruv@tachyon:~/hadoop-0.20.203.0/bin$ ./hadoop -jar
>>>> ~/TestBaumWelch/out/artifacts/TestBaumWelch_jar/TestBaumWelch.jar
>>>> TestBaumWelch.TestBaumWelch
>>>> 
>>>> I am unable to see any logs on the terminal or on the MapReduce web admin
>>>> console (http://localhost:50030/jobtracker.jsp). The admin console does
>>>> not
>>>> even show that any job is running.
>>>> 
>>>> But, I can clearly tell that the program is executing, since I get a
>>> series
>>>> of map 100%, reduce 100% and other INFO messages one after the other and
>>>> because it produces results in the output directories.
>>>> 
>>>> However, when I run the Hadoop's canonical Word Count example which does
>>>> not
>>>> depend on Mahout, I am able to see logs via the web admin console and
>>>> everything goes OK.
>>>> 
>>>> 
>>>> I suspect this is because of SLF4J on my system, there may be more than
>>> one
>>>> class binding or none at all. Here are the warning messages emitted soon
>>>> after I run the program:
>>>> 
>>>> SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
>>>> SLF4J: Defaulting to no-operation (NOP) logger implementation
>>>> SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for
>>> further
>>>> details.
>>>> 
>>>> 
>>>> However, slf4j is clearly on the classpath.
>>>> 
>>>> For hadoop, a portion of echo $CLASSPATH reveals:
>>>> 
>>>> :/home/dhruv/hadoop-0.20.203.0/bin/../lib/slf4j-log4j12-1.4.3.jar
>>>> 
>>>> For mahout:
>>>> 
>>>> 
>>>> 
>>> /home/dhruv/mahout/examples/target/dependency/slf4j-api-1.6.1.jar:/home/dhruv/mahout/examples/target/dependency/slf4j-jcl-1.6.1.jar:/home/dhruv/mahout/examples/target/dependency/slf4j-log4j12-1.6.1.jar:
>>>> 
>>>> 
>>>> Also, during Mahout's "mvn install" the slf4j is being added since I can
>>>> see
>>>> the following messages:
>>>> 
>>>> [INFO] org/slf4j/ already added, skipping
>>>> [INFO] org/slf4j/impl/ already added, skipping
>>>> [INFO] org/slf4j/impl/StaticLoggerBinder.class already added, skipping
>>>> [INFO] org/slf4j/impl/StaticMarkerBinder.class already added, skipping
>>>> [INFO] org/slf4j/impl/StaticMDCBinder.class already added, skipping
>>>> 
>>>> Has any one else run into a problem like this before?
>>>> 
>>> 
>> 
> 
> 
> 
> -- 
> Lance Norskog
> [email protected]

--------------------------------------------
Grant Ingersoll


Reply via email to