Hi Stephane, if you are using the docker images from confluent, a way to set the levels to specific loggers is described here:
http://docs.confluent.io/3.1.1/cp-docker-images/docs/operations/logging.html#log4j-log-levels For Connect, you would need to set the environment variable CONNECT_LOG4J_LOGGERS in a similar way that KAFKA_LOG4J_LOGGERS is set in the "docker run" command described above. Regarding the redirection to stdout, if you are using Docker this is not configurable with the current templates because this allows you to view the logs for each container directly through docker via the command "docker logs <ContainerName>", which is the preferred way. Hope this helps, Konstantine On Mon, Jan 16, 2017 at 9:51 PM, Stephane Maarek < steph...@simplemachines.com.au> wrote: > The kind of output is the following: > > 05:15:34.878 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added > sensor with name connections-closed: > 05:15:34.879 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added > sensor with name connections-created: > 05:15:34.880 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added > sensor with name bytes-sent-received: > 05:15:34.881 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added > sensor with name bytes-sent: > 05:15:34.882 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added > sensor with name bytes-received: > 05:15:34.882 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added > sensor with name select-time: > 05:15:34.884 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added > sensor with name io-time: > 05:15:34.905 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added > sensor with name heartbeat-latency > 05:15:34.906 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added > sensor with name join-latency > 05:15:34.907 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added > sensor with name sync-latency > 05:15:34.970 [DistributedHerder] DEBUG > org.apache.kafka.common.metrics.Metrics - Added sensor with name > connections-closed: > 05:15:34.971 [DistributedHerder] DEBUG > org.apache.kafka.common.metrics.Metrics - Added sensor with name > connections-created: > 05:15:34.971 [DistributedHerder] DEBUG > org.apache.kafka.common.metrics.Metrics - Added sensor with name > bytes-sent-received: > 05:15:34.972 [DistributedHerder] DEBUG > org.apache.kafka.common.metrics.Metrics - Added sensor with name > bytes-sent: > 05:15:34.975 [DistributedHerder] DEBUG > org.apache.kafka.common.metrics.Metrics - Added sensor with name > bytes-received: > 05:15:34.977 [DistributedHerder] DEBUG > org.apache.kafka.common.metrics.Metrics - Added sensor with name > select-time: > 05:15:35.990 [DistributedHerder] DEBUG > org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group > kafka-connect-main has no committed offset for partition > _connect_offsets-39 > 05:15:35.990 [DistributedHerder] DEBUG > org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group > kafka-connect-main has no committed offset for partition _connect_offsets-6 > 05:15:35.990 [DistributedHerder] DEBUG > org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group > kafka-connect-main has no committed offset for partition > _connect_offsets-35 > 05:15:35.990 [DistributedHerder] DEBUG > org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group > kafka-connect-main has no committed offset for partition _connect_offsets-2 > 05:15:35.990 [DistributedHerder] DEBUG > org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group > kafka-connect-main has no committed offset for partition > _connect_offsets-31 > 05:15:35.990 [DistributedHerder] DEBUG > org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group > kafka-connect-main has no committed offset for partition > _connect_offsets-26 > 05:15:35.990 [DistributedHerder] DEBUG > org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group > kafka-connect-main has no committed offset for partition > _connect_offsets-22 > 05:15:35.991 [DistributedHerder] DEBUG > org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group > kafka-connect-main has no committed offset for partition > _connect_offsets-18 > 05:46:58.401 [CLASSPATH traversal thread.] DEBUG > org.reflections.Reflections - could not scan file > groovy/ui/icons/page_copy.png in url > file:/usr/share/java/kafka-connect-hdfs/groovy-all-2.1.6.jar with scanner > TypeAnnotationsScanner > 05:46:58.401 [CLASSPATH traversal thread.] DEBUG > org.reflections.Reflections - could not scan file > groovy/ui/icons/page_copy.png in url > file:/usr/share/java/kafka-connect-hdfs/groovy-all-2.1.6.jar with scanner > SubTypesScanner > > > *How do I stop all these loggers?* > > That’s what my connect-log4j.properties looks like: > > > log4j.rootLogger=INFO, stdout > > log4j.appender.stdout=org.apache.log4j.ConsoleAppender > log4j.appender.stdout.layout=org.apache.log4j.PatternLayout > log4j.appender.stdout.layout.ConversionPattern=[%d] %p %m (%c)%n > > log4j.logger.org.apache.kafka.clients.consumer=INFO, stdout > > I’m surprise because I set INFO at the rootLogger and it doesn’t seem to be > taken into account > > > > > On 16 January 2017 at 7:01:50 pm, Stephane Maarek ( > steph...@simplemachines.com.au) wrote: > > > Hi, > > I created my own connector and I’m launching it in cluster mode, but every > DEBUG statement is still going to the console. > How can I control the log level of Kafka Connect and its associated > connectors? I’m using the confluent docker image btw > > Thanks > Stephane >