The kind of output is the following: 05:15:34.878 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name connections-closed: 05:15:34.879 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name connections-created: 05:15:34.880 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name bytes-sent-received: 05:15:34.881 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name bytes-sent: 05:15:34.882 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name bytes-received: 05:15:34.882 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name select-time: 05:15:34.884 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name io-time: 05:15:34.905 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name heartbeat-latency 05:15:34.906 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name join-latency 05:15:34.907 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name sync-latency 05:15:34.970 [DistributedHerder] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name connections-closed: 05:15:34.971 [DistributedHerder] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name connections-created: 05:15:34.971 [DistributedHerder] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name bytes-sent-received: 05:15:34.972 [DistributedHerder] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name bytes-sent: 05:15:34.975 [DistributedHerder] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name bytes-received: 05:15:34.977 [DistributedHerder] DEBUG org.apache.kafka.common.metrics.Metrics - Added sensor with name select-time: 05:15:35.990 [DistributedHerder] DEBUG org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group kafka-connect-main has no committed offset for partition _connect_offsets-39 05:15:35.990 [DistributedHerder] DEBUG org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group kafka-connect-main has no committed offset for partition _connect_offsets-6 05:15:35.990 [DistributedHerder] DEBUG org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group kafka-connect-main has no committed offset for partition _connect_offsets-35 05:15:35.990 [DistributedHerder] DEBUG org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group kafka-connect-main has no committed offset for partition _connect_offsets-2 05:15:35.990 [DistributedHerder] DEBUG org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group kafka-connect-main has no committed offset for partition _connect_offsets-31 05:15:35.990 [DistributedHerder] DEBUG org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group kafka-connect-main has no committed offset for partition _connect_offsets-26 05:15:35.990 [DistributedHerder] DEBUG org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group kafka-connect-main has no committed offset for partition _connect_offsets-22 05:15:35.991 [DistributedHerder] DEBUG org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group kafka-connect-main has no committed offset for partition _connect_offsets-18 05:46:58.401 [CLASSPATH traversal thread.] DEBUG org.reflections.Reflections - could not scan file groovy/ui/icons/page_copy.png in url file:/usr/share/java/kafka-connect-hdfs/groovy-all-2.1.6.jar with scanner TypeAnnotationsScanner 05:46:58.401 [CLASSPATH traversal thread.] DEBUG org.reflections.Reflections - could not scan file groovy/ui/icons/page_copy.png in url file:/usr/share/java/kafka-connect-hdfs/groovy-all-2.1.6.jar with scanner SubTypesScanner
*How do I stop all these loggers?* That’s what my connect-log4j.properties looks like: log4j.rootLogger=INFO, stdout log4j.appender.stdout=org.apache.log4j.ConsoleAppender log4j.appender.stdout.layout=org.apache.log4j.PatternLayout log4j.appender.stdout.layout.ConversionPattern=[%d] %p %m (%c)%n log4j.logger.org.apache.kafka.clients.consumer=INFO, stdout I’m surprise because I set INFO at the rootLogger and it doesn’t seem to be taken into account On 16 January 2017 at 7:01:50 pm, Stephane Maarek ( steph...@simplemachines.com.au) wrote: Hi, I created my own connector and I’m launching it in cluster mode, but every DEBUG statement is still going to the console. How can I control the log level of Kafka Connect and its associated connectors? I’m using the confluent docker image btw Thanks Stephane