The kind of output is the following:

05:15:34.878 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
sensor with name connections-closed:
05:15:34.879 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
sensor with name connections-created:
05:15:34.880 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
sensor with name bytes-sent-received:
05:15:34.881 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
sensor with name bytes-sent:
05:15:34.882 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
sensor with name bytes-received:
05:15:34.882 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
sensor with name select-time:
05:15:34.884 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
sensor with name io-time:
05:15:34.905 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
sensor with name heartbeat-latency
05:15:34.906 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
sensor with name join-latency
05:15:34.907 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
sensor with name sync-latency
05:15:34.970 [DistributedHerder] DEBUG
org.apache.kafka.common.metrics.Metrics - Added sensor with name
connections-closed:
05:15:34.971 [DistributedHerder] DEBUG
org.apache.kafka.common.metrics.Metrics - Added sensor with name
connections-created:
05:15:34.971 [DistributedHerder] DEBUG
org.apache.kafka.common.metrics.Metrics - Added sensor with name
bytes-sent-received:
05:15:34.972 [DistributedHerder] DEBUG
org.apache.kafka.common.metrics.Metrics - Added sensor with name bytes-sent:
05:15:34.975 [DistributedHerder] DEBUG
org.apache.kafka.common.metrics.Metrics - Added sensor with name
bytes-received:
05:15:34.977 [DistributedHerder] DEBUG
org.apache.kafka.common.metrics.Metrics - Added sensor with name
select-time:
05:15:35.990 [DistributedHerder] DEBUG
org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
kafka-connect-main has no committed offset for partition _connect_offsets-39
05:15:35.990 [DistributedHerder] DEBUG
org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
kafka-connect-main has no committed offset for partition _connect_offsets-6
05:15:35.990 [DistributedHerder] DEBUG
org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
kafka-connect-main has no committed offset for partition _connect_offsets-35
05:15:35.990 [DistributedHerder] DEBUG
org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
kafka-connect-main has no committed offset for partition _connect_offsets-2
05:15:35.990 [DistributedHerder] DEBUG
org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
kafka-connect-main has no committed offset for partition _connect_offsets-31
05:15:35.990 [DistributedHerder] DEBUG
org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
kafka-connect-main has no committed offset for partition _connect_offsets-26
05:15:35.990 [DistributedHerder] DEBUG
org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
kafka-connect-main has no committed offset for partition _connect_offsets-22
05:15:35.991 [DistributedHerder] DEBUG
org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
kafka-connect-main has no committed offset for partition _connect_offsets-18
05:46:58.401 [CLASSPATH traversal thread.] DEBUG
org.reflections.Reflections - could not scan file
groovy/ui/icons/page_copy.png in url
file:/usr/share/java/kafka-connect-hdfs/groovy-all-2.1.6.jar with scanner
TypeAnnotationsScanner
05:46:58.401 [CLASSPATH traversal thread.] DEBUG
org.reflections.Reflections - could not scan file
groovy/ui/icons/page_copy.png in url
file:/usr/share/java/kafka-connect-hdfs/groovy-all-2.1.6.jar with scanner
SubTypesScanner


*How do I stop all these loggers?*

That’s what my connect-log4j.properties looks like:


log4j.rootLogger=INFO, stdout

log4j.appender.stdout=org.apache.log4j.ConsoleAppender
log4j.appender.stdout.layout=org.apache.log4j.PatternLayout
log4j.appender.stdout.layout.ConversionPattern=[%d] %p %m (%c)%n

log4j.logger.org.apache.kafka.clients.consumer=INFO, stdout

I’m surprise because I set INFO at the rootLogger and it doesn’t seem to be
taken into account




On 16 January 2017 at 7:01:50 pm, Stephane Maarek (
steph...@simplemachines.com.au) wrote:


Hi,

I created my own connector and I’m launching it in cluster mode, but every
DEBUG statement is still going to the console.
How can I control the log level of Kafka Connect and its associated
connectors? I’m using the confluent docker image btw

Thanks
Stephane

Reply via email to