Hi Stephane,

if you are using the docker images from confluent, a way to set the levels
to specific loggers is described here:

http://docs.confluent.io/3.1.1/cp-docker-images/docs/operations/logging.html#log4j-log-levels

For Connect, you would need to set the environment variable
CONNECT_LOG4J_LOGGERS in a similar way that KAFKA_LOG4J_LOGGERS is set in
the "docker run" command described above.

Regarding the redirection to stdout, if you are using Docker this is not
configurable with the current templates because this allows you to view the
logs for each container directly through docker via the command "docker
logs <ContainerName>", which is the preferred way.

Hope this helps,
Konstantine


On Mon, Jan 16, 2017 at 9:51 PM, Stephane Maarek <
steph...@simplemachines.com.au> wrote:

> The kind of output is the following:
>
> 05:15:34.878 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name connections-closed:
> 05:15:34.879 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name connections-created:
> 05:15:34.880 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name bytes-sent-received:
> 05:15:34.881 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name bytes-sent:
> 05:15:34.882 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name bytes-received:
> 05:15:34.882 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name select-time:
> 05:15:34.884 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name io-time:
> 05:15:34.905 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name heartbeat-latency
> 05:15:34.906 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name join-latency
> 05:15:34.907 [main] DEBUG org.apache.kafka.common.metrics.Metrics - Added
> sensor with name sync-latency
> 05:15:34.970 [DistributedHerder] DEBUG
> org.apache.kafka.common.metrics.Metrics - Added sensor with name
> connections-closed:
> 05:15:34.971 [DistributedHerder] DEBUG
> org.apache.kafka.common.metrics.Metrics - Added sensor with name
> connections-created:
> 05:15:34.971 [DistributedHerder] DEBUG
> org.apache.kafka.common.metrics.Metrics - Added sensor with name
> bytes-sent-received:
> 05:15:34.972 [DistributedHerder] DEBUG
> org.apache.kafka.common.metrics.Metrics - Added sensor with name
> bytes-sent:
> 05:15:34.975 [DistributedHerder] DEBUG
> org.apache.kafka.common.metrics.Metrics - Added sensor with name
> bytes-received:
> 05:15:34.977 [DistributedHerder] DEBUG
> org.apache.kafka.common.metrics.Metrics - Added sensor with name
> select-time:
> 05:15:35.990 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition
> _connect_offsets-39
> 05:15:35.990 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition _connect_offsets-6
> 05:15:35.990 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition
> _connect_offsets-35
> 05:15:35.990 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition _connect_offsets-2
> 05:15:35.990 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition
> _connect_offsets-31
> 05:15:35.990 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition
> _connect_offsets-26
> 05:15:35.990 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition
> _connect_offsets-22
> 05:15:35.991 [DistributedHerder] DEBUG
> org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - Group
> kafka-connect-main has no committed offset for partition
> _connect_offsets-18
> 05:46:58.401 [CLASSPATH traversal thread.] DEBUG
> org.reflections.Reflections - could not scan file
> groovy/ui/icons/page_copy.png in url
> file:/usr/share/java/kafka-connect-hdfs/groovy-all-2.1.6.jar with scanner
> TypeAnnotationsScanner
> 05:46:58.401 [CLASSPATH traversal thread.] DEBUG
> org.reflections.Reflections - could not scan file
> groovy/ui/icons/page_copy.png in url
> file:/usr/share/java/kafka-connect-hdfs/groovy-all-2.1.6.jar with scanner
> SubTypesScanner
>
>
> *How do I stop all these loggers?*
>
> That’s what my connect-log4j.properties looks like:
>
>
> log4j.rootLogger=INFO, stdout
>
> log4j.appender.stdout=org.apache.log4j.ConsoleAppender
> log4j.appender.stdout.layout=org.apache.log4j.PatternLayout
> log4j.appender.stdout.layout.ConversionPattern=[%d] %p %m (%c)%n
>
> log4j.logger.org.apache.kafka.clients.consumer=INFO, stdout
>
> I’m surprise because I set INFO at the rootLogger and it doesn’t seem to be
> taken into account
>
>
>
>
> On 16 January 2017 at 7:01:50 pm, Stephane Maarek (
> steph...@simplemachines.com.au) wrote:
>
>
> Hi,
>
> I created my own connector and I’m launching it in cluster mode, but every
> DEBUG statement is still going to the console.
> How can I control the log level of Kafka Connect and its associated
> connectors? I’m using the confluent docker image btw
>
> Thanks
> Stephane
>

Reply via email to