Hi Kafka Community,
Sorry due to some changes I had to change the link to the kip.
Here is the updated KIP link:
https://cwiki.apache.org/confluence/display/KAFKA/KIP-1154%3A+Extending+support+for+Microsecond+Precision+for+Kafka+Connect

On Sat, Apr 5, 2025 at 12:14 AM pritam kumar <kumarpritamm...@gmail.com>
wrote:

> Hi Kafka Community,
>
> I’d like to start a discussion on KIP-1153: Extending Support for
> Microsecond Precision for Kafka Connect
> <https://cwiki.apache.org/confluence/display/KAFKA/KIP-1153%3A+Extending+Support+for+Microsecond+Precision+for+Kafka+Connect>
> .
>
> The primary motivation behind this KIP is to enhance the precision of
> timestamp handling in Kafka Connect. Currently, Kafka Connect is limited to
> millisecond-level precision for timestamps. However, many modern data
> formats and platforms have moved beyond this limitation:
>
>    -
>
>    Formats such as *Avro*, *Parquet*, and *ORC* support microsecond (and
>    even nanosecond) precision. For example, Avro specifies support for
>    timestamp-micros (spec link
>    <https://avro.apache.org/docs/1.11.0/spec.html#timestamp-micros>).
>    -
>
>    Sink systems like *Apache Iceberg*, *Delta Lake*, and *Apache Hudi*
>     offer *microsecond and nanosecond precision* for time-based fields,
>    making millisecond precision inadequate for accurate data replication and
>    analytics in some use cases.
>
> This gap can lead to *loss of precision* when transferring data through
> Kafka Connect, especially when interacting with systems that depend on
> high-resolution timestamps for change tracking, event ordering, or
> deduplication.
>
> The goal of this KIP is to:
>
>    -
>
>    Introduce microsecond-level timestamp handling in Kafka Connect schema
>    and data representation.
>    -
>
>    Ensure connectors (both source and sink) can leverage this precision
>    when supported by the underlying data systems.
>    -
>
>    Maintain backward compatibility with existing millisecond-based
>    configurations and data.
>
> We welcome community feedback on:
>
>    -
>
>    Potential implementation concerns or edge cases we should address
>    -
>
>    Suggestions for schema enhancements or conversion strategies
>    -
>
>    Impacts on connector compatibility and testing
>
> Looking forward to your thoughts and input on this proposal!
>
> Thanks.
>  Link to the KIP.
> https://cwiki.apache.org/confluence/display/KAFKA/KIP-1153%3A+Extending+Support+for+Microsecond+Precision+for+Kafka+Connect
> <https://cwiki.apache.org/confluence/display/KAFKA/KIP-1152%3A+Add+transactional+ID+prefix+filter+to+ListTransactions+API>
>
>

Reply via email to