“Exponentially exposed” -> “potentially exposed” On Wed, Feb 21, 2024 at 4:13 PM Li Jin <ice.xell...@gmail.com> wrote:
> Thanks - since we don’t control all the invocation of pq.write_table, I > wonder if there is some configuration for the “default” behavior? > > Also I wonder if there are other API surface that is exponentially exposed > to this, e.g., dataset or pd.Dataframe.to_parquet ? > > Thanks! > Li > > On Wed, Feb 21, 2024 at 3:53 PM Jacek Pliszka <jacek.plis...@gmail.com> > wrote: > >> Hi! >> >> pq.write_table( >> table, config.output_filename, coerce_timestamps="us", >> allow_truncated_timestamps=True, >> ) >> >> allows you to write as us instead of ns. >> >> BR >> >> J >> >> >> śr., 21 lut 2024 o 21:44 Li Jin <ice.xell...@gmail.com> napisał(a): >> >> > Hi, >> > >> > My colleague has informed me that during the Arrow 12->15 upgrade, he >> found >> > that writing a pandas Dataframe with datetime64[ns] to parquet will >> result >> > in nanosecond metadata and nanosecond values. >> > >> > I wonder if this is something configurable to the old behavior so we can >> > enable “nanosecond in parquet” gradually? There are code that reads >> parquet >> > files that don’t handle parquet nanosecond now. >> > >> > Thanks! >> > Li >> > >> >