Some more thoughts. If we support unlimited negative scale, why can't we support unlimited positive scale? e.g. 0.0001 can be decimal(1, 4) instead of (4, 4). I think we need more references here: how other databases deal with decimal type and parse decimal literals?
On Mon, Jan 7, 2019 at 10:36 PM Wenchen Fan <cloud0...@gmail.com> wrote: > I'm OK with it, i.e. fail the write if there are negative-scale decimals > (we need to document it though). We can improve it later in data source v2. > > On Mon, Jan 7, 2019 at 10:09 PM Marco Gaido <marcogaid...@gmail.com> > wrote: > >> In general we can say that some datasources allow them, others fail. At >> the moment, we are doing no casting before writing (so we can state so in >> the doc). But since there is ongoing discussion for DSv2, we can maybe add >> a flag/interface there for "negative scale intollerant" DS and try and cast >> before writing to them. What do you think about this? >> >> Il giorno lun 7 gen 2019 alle ore 15:03 Wenchen Fan <cloud0...@gmail.com> >> ha scritto: >> >>> AFAIK parquet spec says decimal scale can't be negative. If we want to >>> officially support negative-scale decimal, we should clearly define the >>> behavior when writing negative-scale decimals to parquet and other data >>> sources. The most straightforward way is to fail for this case, but maybe >>> we can do something better, like casting decimal(1, -20) to decimal(20, 0) >>> before writing. >>> >>> On Mon, Jan 7, 2019 at 9:32 PM Marco Gaido <marcogaid...@gmail.com> >>> wrote: >>> >>>> Hi Wenchen, >>>> >>>> thanks for your email. I agree adding doc for decimal type, but I am >>>> not sure what you mean speaking of the behavior when writing: we are not >>>> performing any automatic casting before writing; if we want to do that, we >>>> need a design about it I think. >>>> >>>> I am not sure if it makes sense to set a min for it. That would break >>>> backward compatibility (for very weird use case), so I wouldn't do that. >>>> >>>> Thanks, >>>> Marco >>>> >>>> Il giorno lun 7 gen 2019 alle ore 05:53 Wenchen Fan < >>>> cloud0...@gmail.com> ha scritto: >>>> >>>>> I think we need to do this for backward compatibility, and according >>>>> to the discussion in the doc, SQL standard allows negative scale. >>>>> >>>>> To do this, I think the PR should also include a doc for the decimal >>>>> type, like the definition of precision and scale(this one >>>>> <https://stackoverflow.com/questions/35435691/bigdecimal-precision-and-scale> >>>>> looks pretty good), and the result type of decimal operations, and the >>>>> behavior when writing out decimals(e.g. we can cast decimal(1, -20) to >>>>> decimal(20, 0) before writing). >>>>> >>>>> Another question is, shall we set a min scale? e.g. shall we allow >>>>> decimal(1, -10000000)? >>>>> >>>>> On Thu, Oct 25, 2018 at 9:49 PM Marco Gaido <marcogaid...@gmail.com> >>>>> wrote: >>>>> >>>>>> Hi all, >>>>>> >>>>>> a bit more than one month ago, I sent a proposal for handling >>>>>> properly decimals with negative scales in our operations. This is a long >>>>>> standing problem in our codebase as we derived our rules from Hive and >>>>>> SQLServer where negative scales are forbidden, while in Spark they are >>>>>> not. >>>>>> >>>>>> The discussion has been stale for a while now. No more comments on >>>>>> the design doc: >>>>>> https://docs.google.com/document/d/17ScbMXJ83bO9lx8hB_jeJCSryhT9O_HDEcixDq0qmPk/edit#heading=h.x7062zmkubwm >>>>>> . >>>>>> >>>>>> So I am writing this e-mail in order to check whether there are more >>>>>> comments on it or we can go ahead with the PR. >>>>>> >>>>>> Thanks, >>>>>> Marco >>>>>> >>>>>