> > If yes then `timestamp_as_object` keyword arguments seems like a new > feature, so strictly speaking it's not a regression compared to the > previous release.
Yes, I don't think we should be releasing new features that are know to be half baked and based on discussions elsewhere will likely need a backward compatibility mode just in case users come to rely on the flawed implementation. I think we should remove or cause the flag to error for the 1.0 release at least, so we aren't digging ourselves further into a hole. On Mon, Jul 20, 2020 at 12:41 PM Krisztián Szűcs <szucs.kriszt...@gmail.com> wrote: > The conversations in the pull requests are pretty broad so I'm just > guessing, but do you refer that `to_pandas(timestamp_as_object=True)` > drops the timezone information? > If yes then `timestamp_as_object` keyword arguments seems like a new > feature, so strictly speaking it's not a regression compared to the > previous release. > > I agree that we shouldn't leave known bugs (I don't like it either), > but I'm afraid proper timezone support will require more effort. Like > currently we also strip timezone information when converting from > datetime.time(..., tzinfo) objects, or the missing timezone support in > the temporal casts. > > On Mon, Jul 20, 2020 at 7:36 PM Micah Kornfield <emkornfi...@gmail.com> > wrote: > > > > I just wanted to clarify. doing a full rollback of the patch means that > https://issues.apache.org/jira/browse/ARROW-5359 would get released out > of the gate with a bug in it. > > > > On Mon, Jul 20, 2020 at 7:48 AM Antoine Pitrou <anto...@python.org> > wrote: > >> > >> > >> If the release condition is for the regression to be fixed in less than > >> 24 hours (less than 12 hours now?), I think we should simply revert the > >> original PR and work on a fix more leisurely for 1.1.0 (or even 1.0.1). > >> > >> Unless it really causes havoc for Spark users, in which case a > >> circumvention should be found. > >> > >> Regards > >> > >> Antoine. > >> > >> > >> Le 20/07/2020 à 16:46, Krisztián Szűcs a écrit : > >> > If I understand correctly we used to just store the timestamp and the > >> > timezone if an explicit arrow type was passed during the python->arrow > >> > conversion, but the timestamp values were not changed in any way. > >> > Micah's current patch changes the python->arrow conversion behavior to > >> > normalize all values to utc timestamps. > >> > > >> > While it's definitely an improvement over the previously ignored > >> > timezones, I'm not sure that it won't cause unexpected regressions in > >> > the users' codebases. > >> > I'm still trying to better understand the issue and its compatibility > >> > implications, but my intuition tells me that we should apply the > >> > reversion instead and properly handle the datetime value conversions > >> > in an upcoming minor release. > >> > > >> > Either way we should move this conversation to the pull request [1], > >> > because the code snippets pasted here are hardly readable. > >> > > >> > [1]: https://github.com/apache/arrow/pull/7805 > >> > > >> > On Mon, Jul 20, 2020 at 9:40 AM Sutou Kouhei <k...@clear-code.com> > wrote: > >> >> > >> >> Done: > https://github.com/apache/arrow/pull/7805#issuecomment-660855376 > >> >> > >> >> We can use ...-3.8-... not ...-3.7-... because we don't have > >> >> ...-3.7-... task in > >> >> https://github.com/apache/arrow/blob/master/dev/tasks/tasks.yml. > >> >> > >> >> In < > cak7z5t8hqcsd3meg42cuzkscpjr3zndsvrjmm8vied0gzto...@mail.gmail.com> > >> >> "Re: [VOTE] Release Apache Arrow 1.0.0 - RC1" on Mon, 20 Jul 2020 > 00:14:00 -0700, > >> >> Micah Kornfield <emkornfi...@gmail.com> wrote: > >> >> > >> >>> FYI, I'm not sure if it is a permissions issue or I've done > something wrong > >> >>> but github-actions does not seem to be responding to > "@github-actions > >> >>> <https://github.com/github-actions> crossbow submit > >> >>> test-conda-python-3.7-spark-master" when I enter it. If someone > could kick > >> >>> off the spark integration test I would be grateful. > >> >>> > >> >>> On Mon, Jul 20, 2020 at 12:09 AM Micah Kornfield < > emkornfi...@gmail.com> > >> >>> wrote: > >> >>> > >> >>>> Thanks Bryan. I cherry-picked your change onto my change [1] > which now > >> >>>> honors timezone aware datetime objects on ingestion. I've kicked > off the > >> >>>> spark integration tests. > >> >>>> > >> >>>> If this change doesn't work I think the correct course of action > is to > >> >>>> provide an environment variable in python to turn back to the old > behavior > >> >>>> (ignoring timezones on conversion). I think honoring timezone > information > >> >>>> where possible is a strict improvement but I agree we should give > users an > >> >>>> option to not break if they wish to upgrade to the latest > version. I need > >> >>>> to get some sleep but I will have another PR posted tomorrow > evening if the > >> >>>> current one doesn't unblock the release. > >> >>>> > >> >>>> [1] https://github.com/apache/arrow/pull/7805 > >> >>>> > >> >>>> On Sun, Jul 19, 2020 at 10:50 PM Bryan Cutler <cutl...@gmail.com> > wrote: > >> >>>> > >> >>>>> I'd rather not see ARROW-9223 reverted, if possible. I will put > up my > >> >>>>> hacked patch to Spark for this so we can test against it if > needed, and > >> >>>>> could share my branch if anyone else wants to test it locally. > >> >>>>> > >> >>>>> On Sun, Jul 19, 2020 at 7:35 PM Micah Kornfield < > emkornfi...@gmail.com> > >> >>>>> wrote: > >> >>>>> > >> >>>>>> I'll spend some time tonight on it and if I can't get round trip > working > >> >>>>>> I'll handle reverting > >> >>>>>> > >> >>>>>> On Sunday, July 19, 2020, Wes McKinney <wesmck...@gmail.com> > wrote: > >> >>>>>> > >> >>>>>>> On Sun, Jul 19, 2020 at 7:33 PM Neal Richardson > >> >>>>>>> <neal.p.richard...@gmail.com> wrote: > >> >>>>>>>> > >> >>>>>>>> It sounds like you may have identified a pyarrow bug, which > sounds > >> >>>>> not > >> >>>>>>>> good, though I don't know enough about the broader context to > know > >> >>>>>>> whether > >> >>>>>>>> this is (1) worse than 0.17 or (2) release blocking. I defer to > >> >>>>> y'all > >> >>>>>> who > >> >>>>>>>> know better. > >> >>>>>>>> > >> >>>>>>>> If there are quirks in how Spark handles timezone-naive > timestamps, > >> >>>>>>>> shouldn't the fix/workaround go in pyspark, not pyarrow? For > what > >> >>>>> it's > >> >>>>>>>> worth, I dealt with similar Spark timezone issues in R > recently: > >> >>>>>>>> https://github.com/sparklyr/sparklyr/issues/2439 I handled > with it > >> >>>>> (in > >> >>>>>>>> sparklyr, not the arrow R package) by always setting a > timezone when > >> >>>>>>>> sending data to Spark. Not ideal but it made the numbers > "right". > >> >>>>>>> > >> >>>>>>> Since people are running this code in production we need to be > careful > >> >>>>>>> about disrupting them. Unfortunately I'm at the limit of how > much time > >> >>>>>>> I can spend on this, but releasing with ARROW-9223 as is > (without > >> >>>>>>> being partially or fully reverted) makes me deeply > uncomfortable. So I > >> >>>>>>> hope the matter can be resolved. > >> >>>>>>> > >> >>>>>>>> Neal > >> >>>>>>>> > >> >>>>>>>> > >> >>>>>>>> On Sun, Jul 19, 2020 at 5:13 PM Wes McKinney < > wesmck...@gmail.com> > >> >>>>>>> wrote: > >> >>>>>>>> > >> >>>>>>>>> Honestly I think reverting is the best option. This change > >> >>>>> evidently > >> >>>>>>>>> needs more time to "season" and perhaps this is motivation to > >> >>>>> enhance > >> >>>>>>>>> test coverage in a number of places. > >> >>>>>>>>> > >> >>>>>>>>> On Sun, Jul 19, 2020 at 7:11 PM Wes McKinney < > wesmck...@gmail.com > >> >>>>>> > >> >>>>>>> wrote: > >> >>>>>>>>>> > >> >>>>>>>>>> I am OK with any solution that doesn't delay the production > of > >> >>>>> the > >> >>>>>>>>>> next RC by more than 24 hours > >> >>>>>>>>>> > >> >>>>>>>>>> On Sun, Jul 19, 2020 at 7:08 PM Micah Kornfield < > >> >>>>>>> emkornfi...@gmail.com> > >> >>>>>>>>> wrote: > >> >>>>>>>>>>> > >> >>>>>>>>>>> If I read the example right it looks like constructing from > >> >>>>>> python > >> >>>>>>>>> types > >> >>>>>>>>>>> isn't keeping timezones into in place? I can try make a > patch > >> >>>>>> that > >> >>>>>>>>> fixes > >> >>>>>>>>>>> that tonight or would the preference be to revert my patch > >> >>>>> (note > >> >>>>>> I > >> >>>>>>>>> think > >> >>>>>>>>>>> another bug with a prior bug was fixed in my PR as well) > >> >>>>>>>>>>> > >> >>>>>>>>>>> -Micah > >> >>>>>>>>>>> > >> >>>>>>>>>>> On Sunday, July 19, 2020, Wes McKinney <wesmck...@gmail.com > > > >> >>>>>>> wrote: > >> >>>>>>>>>>> > >> >>>>>>>>>>>> I think I see the problem now: > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> In [40]: parr > >> >>>>>>>>>>>> Out[40]: > >> >>>>>>>>>>>> 0 {'f0': 1969-12-31 16:00:00-08:00} > >> >>>>>>>>>>>> 1 {'f0': 1969-12-31 16:00:00.000001-08:00} > >> >>>>>>>>>>>> 2 {'f0': 1969-12-31 16:00:00.000002-08:00} > >> >>>>>>>>>>>> dtype: object > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> In [41]: parr[0]['f0'] > >> >>>>>>>>>>>> Out[41]: datetime.datetime(1969, 12, 31, 16, 0, > >> >>>>>> tzinfo=<DstTzInfo > >> >>>>>>>>>>>> 'America/Los_Angeles' PST-1 day, 16:00:00 STD>) > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> In [42]: pa.array(parr) > >> >>>>>>>>>>>> Out[42]: > >> >>>>>>>>>>>> <pyarrow.lib.StructArray object at 0x7f0893706a60> > >> >>>>>>>>>>>> -- is_valid: all not null > >> >>>>>>>>>>>> -- child 0 type: timestamp[us] > >> >>>>>>>>>>>> [ > >> >>>>>>>>>>>> 1969-12-31 16:00:00.000000, > >> >>>>>>>>>>>> 1969-12-31 16:00:00.000001, > >> >>>>>>>>>>>> 1969-12-31 16:00:00.000002 > >> >>>>>>>>>>>> ] > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> In [43]: pa.array(parr).field(0).type > >> >>>>>>>>>>>> Out[43]: TimestampType(timestamp[us]) > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> On 0.17.1 > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> In [8]: arr = pa.array([0, 1, 2], type=pa.timestamp('us', > >> >>>>>>>>>>>> 'America/Los_Angeles')) > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> In [9]: arr > >> >>>>>>>>>>>> Out[9]: > >> >>>>>>>>>>>> <pyarrow.lib.TimestampArray object at 0x7f9dede69d00> > >> >>>>>>>>>>>> [ > >> >>>>>>>>>>>> 1970-01-01 00:00:00.000000, > >> >>>>>>>>>>>> 1970-01-01 00:00:00.000001, > >> >>>>>>>>>>>> 1970-01-01 00:00:00.000002 > >> >>>>>>>>>>>> ] > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> In [10]: struct_arr = pa.StructArray.from_arrays([arr], > >> >>>>>>> names=['f0']) > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> In [11]: struct_arr > >> >>>>>>>>>>>> Out[11]: > >> >>>>>>>>>>>> <pyarrow.lib.StructArray object at 0x7f9ded0016e0> > >> >>>>>>>>>>>> -- is_valid: all not null > >> >>>>>>>>>>>> -- child 0 type: timestamp[us, tz=America/Los_Angeles] > >> >>>>>>>>>>>> [ > >> >>>>>>>>>>>> 1970-01-01 00:00:00.000000, > >> >>>>>>>>>>>> 1970-01-01 00:00:00.000001, > >> >>>>>>>>>>>> 1970-01-01 00:00:00.000002 > >> >>>>>>>>>>>> ] > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> In [12]: struct_arr.to_pandas() > >> >>>>>>>>>>>> Out[12]: > >> >>>>>>>>>>>> 0 {'f0': 1970-01-01 00:00:00} > >> >>>>>>>>>>>> 1 {'f0': 1970-01-01 00:00:00.000001} > >> >>>>>>>>>>>> 2 {'f0': 1970-01-01 00:00:00.000002} > >> >>>>>>>>>>>> dtype: object > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> In [13]: pa.array(struct_arr.to_pandas()) > >> >>>>>>>>>>>> Out[13]: > >> >>>>>>>>>>>> <pyarrow.lib.StructArray object at 0x7f9ded003210> > >> >>>>>>>>>>>> -- is_valid: all not null > >> >>>>>>>>>>>> -- child 0 type: timestamp[us] > >> >>>>>>>>>>>> [ > >> >>>>>>>>>>>> 1970-01-01 00:00:00.000000, > >> >>>>>>>>>>>> 1970-01-01 00:00:00.000001, > >> >>>>>>>>>>>> 1970-01-01 00:00:00.000002 > >> >>>>>>>>>>>> ] > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> In [14]: pa.array(struct_arr.to_pandas()).type > >> >>>>>>>>>>>> Out[14]: StructType(struct<f0: timestamp[us]>) > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> So while the time zone is getting stripped in both cases, > >> >>>>> the > >> >>>>>>> failure > >> >>>>>>>>>>>> to round trip is a problem. If we are going to attach the > >> >>>>> time > >> >>>>>>> zone > >> >>>>>>>>> in > >> >>>>>>>>>>>> to_pandas() then we need to respect it when going the other > >> >>>>>> way. > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> This looks like a regression to me and so I'm inclined to > >> >>>>>> revise > >> >>>>>>> my > >> >>>>>>>>>>>> vote on the release to -0/-1 > >> >>>>>>>>>>>> > >> >>>>>>>>>>>> On Sun, Jul 19, 2020 at 6:46 PM Wes McKinney < > >> >>>>>>> wesmck...@gmail.com> > >> >>>>>>>>> wrote: > >> >>>>>>>>>>>>> > >> >>>>>>>>>>>>> Ah I forgot that this is a "feature" of nanosecond > >> >>>>> timestamps > >> >>>>>>>>>>>>> > >> >>>>>>>>>>>>> In [21]: arr = pa.array([0, 1, 2], type=pa.timestamp('us', > >> >>>>>>>>>>>>> 'America/Los_Angeles')) > >> >>>>>>>>>>>>> > >> >>>>>>>>>>>>> In [22]: struct_arr = pa.StructArray.from_arrays([arr], > >> >>>>>>>>> names=['f0']) > >> >>>>>>>>>>>>> > >> >>>>>>>>>>>>> In [23]: struct_arr.to_pandas() > >> >>>>>>>>>>>>> Out[23]: > >> >>>>>>>>>>>>> 0 {'f0': 1969-12-31 16:00:00-08:00} > >> >>>>>>>>>>>>> 1 {'f0': 1969-12-31 16:00:00.000001-08:00} > >> >>>>>>>>>>>>> 2 {'f0': 1969-12-31 16:00:00.000002-08:00} > >> >>>>>>>>>>>>> dtype: object > >> >>>>>>>>>>>>> > >> >>>>>>>>>>>>> So this is working as intended, such as it is > >> >>>>>>>>>>>>> > >> >>>>>>>>>>>>> On Sun, Jul 19, 2020 at 6:40 PM Wes McKinney < > >> >>>>>>> wesmck...@gmail.com> > >> >>>>>>>>>>>> wrote: > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> There seems to be other broken StructArray stuff > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> In [14]: arr = pa.array([0, 1, 2], > >> >>>>> type=pa.timestamp('ns', > >> >>>>>>>>>>>>>> 'America/Los_Angeles')) > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> In [15]: struct_arr = pa.StructArray.from_arrays([arr], > >> >>>>>>>>> names=['f0']) > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> In [16]: struct_arr > >> >>>>>>>>>>>>>> Out[16]: > >> >>>>>>>>>>>>>> <pyarrow.lib.StructArray object at 0x7f089370f590> > >> >>>>>>>>>>>>>> -- is_valid: all not null > >> >>>>>>>>>>>>>> -- child 0 type: timestamp[ns, tz=America/Los_Angeles] > >> >>>>>>>>>>>>>> [ > >> >>>>>>>>>>>>>> 1970-01-01 00:00:00.000000000, > >> >>>>>>>>>>>>>> 1970-01-01 00:00:00.000000001, > >> >>>>>>>>>>>>>> 1970-01-01 00:00:00.000000002 > >> >>>>>>>>>>>>>> ] > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> In [17]: struct_arr.to_pandas() > >> >>>>>>>>>>>>>> Out[17]: > >> >>>>>>>>>>>>>> 0 {'f0': 0} > >> >>>>>>>>>>>>>> 1 {'f0': 1} > >> >>>>>>>>>>>>>> 2 {'f0': 2} > >> >>>>>>>>>>>>>> dtype: object > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> All in all it appears that this part of the project > >> >>>>> needs > >> >>>>>>> some > >> >>>>>>>>> TLC > >> >>>>>>>>>>>>>> > >> >>>>>>>>>>>>>> On Sun, Jul 19, 2020 at 6:16 PM Wes McKinney < > >> >>>>>>>>> wesmck...@gmail.com> > >> >>>>>>>>>>>> wrote: > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> Well, the problem is that time zones are really > >> >>>>> finicky > >> >>>>>>>>> comparing > >> >>>>>>>>>>>>>>> Spark (which uses a localtime interpretation of > >> >>>>>> timestamps > >> >>>>>>>>> without > >> >>>>>>>>>>>>>>> time zone) and Arrow (which has naive timestamps -- a > >> >>>>>>> concept > >> >>>>>>>>> similar > >> >>>>>>>>>>>>>>> but different from the SQL concept TIMESTAMP WITHOUT > >> >>>>> TIME > >> >>>>>>> ZONE > >> >>>>>>>>> -- and > >> >>>>>>>>>>>>>>> tz-aware timestamps). So somewhere there is a time > >> >>>>> zone > >> >>>>>>> being > >> >>>>>>>>>>>> stripped > >> >>>>>>>>>>>>>>> or applied/localized which may result in the > >> >>>>> transferred > >> >>>>>>> data > >> >>>>>>>>> to/from > >> >>>>>>>>>>>>>>> Spark being shifted by the time zone offset. I think > >> >>>>> it's > >> >>>>>>>>> important > >> >>>>>>>>>>>>>>> that we determine what the problem is -- if it's a > >> >>>>>> problem > >> >>>>>>>>> that has > >> >>>>>>>>>>>> to > >> >>>>>>>>>>>>>>> be fixed in Arrow (and it's not clear to me that it > >> >>>>> is) > >> >>>>>>> it's > >> >>>>>>>>> worth > >> >>>>>>>>>>>>>>> spending some time to understand what's going on to > >> >>>>> avoid > >> >>>>>>> the > >> >>>>>>>>>>>>>>> possibility of patch release on account of this. > >> >>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>> On Sun, Jul 19, 2020 at 6:12 PM Neal Richardson > >> >>>>>>>>>>>>>>> <neal.p.richard...@gmail.com> wrote: > >> >>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>> If it’s a display problem, should it block the > >> >>>>> release? > >> >>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>> Sent from my iPhone > >> >>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>> On Jul 19, 2020, at 3:57 PM, Wes McKinney < > >> >>>>>>>>> wesmck...@gmail.com> > >> >>>>>>>>>>>> wrote: > >> >>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>> I opened https://issues.apache.org/ > >> >>>>>>> jira/browse/ARROW-9525 > >> >>>>>>>>>>>> about the > >> >>>>>>>>>>>>>>>>> display problem. My guess is that there are other > >> >>>>>>> problems > >> >>>>>>>>>>>> lurking > >> >>>>>>>>>>>>>>>>> here > >> >>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>> On Sun, Jul 19, 2020 at 5:54 PM Wes McKinney < > >> >>>>>>>>>>>> wesmck...@gmail.com> wrote: > >> >>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>> hi Bryan, > >> >>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>> This is a display bug > >> >>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>> In [6]: arr = pa.array([0, 1, 2], > >> >>>>>>> type=pa.timestamp('ns', > >> >>>>>>>>>>>>>>>>>> 'America/Los_Angeles')) > >> >>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>> In [7]: arr.view('int64') > >> >>>>>>>>>>>>>>>>>> Out[7]: > >> >>>>>>>>>>>>>>>>>> <pyarrow.lib.Int64Array object at 0x7fd1b8aaef30> > >> >>>>>>>>>>>>>>>>>> [ > >> >>>>>>>>>>>>>>>>>> 0, > >> >>>>>>>>>>>>>>>>>> 1, > >> >>>>>>>>>>>>>>>>>> 2 > >> >>>>>>>>>>>>>>>>>> ] > >> >>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>> In [8]: arr > >> >>>>>>>>>>>>>>>>>> Out[8]: > >> >>>>>>>>>>>>>>>>>> <pyarrow.lib.TimestampArray object at > >> >>>>>> 0x7fd1b8aae6e0> > >> >>>>>>>>>>>>>>>>>> [ > >> >>>>>>>>>>>>>>>>>> 1970-01-01 00:00:00.000000000, > >> >>>>>>>>>>>>>>>>>> 1970-01-01 00:00:00.000000001, > >> >>>>>>>>>>>>>>>>>> 1970-01-01 00:00:00.000000002 > >> >>>>>>>>>>>>>>>>>> ] > >> >>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>> In [9]: arr.to_pandas() > >> >>>>>>>>>>>>>>>>>> Out[9]: > >> >>>>>>>>>>>>>>>>>> 0 1969-12-31 16:00:00-08:00 > >> >>>>>>>>>>>>>>>>>> 1 1969-12-31 16:00:00.000000001-08:00 > >> >>>>>>>>>>>>>>>>>> 2 1969-12-31 16:00:00.000000002-08:00 > >> >>>>>>>>>>>>>>>>>> dtype: datetime64[ns, America/Los_Angeles] > >> >>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>> the repr of TimestampArray doesn't take into > >> >>>>> account > >> >>>>>>> the > >> >>>>>>>>>>>> timezone > >> >>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>> In [10]: arr[0] > >> >>>>>>>>>>>>>>>>>> Out[10]: <pyarrow.TimestampScalar: > >> >>>>>>> Timestamp('1969-12-31 > >> >>>>>>>>>>>>>>>>>> 16:00:00-0800', tz='America/Los_Angeles')> > >> >>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>> So if it's incorrect, the problem is happening > >> >>>>>>> somewhere > >> >>>>>>>>> before > >> >>>>>>>>>>>> or > >> >>>>>>>>>>>>>>>>>> while the StructArray is being created. If I had > >> >>>>> to > >> >>>>>>> guess > >> >>>>>>>>> it's > >> >>>>>>>>>>>> caused > >> >>>>>>>>>>>>>>>>>> by the tzinfo of the datetime.datetime values not > >> >>>>>>> being > >> >>>>>>>>> handled > >> >>>>>>>>>>>> in the > >> >>>>>>>>>>>>>>>>>> way that they were before > >> >>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>> On Sun, Jul 19, 2020 at 5:19 PM Wes McKinney < > >> >>>>>>>>>>>> wesmck...@gmail.com> wrote: > >> >>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>> Well this is not good and pretty disappointing > >> >>>>>> given > >> >>>>>>>>> that we > >> >>>>>>>>>>>> had nearly a month to sort through the implications of > >> >>>>> Micah’s > >> >>>>>>>>> patch. We > >> >>>>>>>>>>>> should try to resolve this ASAP > >> >>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>> On Sun, Jul 19, 2020 at 5:10 PM Bryan Cutler < > >> >>>>>>>>>>>> cutl...@gmail.com> wrote: > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> +0 (non-binding) > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> I ran verification script for binaries and then > >> >>>>>>> source, > >> >>>>>>>>> as > >> >>>>>>>>>>>> below, and both > >> >>>>>>>>>>>>>>>>>>>> look good > >> >>>>>>>>>>>>>>>>>>>> ARROW_TMPDIR=/tmp/arrow-test TEST_DEFAULT=0 > >> >>>>>>>>> TEST_SOURCE=1 > >> >>>>>>>>>>>> TEST_CPP=1 > >> >>>>>>>>>>>>>>>>>>>> TEST_PYTHON=1 TEST_JAVA=1 > >> >>>>> TEST_INTEGRATION_CPP=1 > >> >>>>>>>>>>>> TEST_INTEGRATION_JAVA=1 > >> >>>>>>>>>>>>>>>>>>>> dev/release/verify-release-candidate.sh source > >> >>>>>>> 1.0.0 1 > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> I tried to patch Spark locally to verify the > >> >>>>>> recent > >> >>>>>>>>> change in > >> >>>>>>>>>>>> nested > >> >>>>>>>>>>>>>>>>>>>> timestamps and was not able to get things > >> >>>>> working > >> >>>>>>> quite > >> >>>>>>>>>>>> right, but I'm not > >> >>>>>>>>>>>>>>>>>>>> sure if the problem is in Spark, Arrow or my > >> >>>>>> patch - > >> >>>>>>>>> hence my > >> >>>>>>>>>>>> vote of +0. > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> Here is what I'm seeing > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> ``` > >> >>>>>>>>>>>>>>>>>>>> (Input as datetime) > >> >>>>>>>>>>>>>>>>>>>> datetime.datetime(2018, 3, 10, 0, 0) > >> >>>>>>>>>>>>>>>>>>>> datetime.datetime(2018, 3, 15, 0, 0) > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> (Struct Array) > >> >>>>>>>>>>>>>>>>>>>> -- is_valid: all not null > >> >>>>>>>>>>>>>>>>>>>> -- child 0 type: timestamp[us, > >> >>>>>>> tz=America/Los_Angeles] > >> >>>>>>>>>>>>>>>>>>>> [ > >> >>>>>>>>>>>>>>>>>>>> 2018-03-10 00:00:00.000000, > >> >>>>>>>>>>>>>>>>>>>> 2018-03-10 00:00:00.000000 > >> >>>>>>>>>>>>>>>>>>>> ] > >> >>>>>>>>>>>>>>>>>>>> -- child 1 type: timestamp[us, > >> >>>>>>> tz=America/Los_Angeles] > >> >>>>>>>>>>>>>>>>>>>> [ > >> >>>>>>>>>>>>>>>>>>>> 2018-03-15 00:00:00.000000, > >> >>>>>>>>>>>>>>>>>>>> 2018-03-15 00:00:00.000000 > >> >>>>>>>>>>>>>>>>>>>> ] > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> (Flattened Arrays) > >> >>>>>>>>>>>>>>>>>>>> types [TimestampType(timestamp[us, > >> >>>>>>>>> tz=America/Los_Angeles]), > >> >>>>>>>>>>>>>>>>>>>> TimestampType(timestamp[us, > >> >>>>>>> tz=America/Los_Angeles])] > >> >>>>>>>>>>>>>>>>>>>> [<pyarrow.lib.TimestampArray object at > >> >>>>>>> 0x7ffbbd88f520> > >> >>>>>>>>>>>>>>>>>>>> [ > >> >>>>>>>>>>>>>>>>>>>> 2018-03-10 00:00:00.000000, > >> >>>>>>>>>>>>>>>>>>>> 2018-03-10 00:00:00.000000 > >> >>>>>>>>>>>>>>>>>>>> ], <pyarrow.lib.TimestampArray object at > >> >>>>>>> 0x7ffba958be50> > >> >>>>>>>>>>>>>>>>>>>> [ > >> >>>>>>>>>>>>>>>>>>>> 2018-03-15 00:00:00.000000, > >> >>>>>>>>>>>>>>>>>>>> 2018-03-15 00:00:00.000000 > >> >>>>>>>>>>>>>>>>>>>> ]] > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> (Pandas Conversion) > >> >>>>>>>>>>>>>>>>>>>> [ > >> >>>>>>>>>>>>>>>>>>>> 0 2018-03-09 16:00:00-08:00 > >> >>>>>>>>>>>>>>>>>>>> 1 2018-03-09 16:00:00-08:00 > >> >>>>>>>>>>>>>>>>>>>> dtype: datetime64[ns, America/Los_Angeles], > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> 0 2018-03-14 17:00:00-07:00 > >> >>>>>>>>>>>>>>>>>>>> 1 2018-03-14 17:00:00-07:00 > >> >>>>>>>>>>>>>>>>>>>> dtype: datetime64[ns, America/Los_Angeles]] > >> >>>>>>>>>>>>>>>>>>>> ``` > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> Based on output of existing a correct timestamp > >> >>>>>>> udf, it > >> >>>>>>>>> looks > >> >>>>>>>>>>>> like the > >> >>>>>>>>>>>>>>>>>>>> pyarrow Struct Array values are wrong and > >> >>>>> that's > >> >>>>>>> carried > >> >>>>>>>>>>>> through the > >> >>>>>>>>>>>>>>>>>>>> flattened arrays, causing the Pandas values to > >> >>>>>> have > >> >>>>>>> a > >> >>>>>>>>>>>> negative offset. > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> Here is output from a working udf with > >> >>>>> timestamp, > >> >>>>>>> the > >> >>>>>>>>> pyarrow > >> >>>>>>>>>>>> Array > >> >>>>>>>>>>>>>>>>>>>> displays in UTC time, I believe. > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> ``` > >> >>>>>>>>>>>>>>>>>>>> (Timestamp Array) > >> >>>>>>>>>>>>>>>>>>>> type timestamp[us, tz=America/Los_Angeles] > >> >>>>>>>>>>>>>>>>>>>> [ > >> >>>>>>>>>>>>>>>>>>>> [ > >> >>>>>>>>>>>>>>>>>>>> 1969-01-01 09:01:01.000000 > >> >>>>>>>>>>>>>>>>>>>> ] > >> >>>>>>>>>>>>>>>>>>>> ] > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> (Pandas Conversion) > >> >>>>>>>>>>>>>>>>>>>> 0 1969-01-01 01:01:01-08:00 > >> >>>>>>>>>>>>>>>>>>>> Name: _0, dtype: datetime64[ns, > >> >>>>>> America/Los_Angeles] > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> (Timezone Localized) > >> >>>>>>>>>>>>>>>>>>>> 0 1969-01-01 01:01:01 > >> >>>>>>>>>>>>>>>>>>>> Name: _0, dtype: datetime64[ns] > >> >>>>>>>>>>>>>>>>>>>> ``` > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> I'll have to dig in further at another time and > >> >>>>>>> debug > >> >>>>>>>>> where > >> >>>>>>>>>>>> the values go > >> >>>>>>>>>>>>>>>>>>>> wrong. > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>> On Sat, Jul 18, 2020 at 9:51 PM Micah > >> >>>>> Kornfield < > >> >>>>>>>>>>>> emkornfi...@gmail.com> > >> >>>>>>>>>>>>>>>>>>>> wrote: > >> >>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>> +1 (binding) > >> >>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>> Ran wheel and binary tests on ubuntu 19.04 > >> >>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>> On Fri, Jul 17, 2020 at 2:25 PM Neal > >> >>>>> Richardson < > >> >>>>>>>>>>>>>>>>>>>>> neal.p.richard...@gmail.com> > >> >>>>>>>>>>>>>>>>>>>>> wrote: > >> >>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>> +1 (binding) > >> >>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>> In addition to the usual verification on > >> >>>>>>>>>>>>>>>>>>>>>> https://github.com/apache/arrow/pull/7787, > >> >>>>> I've > >> >>>>>>>>>>>> successfully staged the > >> >>>>>>>>>>>>>>>>>>>>> R > >> >>>>>>>>>>>>>>>>>>>>>> binary artifacts on Windows ( > >> >>>>>>>>>>>>>>>>>>>>>> https://github.com/r-windows/ > >> >>>>>>> rtools-packages/pull/126 > >> >>>>>>>>> ), > >> >>>>>>>>>>>> macOS ( > >> >>>>>>>>>>>>>>>>>>>>>> > >> >>>>>> https://github.com/autobrew/homebrew-core/pull/12 > >> >>>>>>> ), > >> >>>>>>>>> and > >> >>>>>>>>>>>> Linux ( > >> >>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>> https://github.com/ursa-labs/arrow-r-nightly/actions/runs/ > >> >>>>>>>>>>>> 172977277) > >> >>>>>>>>>>>>>>>>>>>>> using > >> >>>>>>>>>>>>>>>>>>>>>> the release candidate. > >> >>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>> And I agree with the judgment about skipping > >> >>>>> a > >> >>>>>> JS > >> >>>>>>>>> release > >> >>>>>>>>>>>> artifact. Looks > >> >>>>>>>>>>>>>>>>>>>>>> like there hasn't been a code change since > >> >>>>>>> October so > >> >>>>>>>>>>>> there's no point. > >> >>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>> Neal > >> >>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>> On Fri, Jul 17, 2020 at 10:37 AM Wes > >> >>>>> McKinney < > >> >>>>>>>>>>>> wesmck...@gmail.com> > >> >>>>>>>>>>>>>>>>>>>>> wrote: > >> >>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>> I see the JS failures as well. I think it > >> >>>>> is a > >> >>>>>>>>> failure > >> >>>>>>>>>>>> localized to > >> >>>>>>>>>>>>>>>>>>>>>>> newer Node versions since our JavaScript CI > >> >>>>>> works > >> >>>>>>>>> fine. I > >> >>>>>>>>>>>> don't think > >> >>>>>>>>>>>>>>>>>>>>>>> it should block the release given the lack > >> >>>>> of > >> >>>>>>>>> development > >> >>>>>>>>>>>> activity in > >> >>>>>>>>>>>>>>>>>>>>>>> JavaScript [1] -- if any JS devs are > >> >>>>> concerned > >> >>>>>>> about > >> >>>>>>>>>>>> publishing an > >> >>>>>>>>>>>>>>>>>>>>>>> artifact then we can skip pushing it to NPM > >> >>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>> @Ryan it seems it may be something > >> >>>>> environment > >> >>>>>>>>> related on > >> >>>>>>>>>>>> your > >> >>>>>>>>>>>>>>>>>>>>>>> machine, I'm on Ubuntu 18.04 and have not > >> >>>>> seen > >> >>>>>>> this. > >> >>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>> On > >> >>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>> * Python 3.8 wheel's tests are failed. > >> >>>>> 3.5, > >> >>>>>> 3.6 > >> >>>>>>>>> and 3.7 > >> >>>>>>>>>>>>>>>>>>>>>>>> are passed. It seems that -larrow and > >> >>>>>>>>> -larrow_python > >> >>>>>>>>>>>> for > >> >>>>>>>>>>>>>>>>>>>>>>>> Cython are failed. > >> >>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>> I suspect this is related to > >> >>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>> https://github.com/apache/arrow/commit/ > >> >>>>>>>>>>>> 120c21f4bf66d2901b3a353a1f67bac3c3355924#diff- > >> >>>>>>>>>>>> 0f69784b44040448d17d0e4e8a641fe8 > >> >>>>>>>>>>>>>>>>>>>>>>> , > >> >>>>>>>>>>>>>>>>>>>>>>> but I don't think it's a blocking issue > >> >>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>> [1]: > >> >>>>>>>>> https://github.com/apache/arrow/commits/master/js > >> >>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>> On Fri, Jul 17, 2020 at 9:42 AM Ryan Murray > >> >>>>> < > >> >>>>>>>>>>>> rym...@dremio.com> wrote: > >> >>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>> I've tested Java and it looks good. However > >> >>>>>> the > >> >>>>>>>>> verify > >> >>>>>>>>>>>> script keeps > >> >>>>>>>>>>>>>>>>>>>>> on > >> >>>>>>>>>>>>>>>>>>>>>>>> bailing with protobuf related errors: > >> >>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>> 'cpp/build/orc_ep-prefix/src/orc_ep-build/c++/src/orc_ > >> >>>>>>>>>>>> proto.pb.cc' > >> >>>>>>>>>>>>>>>>>>>>> and > >> >>>>>>>>>>>>>>>>>>>>>>>> friends cant find protobuf definitions. A > >> >>>>> bit > >> >>>>>>> odd as > >> >>>>>>>>>>>> cmake can see > >> >>>>>>>>>>>>>>>>>>>>>>> protobuf > >> >>>>>>>>>>>>>>>>>>>>>>>> headers and builds directly off master work > >> >>>>>> just > >> >>>>>>>>> fine. > >> >>>>>>>>>>>> Has anyone > >> >>>>>>>>>>>>>>>>>>>>> else > >> >>>>>>>>>>>>>>>>>>>>>>>> experienced this? I am on ubutnu 18.04 > >> >>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>> On Fri, Jul 17, 2020 at 10:49 AM Antoine > >> >>>>>> Pitrou > >> >>>>>>> < > >> >>>>>>>>>>>> anto...@python.org> > >> >>>>>>>>>>>>>>>>>>>>>>> wrote: > >> >>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>> +1 (binding). I tested on Ubuntu 18.04. > >> >>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>> * Wheels verification went fine. > >> >>>>>>>>>>>>>>>>>>>>>>>>> * Source verification went fine with CUDA > >> >>>>>>> enabled > >> >>>>>>>>> and > >> >>>>>>>>>>>>>>>>>>>>>>>>> TEST_INTEGRATION_JS=0 TEST_JS=0. > >> >>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>> I didn't test the binaries. > >> >>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>> Regards > >> >>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>> Antoine. > >> >>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>> Le 17/07/2020 à 03:41, Krisztián Szűcs a > >> >>>>>> écrit > >> >>>>>>> : > >> >>>>>>>>>>>>>>>>>>>>>>>>>> Hi, > >> >>>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>>> I would like to propose the second > >> >>>>> release > >> >>>>>>>>> candidate > >> >>>>>>>>>>>> (RC1) of > >> >>>>>>>>>>>>>>>>>>>>>> Apache > >> >>>>>>>>>>>>>>>>>>>>>>>>>> Arrow version 1.0.0. > >> >>>>>>>>>>>>>>>>>>>>>>>>>> This is a major release consisting of 826 > >> >>>>>>>>> resolved JIRA > >> >>>>>>>>>>>>>>>>>>>>> issues[1]. > >> >>>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>>> The verification of the first release > >> >>>>>>> candidate > >> >>>>>>>>> (RC0) > >> >>>>>>>>>>>> has failed > >> >>>>>>>>>>>>>>>>>>>>>>> [0], and > >> >>>>>>>>>>>>>>>>>>>>>>>>>> the packaging scripts were unable to > >> >>>>> produce > >> >>>>>>> two > >> >>>>>>>>>>>> wheels. Compared > >> >>>>>>>>>>>>>>>>>>>>>>>>>> to RC0 this release candidate includes > >> >>>>>>> additional > >> >>>>>>>>>>>> patches for the > >> >>>>>>>>>>>>>>>>>>>>>>>>>> following bugs: ARROW-9506, ARROW-9504, > >> >>>>>>>>> ARROW-9497, > >> >>>>>>>>>>>>>>>>>>>>>>>>>> ARROW-9500, ARROW-9499. > >> >>>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>>> This release candidate is based on > >> >>>>> commit: > >> >>>>>>>>>>>>>>>>>>>>>>>>>> bc0649541859095ee77d03a7b891ea8d6e2fd641 > >> >>>>> [2] > >> >>>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>>> The source release rc1 is hosted at [3]. > >> >>>>>>>>>>>>>>>>>>>>>>>>>> The binary artifacts are hosted at > >> >>>>>>> [4][5][6][7]. > >> >>>>>>>>>>>>>>>>>>>>>>>>>> The changelog is located at [8]. > >> >>>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>>> Please download, verify checksums and > >> >>>>>>> signatures, > >> >>>>>>>>> run > >> >>>>>>>>>>>> the unit > >> >>>>>>>>>>>>>>>>>>>>>> tests, > >> >>>>>>>>>>>>>>>>>>>>>>>>>> and vote on the release. See [9] for how > >> >>>>> to > >> >>>>>>>>> validate a > >> >>>>>>>>>>>> release > >> >>>>>>>>>>>>>>>>>>>>>>> candidate. > >> >>>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>>> The vote will be open for at least 72 > >> >>>>> hours. > >> >>>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>>> [ ] +1 Release this as Apache Arrow 1.0.0 > >> >>>>>>>>>>>>>>>>>>>>>>>>>> [ ] +0 > >> >>>>>>>>>>>>>>>>>>>>>>>>>> [ ] -1 Do not release this as Apache > >> >>>>> Arrow > >> >>>>>>> 1.0.0 > >> >>>>>>>>>>>> because... > >> >>>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>>> [0]: > >> >>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>> https://github.com/apache/arrow/pull/7778#issuecomment- > >> >>>>>>>>>>>> 659065370 > >> >>>>>>>>>>>>>>>>>>>>>>>>>> [1]: > >> >>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>> https://issues.apache.org/ > >> >>>>>>> jira/issues/?jql=project%20% > >> >>>>>>>>>>>> 3D%20ARROW%20AND%20status%20in%20%28Resolved%2C% > >> >>>>>>> 20Closed%29%20AND% > >> >>>>>>>>>>>> 20fixVersion%20%3D%201.0.0 > >> >>>>>>>>>>>>>>>>>>>>>>>>>> [2]: > >> >>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>> https://github.com/apache/arrow/tree/ > >> >>>>>>>>>>>> bc0649541859095ee77d03a7b891ea8d6e2fd641 > >> >>>>>>>>>>>>>>>>>>>>>>>>>> [3]: > >> >>>>>>>>>>>>>>>>>>>>>>> https://dist.apache.org/repos/ > >> >>>>>>>>>>>> dist/dev/arrow/apache-arrow-1.0.0-rc1 > >> >>>>>>>>>>>>>>>>>>>>>>>>>> [4]: https://bintray.com/apache/ > >> >>>>>>>>>>>> arrow/centos-rc/1.0.0-rc1 > >> >>>>>>>>>>>>>>>>>>>>>>>>>> [5]: https://bintray.com/apache/ > >> >>>>>>>>>>>> arrow/debian-rc/1.0.0-rc1 > >> >>>>>>>>>>>>>>>>>>>>>>>>>> [6]: https://bintray.com/apache/ > >> >>>>>>>>>>>> arrow/python-rc/1.0.0-rc1 > >> >>>>>>>>>>>>>>>>>>>>>>>>>> [7]: https://bintray.com/apache/ > >> >>>>>>>>>>>> arrow/ubuntu-rc/1.0.0-rc1 > >> >>>>>>>>>>>>>>>>>>>>>>>>>> [8]: > >> >>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>> https://github.com/apache/arrow/blob/ > >> >>>>>>>>>>>> bc0649541859095ee77d03a7b891ea8d6e2fd641/CHANGELOG.md > >> >>>>>>>>>>>>>>>>>>>>>>>>>> [9]: > >> >>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>> https://cwiki.apache.org/ > >> >>>>>>> confluence/display/ARROW/How+ > >> >>>>>>>>>>>> to+Verify+Release+Candidates > >> >>>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>>>>>>>>>>> > >> >>>>>>>>>>>> > >> >>>>>>>>> > >> >>>>>>> > >> >>>>>> > >> >>>>> > >> >>>> >