[arrow] Diff for: [GitHub] kszucs closed pull request #3403: ARROW-4260: [Python] NumPy buffer protocol failure
diff --git a/ci/appveyor-cpp-build.bat b/ci/appveyor-cpp-build.bat index 387dd55d18..78f5e41928 100644 --- a/ci/appveyor-cpp-build.bat +++ b/ci/appveyor-cpp-build.bat @@ -95,11 +95,12 @@ if "%JOB%" == "Build_Debug" ( exit /B 0 ) -conda create -n arrow -q -y ^ +conda create -n arrow -q -y -c conda-forge ^ + --file=ci\conda_env_python.yml ^ python=%PYTHON% ^ - six pytest setuptools numpy pandas cython hypothesis ^ - thrift-cpp=0.11.0 boost-cpp ^ - -c conda-forge + numpy=1.14 ^ + thrift-cpp=0.11 ^ + boost-cpp call activate arrow @@ -109,9 +110,9 @@ set BOOST_LIBRARYDIR=%CONDA_PREFIX%\Library\lib if "%JOB%" == "Toolchain" ( @rem Install pre-built "toolchain" packages for faster builds - conda install -q -y --file=ci\conda_env_cpp.yml ^ -python=%PYTHON% ^ --c conda-forge + conda install -q -y -c conda-forge ^ +--file=ci\conda_env_cpp.yml ^ +python=%PYTHON% set ARROW_BUILD_TOOLCHAIN=%CONDA_PREFIX%\Library ) diff --git a/ci/cpp-msvc-build-main.bat b/ci/cpp-msvc-build-main.bat index 644170775d..d9d7e548dd 100644 --- a/ci/cpp-msvc-build-main.bat +++ b/ci/cpp-msvc-build-main.bat @@ -72,7 +72,7 @@ popd pushd python -pip install pickle5 +pip install -r requirements.txt pickle5 set PYARROW_CXXFLAGS=%ARROW_CXXFLAGS% set PYARROW_CMAKE_GENERATOR=%GENERATOR% diff --git a/ci/travis_script_python.sh b/ci/travis_script_python.sh index e9a1122755..c3e2d1903c 100755 --- a/ci/travis_script_python.sh +++ b/ci/travis_script_python.sh @@ -50,7 +50,7 @@ conda create -y -q -p $CONDA_ENV_DIR \ nomkl \ cmake \ pip \ - numpy=1.13.1 \ + numpy=1.14 \ python=${PYTHON_VERSION} \ ${CONDA_JVM_DEPS} @@ -124,7 +124,7 @@ $ARROW_CPP_BUILD_DIR/$ARROW_BUILD_TYPE/arrow-python-test pushd $ARROW_PYTHON_DIR # Other stuff pip install -pip install -q -r requirements.txt +pip install -r requirements.txt if [ "$PYTHON_VERSION" == "3.6" ]; then pip install -q pickle5 diff --git a/python/pyarrow/tests/test_convert_pandas.py b/python/pyarrow/tests/test_convert_pandas.py index 466d2e9562..264b51c32f 100644 --- a/python/pyarrow/tests/test_convert_pandas.py +++ b/python/pyarrow/tests/test_convert_pandas.py @@ -23,6 +23,7 @@ from collections import OrderedDict from datetime import date, datetime, time, timedelta +from distutils.version import LooseVersion import hypothesis as h import hypothesis.extra.pytz as tzst @@ -2224,8 +2225,6 @@ def test_safe_unsafe_casts(self): def _fully_loaded_dataframe_example(): -from distutils.version import LooseVersion - index = pd.MultiIndex.from_arrays([ pd.date_range('2000-01-01', periods=5).repeat(2), np.tile(np.array(['foo', 'bar'], dtype=object), 5) @@ -2271,6 +2270,8 @@ def _check_serialize_components_roundtrip(df): tm.assert_frame_equal(df, deserialized) +@pytest.mark.skipif(LooseVersion(np.__version__) >= '0.16', +reason='Until numpy/numpy#12745 is resolved') def test_serialize_deserialize_pandas(): # ARROW-1784, serialize and deserialize DataFrame by decomposing # BlockManager With regards, Apache Git Services
[arrow] branch master updated: ARROW-4260: [Python] NumPy buffer protocol failure
This is an automated email from the ASF dual-hosted git repository. kszucs pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/arrow.git The following commit(s) were added to refs/heads/master by this push: new 09d3496 ARROW-4260: [Python] NumPy buffer protocol failure 09d3496 is described below commit 09d34964996a718beb2def4e4173974f8a3d862e Author: Krisztián Szűcs AuthorDate: Tue Jan 15 17:02:26 2019 +0100 ARROW-4260: [Python] NumPy buffer protocol failure The following error only occurred when both numpy=1.16 and pickle5 was installed. It's a numpy issue https://github.com/numpy/numpy/issues/12745, so skipping the test until it's resolved. ``` == FAILURES === __ test_serialize_deserialize_pandas __ ValueError: cannot include dtype 'M' in a buffer The above exception was the direct cause of the following exception: def test_serialize_deserialize_pandas(): # ARROW-1784, serialize and deserialize DataFrame by decomposing # BlockManager df = _fully_loaded_dataframe_example() > _check_serialize_components_roundtrip(df) pyarrow/tests/test_convert_pandas.py:2278: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ pyarrow/tests/test_convert_pandas.py:2268: in _check_serialize_components_roundtrip components = ctx.serialize(df).to_components() pyarrow/serialization.pxi:198: in pyarrow.lib.SerializationContext.serialize return serialize(obj, context=self) pyarrow/serialization.pxi:345: in pyarrow.lib.serialize check_status(SerializeObject(context, wrapped_value, )) pyarrow/serialization.pxi:153: in pyarrow.lib.SerializationContext._serialize_callback serialized_obj = {"data": self.custom_serializers[type_id](obj)} _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ x = MultiIndex(levels=[[2000-01-01 00:00:00, 2000-01-02 00:00:00, 2000-01-03 00:00:00, 2000-01-04 00:00:00, 2000-01-05 00:00:00], ['bar', 'foo']], labels=[[0, 0, 1, 1, 2, 2, 3, 3, 4, 4], [1, 0, 1, 0, 1, 0, 1, 0, 1, 0]]) def _pickle_to_buffer(x): > pickled = builtin_pickle.dumps(x, protocol=builtin_pickle.HIGHEST_PROTOCOL) E SystemError: returned a result with an error set ``` Author: Krisztián Szűcs Closes #3403 from kszucs/travis_numpy_version and squashes the following commits: e3162a8b pin to minor 77e9b64f reason 301226d1 skip serialize test 39a7161d appveyor ae1f16d9 pin numpy to 1.14 --- ci/appveyor-cpp-build.bat | 15 --- ci/cpp-msvc-build-main.bat | 2 +- ci/travis_script_python.sh | 4 ++-- python/pyarrow/tests/test_convert_pandas.py | 5 +++-- 4 files changed, 14 insertions(+), 12 deletions(-) diff --git a/ci/appveyor-cpp-build.bat b/ci/appveyor-cpp-build.bat index 387dd55..78f5e41 100644 --- a/ci/appveyor-cpp-build.bat +++ b/ci/appveyor-cpp-build.bat @@ -95,11 +95,12 @@ if "%JOB%" == "Build_Debug" ( exit /B 0 ) -conda create -n arrow -q -y ^ +conda create -n arrow -q -y -c conda-forge ^ + --file=ci\conda_env_python.yml ^ python=%PYTHON% ^ - six pytest setuptools numpy pandas cython hypothesis ^ - thrift-cpp=0.11.0 boost-cpp ^ - -c conda-forge + numpy=1.14 ^ + thrift-cpp=0.11 ^ + boost-cpp call activate arrow @@ -109,9 +110,9 @@ set BOOST_LIBRARYDIR=%CONDA_PREFIX%\Library\lib if "%JOB%" == "Toolchain" ( @rem Install pre-built "toolchain" packages for faster builds - conda install -q -y --file=ci\conda_env_cpp.yml ^ -python=%PYTHON% ^ --c conda-forge + conda install -q -y -c conda-forge ^ +--file=ci\conda_env_cpp.yml ^ +python=%PYTHON% set ARROW_BUILD_TOOLCHAIN=%CONDA_PREFIX%\Library ) diff --git a/ci/cpp-msvc-build-main.bat b/ci/cpp-msvc-build-main.bat index 6441707..d9d7e54 100644 --- a/ci/cpp-msvc-build-main.bat +++ b/ci/cpp-msvc-build-main.bat @@ -72,7 +72,7 @@ popd pushd python -pip install pickle5 +pip install -r requirements.txt pickle5 set PYARROW_CXXFLAGS=%ARROW_CXXFLAGS% set PYARROW_CMAKE_GENERATOR=%GENERATOR% diff --git a/ci/travis_script_python.sh b/ci/travis_script_python.sh index e9a1122..c3e2d19 100755 --- a/ci/travis_script_python.sh +++ b/ci/travis_script_python.sh @@ -50,7 +50,7 @@ conda create -y -q -p $CONDA_ENV_DIR \ nomkl \ cmake \ pip \ - numpy=1.13.1 \ + numpy=1.14 \ python=${PYTHON_VERSION} \ ${CONDA_JVM_DEPS} @@ -124,7 +124,7 @@
[arrow] branch master updated: ARROW-4266: [Python][CI] Disable ORC tests in dask integration test
This is an automated email from the ASF dual-hosted git repository. kszucs pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/arrow.git The following commit(s) were added to refs/heads/master by this push: new 5a7507c ARROW-4266: [Python][CI] Disable ORC tests in dask integration test 5a7507c is described below commit 5a7507c9f28f5175c367f483accc072891f53905 Author: Krisztián Szűcs AuthorDate: Tue Jan 15 17:31:48 2019 +0100 ARROW-4266: [Python][CI] Disable ORC tests in dask integration test [kszucs/crossbow/build-415](https://github.com/kszucs/crossbow/branches/all?utf8=%E2%9C%93=build-415) Author: Krisztián Szűcs Closes #3405 from kszucs/ARROW-4266 and squashes the following commits: 16bc7642 don't run orc tests --- integration/dask/runtest.sh | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/integration/dask/runtest.sh b/integration/dask/runtest.sh index 9a37e0a..baf9ccf 100755 --- a/integration/dask/runtest.sh +++ b/integration/dask/runtest.sh @@ -29,6 +29,7 @@ python -c "import pyarrow.parquet" # pytest -sv --pyargs dask.bytes.tests.test_hdfs # pytest -sv --pyargs dask.bytes.tests.test_local -pytest -v --pyargs dask.dataframe.io.tests.test_orc +# TODO(kszucs): re-enable it, for more see ARROW-3910 +# pytest -v --pyargs dask.dataframe.io.tests.test_orc pytest -v --pyargs dask.dataframe.io.tests.test_parquet pytest -v --pyargs dask.dataframe.tests.test_dataframe
[arrow] Diff for: [GitHub] kszucs closed pull request #3405: ARROW-4266: [Python][CI] Disable ORC tests in dask integration test
diff --git a/integration/dask/runtest.sh b/integration/dask/runtest.sh index 9a37e0a67b..baf9ccf445 100755 --- a/integration/dask/runtest.sh +++ b/integration/dask/runtest.sh @@ -29,6 +29,7 @@ python -c "import pyarrow.parquet" # pytest -sv --pyargs dask.bytes.tests.test_hdfs # pytest -sv --pyargs dask.bytes.tests.test_local -pytest -v --pyargs dask.dataframe.io.tests.test_orc +# TODO(kszucs): re-enable it, for more see ARROW-3910 +# pytest -v --pyargs dask.dataframe.io.tests.test_orc pytest -v --pyargs dask.dataframe.io.tests.test_parquet pytest -v --pyargs dask.dataframe.tests.test_dataframe With regards, Apache Git Services
[arrow] Diff for: [GitHub] wesm closed pull request #3395: ARROW-4258: [Python] Safe cast fails from numpy float64 array with nans to integer
diff --git a/cpp/src/arrow/python/numpy_to_arrow.cc b/cpp/src/arrow/python/numpy_to_arrow.cc index aada6bf598..a944b80914 100644 --- a/cpp/src/arrow/python/numpy_to_arrow.cc +++ b/cpp/src/arrow/python/numpy_to_arrow.cc @@ -443,8 +443,8 @@ inline Status NumPyConverter::ConvertData(std::shared_ptr* data) { RETURN_NOT_OK(NumPyDtypeToArrow(reinterpret_cast(dtype_), _type)); if (!input_type->Equals(*type_)) { -RETURN_NOT_OK(CastBuffer(input_type, *data, length_, nullptr, 0, type_, cast_options_, - pool_, data)); +RETURN_NOT_OK(CastBuffer(input_type, *data, length_, null_bitmap_, null_count_, type_, + cast_options_, pool_, data)); } return Status::OK(); @@ -477,8 +477,8 @@ inline Status NumPyConverter::ConvertData(std::shared_ptr* d } else { RETURN_NOT_OK(NumPyDtypeToArrow(reinterpret_cast(dtype_), _type)); if (!input_type->Equals(*type_)) { - RETURN_NOT_OK(CastBuffer(input_type, *data, length_, nullptr, 0, type_, - cast_options_, pool_, data)); + RETURN_NOT_OK(CastBuffer(input_type, *data, length_, null_bitmap_, null_count_, + type_, cast_options_, pool_, data)); } } @@ -518,8 +518,8 @@ inline Status NumPyConverter::ConvertData(std::shared_ptr* d } else { RETURN_NOT_OK(NumPyDtypeToArrow(reinterpret_cast(dtype_), _type)); if (!input_type->Equals(*type_)) { - RETURN_NOT_OK(CastBuffer(input_type, *data, length_, nullptr, 0, type_, - cast_options_, pool_, data)); + RETURN_NOT_OK(CastBuffer(input_type, *data, length_, null_bitmap_, null_count_, + type_, cast_options_, pool_, data)); } } diff --git a/python/pyarrow/tests/test_convert_pandas.py b/python/pyarrow/tests/test_convert_pandas.py index 264b51c32f..9bee9053c5 100644 --- a/python/pyarrow/tests/test_convert_pandas.py +++ b/python/pyarrow/tests/test_convert_pandas.py @@ -2224,6 +2224,15 @@ def test_safe_unsafe_casts(self): assert table.column('B').type == pa.int32() +def test_safe_cast_from_float_with_nans_to_int(): +# TODO(kszucs): write tests for creating Date32 and Date64 arrays, see +# ARROW-4258 and https://github.com/apache/arrow/pull/3395 +values = pd.Series([1, 2, None, 4]) +arr = pa.Array.from_pandas(values, type=pa.int32(), safe=True) +expected = pa.array([1, 2, None, 4], type=pa.int32()) +assert arr.equals(expected) + + def _fully_loaded_dataframe_example(): index = pd.MultiIndex.from_arrays([ pd.date_range('2000-01-01', periods=5).repeat(2), diff --git a/python/pyarrow/tests/test_plasma_tf_op.py b/python/pyarrow/tests/test_plasma_tf_op.py index e239055209..53ecae217e 100644 --- a/python/pyarrow/tests/test_plasma_tf_op.py +++ b/python/pyarrow/tests/test_plasma_tf_op.py @@ -82,6 +82,7 @@ def FromPlasma(): @pytest.mark.plasma @pytest.mark.tensorflow +@pytest.mark.skip(reason='Until ARROW-4259 is resolved') def test_plasma_tf_op(use_gpu=False): import pyarrow.plasma as plasma import tensorflow as tf With regards, Apache Git Services
[arrow] branch master updated: ARROW-4258: [Python] Safe cast fails from numpy float64 array with nans to integer
This is an automated email from the ASF dual-hosted git repository. wesm pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/arrow.git The following commit(s) were added to refs/heads/master by this push: new 18c0e82 ARROW-4258: [Python] Safe cast fails from numpy float64 array with nans to integer 18c0e82 is described below commit 18c0e8241b5af5b98d7238e64753e44e43a598a8 Author: Krisztián Szűcs AuthorDate: Tue Jan 15 11:42:36 2019 -0600 ARROW-4258: [Python] Safe cast fails from numpy float64 array with nans to integer I'll write more tests to cover all three occurrences... Author: Krisztián Szűcs Closes #3395 from kszucs/ARROW-4258 and squashes the following commits: 85129b41f skip test_plasma_tf_op 6452792f2 lint 76300d1d0 fix --- cpp/src/arrow/python/numpy_to_arrow.cc | 12 ++-- python/pyarrow/tests/test_convert_pandas.py | 9 + python/pyarrow/tests/test_plasma_tf_op.py | 1 + 3 files changed, 16 insertions(+), 6 deletions(-) diff --git a/cpp/src/arrow/python/numpy_to_arrow.cc b/cpp/src/arrow/python/numpy_to_arrow.cc index aada6bf..a944b80 100644 --- a/cpp/src/arrow/python/numpy_to_arrow.cc +++ b/cpp/src/arrow/python/numpy_to_arrow.cc @@ -443,8 +443,8 @@ inline Status NumPyConverter::ConvertData(std::shared_ptr* data) { RETURN_NOT_OK(NumPyDtypeToArrow(reinterpret_cast(dtype_), _type)); if (!input_type->Equals(*type_)) { -RETURN_NOT_OK(CastBuffer(input_type, *data, length_, nullptr, 0, type_, cast_options_, - pool_, data)); +RETURN_NOT_OK(CastBuffer(input_type, *data, length_, null_bitmap_, null_count_, type_, + cast_options_, pool_, data)); } return Status::OK(); @@ -477,8 +477,8 @@ inline Status NumPyConverter::ConvertData(std::shared_ptr* d } else { RETURN_NOT_OK(NumPyDtypeToArrow(reinterpret_cast(dtype_), _type)); if (!input_type->Equals(*type_)) { - RETURN_NOT_OK(CastBuffer(input_type, *data, length_, nullptr, 0, type_, - cast_options_, pool_, data)); + RETURN_NOT_OK(CastBuffer(input_type, *data, length_, null_bitmap_, null_count_, + type_, cast_options_, pool_, data)); } } @@ -518,8 +518,8 @@ inline Status NumPyConverter::ConvertData(std::shared_ptr* d } else { RETURN_NOT_OK(NumPyDtypeToArrow(reinterpret_cast(dtype_), _type)); if (!input_type->Equals(*type_)) { - RETURN_NOT_OK(CastBuffer(input_type, *data, length_, nullptr, 0, type_, - cast_options_, pool_, data)); + RETURN_NOT_OK(CastBuffer(input_type, *data, length_, null_bitmap_, null_count_, + type_, cast_options_, pool_, data)); } } diff --git a/python/pyarrow/tests/test_convert_pandas.py b/python/pyarrow/tests/test_convert_pandas.py index 264b51c..9bee905 100644 --- a/python/pyarrow/tests/test_convert_pandas.py +++ b/python/pyarrow/tests/test_convert_pandas.py @@ -2224,6 +2224,15 @@ class TestConvertMisc(object): assert table.column('B').type == pa.int32() +def test_safe_cast_from_float_with_nans_to_int(): +# TODO(kszucs): write tests for creating Date32 and Date64 arrays, see +# ARROW-4258 and https://github.com/apache/arrow/pull/3395 +values = pd.Series([1, 2, None, 4]) +arr = pa.Array.from_pandas(values, type=pa.int32(), safe=True) +expected = pa.array([1, 2, None, 4], type=pa.int32()) +assert arr.equals(expected) + + def _fully_loaded_dataframe_example(): index = pd.MultiIndex.from_arrays([ pd.date_range('2000-01-01', periods=5).repeat(2), diff --git a/python/pyarrow/tests/test_plasma_tf_op.py b/python/pyarrow/tests/test_plasma_tf_op.py index e239055..53ecae2 100644 --- a/python/pyarrow/tests/test_plasma_tf_op.py +++ b/python/pyarrow/tests/test_plasma_tf_op.py @@ -82,6 +82,7 @@ def run_tensorflow_test_with_dtype(tf, plasma, plasma_store_name, @pytest.mark.plasma @pytest.mark.tensorflow +@pytest.mark.skip(reason='Until ARROW-4259 is resolved') def test_plasma_tf_op(use_gpu=False): import pyarrow.plasma as plasma import tensorflow as tf
[arrow] Diff for: [GitHub] kszucs merged pull request #3406: [CI] Temporary fix for conda-forge migration
diff --git a/ci/appveyor-cpp-build.bat b/ci/appveyor-cpp-build.bat index 78f5e41928..5b653a9050 100644 --- a/ci/appveyor-cpp-build.bat +++ b/ci/appveyor-cpp-build.bat @@ -95,7 +95,7 @@ if "%JOB%" == "Build_Debug" ( exit /B 0 ) -conda create -n arrow -q -y -c conda-forge ^ +conda create -n arrow -q -y -c conda-forge/label/cf201901 ^ --file=ci\conda_env_python.yml ^ python=%PYTHON% ^ numpy=1.14 ^ @@ -110,7 +110,7 @@ set BOOST_LIBRARYDIR=%CONDA_PREFIX%\Library\lib if "%JOB%" == "Toolchain" ( @rem Install pre-built "toolchain" packages for faster builds - conda install -q -y -c conda-forge ^ + conda install -q -y -c conda-forge/label/cf201901 ^ --file=ci\conda_env_cpp.yml ^ python=%PYTHON% diff --git a/ci/appveyor-cpp-setup.bat b/ci/appveyor-cpp-setup.bat index 4cae2cb1ea..2a46d130ea 100644 --- a/ci/appveyor-cpp-setup.bat +++ b/ci/appveyor-cpp-setup.bat @@ -39,7 +39,7 @@ if defined need_vcvarsall ( ) ) -if "%GENERATOR%"=="Ninja" conda install -y -q ninja +if "%GENERATOR%"=="Ninja" conda install -y -q ninja -c conda-forge/label/cf201901 if "%USE_CLCACHE%" == "true" ( @rem Use clcache for faster builds diff --git a/ci/appveyor-cpp-test-cmake-script.bat b/ci/appveyor-cpp-test-cmake-script.bat index 415406c4ac..d55a24dccf 100644 --- a/ci/appveyor-cpp-test-cmake-script.bat +++ b/ci/appveyor-cpp-test-cmake-script.bat @@ -19,8 +19,8 @@ @rem Validate cmake script behaviour on missed lib in toolchain set CONDA_ENV=arrow-cmake-tests-libs -conda create -n %CONDA_ENV% -q -y -conda install -n %CONDA_ENV% -q -y -c conda-forge boost-cpp +conda create -n %CONDA_ENV% -q -y -c conda-forge/label/cf201901 +conda install -n %CONDA_ENV% -q -y -c conda-forge/label/cf201901 boost-cpp call activate %CONDA_ENV% set BUILD_DIR=cpp\build-cmake-test @@ -161,8 +161,8 @@ call deactivate @rem Validate libs availability in conda toolchain set CONDA_ENV=arrow-cmake-tests-toolchain -conda create -n %CONDA_ENV% -q -y -conda install -n %CONDA_ENV% -q -y -c conda-forge ^ +conda create -n %CONDA_ENV% -q -y -c conda-forge/label/cf201901 +conda install -n %CONDA_ENV% -q -y -c conda-forge/label/cf201901 ^ --file=ci\conda_env_cpp.yml call activate %CONDA_ENV% diff --git a/ci/cpp-msvc-build-main.bat b/ci/cpp-msvc-build-main.bat index d9d7e548dd..6629060a47 100644 --- a/ci/cpp-msvc-build-main.bat +++ b/ci/cpp-msvc-build-main.bat @@ -105,7 +105,7 @@ popd call deactivate -conda create -n wheel_test -q -y python=%PYTHON% || exit /B +conda create -n wheel_test -q -y -c conda-forge/label/cf201901 python=%PYTHON% || exit /B call activate wheel_test diff --git a/ci/travis_before_script_c_glib.sh b/ci/travis_before_script_c_glib.sh index e8dd0cdc80..8507f779b3 100755 --- a/ci/travis_before_script_c_glib.sh +++ b/ci/travis_before_script_c_glib.sh @@ -23,7 +23,7 @@ source $TRAVIS_BUILD_DIR/ci/travis_env_common.sh source $TRAVIS_BUILD_DIR/ci/travis_install_conda.sh -conda create -n meson -y -q python=3.6 +conda create -n meson -y -q -c conda-forge/label/cf201901 python=3.6 conda activate meson pip install meson @@ -36,7 +36,7 @@ else autoconf-archive \ gtk-doc-tools \ libgirepository1.0-dev - conda install -q -y ninja + conda install -q -y ninja -c conda-forge/label/cf201901 fi gem install test-unit gobject-introspection diff --git a/ci/travis_install_toolchain.sh b/ci/travis_install_toolchain.sh index 82031e8fd3..aa50d5966d 100755 --- a/ci/travis_install_toolchain.sh +++ b/ci/travis_install_toolchain.sh @@ -27,7 +27,7 @@ if [ ! -e $CPP_TOOLCHAIN ]; then fi # Set up C++ toolchain from conda-forge packages for faster builds -conda create -y -q -p $CPP_TOOLCHAIN \ +conda create -y -q -p $CPP_TOOLCHAIN -c conda-forge/label/cf201901 \ --file=$TRAVIS_BUILD_DIR/ci/conda_env_cpp.yml \ ${CONDA_LLVM} \ ccache \ diff --git a/ci/travis_script_integration.sh b/ci/travis_script_integration.sh index 342db58b5d..5571ebc0b0 100755 --- a/ci/travis_script_integration.sh +++ b/ci/travis_script_integration.sh @@ -43,14 +43,14 @@ popd pushd $ARROW_INTEGRATION_DIR CONDA_ENV_NAME=arrow-integration-test -conda create -y -q -n $CONDA_ENV_NAME python=3.5 +conda create -y -q -n $CONDA_ENV_NAME -c conda-forge/label/cf201901 python=3.5 conda activate $CONDA_ENV_NAME # faster builds, please -conda install -y nomkl +conda install -y nomkl -c conda-forge/label/cf201901 # Expensive dependencies install from Continuum package repo -conda install -y pip numpy six +conda install -y pip numpy six -c conda-forge/label/cf201901 # ARROW-4008: Create a directory to write temporary files since /tmp can be # unstable in Travis CI diff --git a/ci/travis_script_manylinux.sh b/ci/travis_script_manylinux.sh index 588d0f9a7b..59b818af69 100755 --- a/ci/travis_script_manylinux.sh +++ b/ci/travis_script_manylinux.sh @@ -34,7 +34,7 @@ source $TRAVIS_BUILD_DIR/ci/travis_install_conda.sh PYTHON_VERSION=3.6
[arrow] branch master updated: [CI] Temporary fix for conda-forge migration (#3406)
This is an automated email from the ASF dual-hosted git repository. kszucs pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/arrow.git The following commit(s) were added to refs/heads/master by this push: new 143558e [CI] Temporary fix for conda-forge migration (#3406) 143558e is described below commit 143558e1726d6468f4d624ad53d707f9df8a02a5 Author: Krisztián Szűcs AuthorDate: Tue Jan 15 22:00:20 2019 +0100 [CI] Temporary fix for conda-forge migration (#3406) * pin conda forge channel * failed building wheel for pickle5 * verbose and optional pickle5 install --- ci/appveyor-cpp-build.bat | 4 ++-- ci/appveyor-cpp-setup.bat | 2 +- ci/appveyor-cpp-test-cmake-script.bat | 8 ci/cpp-msvc-build-main.bat| 2 +- ci/travis_before_script_c_glib.sh | 4 ++-- ci/travis_install_toolchain.sh| 2 +- ci/travis_script_integration.sh | 6 +++--- ci/travis_script_manylinux.sh | 2 +- ci/travis_script_python.sh| 11 ++- 9 files changed, 21 insertions(+), 20 deletions(-) diff --git a/ci/appveyor-cpp-build.bat b/ci/appveyor-cpp-build.bat index 78f5e41..5b653a9 100644 --- a/ci/appveyor-cpp-build.bat +++ b/ci/appveyor-cpp-build.bat @@ -95,7 +95,7 @@ if "%JOB%" == "Build_Debug" ( exit /B 0 ) -conda create -n arrow -q -y -c conda-forge ^ +conda create -n arrow -q -y -c conda-forge/label/cf201901 ^ --file=ci\conda_env_python.yml ^ python=%PYTHON% ^ numpy=1.14 ^ @@ -110,7 +110,7 @@ set BOOST_LIBRARYDIR=%CONDA_PREFIX%\Library\lib if "%JOB%" == "Toolchain" ( @rem Install pre-built "toolchain" packages for faster builds - conda install -q -y -c conda-forge ^ + conda install -q -y -c conda-forge/label/cf201901 ^ --file=ci\conda_env_cpp.yml ^ python=%PYTHON% diff --git a/ci/appveyor-cpp-setup.bat b/ci/appveyor-cpp-setup.bat index 4cae2cb..2a46d13 100644 --- a/ci/appveyor-cpp-setup.bat +++ b/ci/appveyor-cpp-setup.bat @@ -39,7 +39,7 @@ if defined need_vcvarsall ( ) ) -if "%GENERATOR%"=="Ninja" conda install -y -q ninja +if "%GENERATOR%"=="Ninja" conda install -y -q ninja -c conda-forge/label/cf201901 if "%USE_CLCACHE%" == "true" ( @rem Use clcache for faster builds diff --git a/ci/appveyor-cpp-test-cmake-script.bat b/ci/appveyor-cpp-test-cmake-script.bat index 415406c..d55a24d 100644 --- a/ci/appveyor-cpp-test-cmake-script.bat +++ b/ci/appveyor-cpp-test-cmake-script.bat @@ -19,8 +19,8 @@ @rem Validate cmake script behaviour on missed lib in toolchain set CONDA_ENV=arrow-cmake-tests-libs -conda create -n %CONDA_ENV% -q -y -conda install -n %CONDA_ENV% -q -y -c conda-forge boost-cpp +conda create -n %CONDA_ENV% -q -y -c conda-forge/label/cf201901 +conda install -n %CONDA_ENV% -q -y -c conda-forge/label/cf201901 boost-cpp call activate %CONDA_ENV% set BUILD_DIR=cpp\build-cmake-test @@ -161,8 +161,8 @@ call deactivate @rem Validate libs availability in conda toolchain set CONDA_ENV=arrow-cmake-tests-toolchain -conda create -n %CONDA_ENV% -q -y -conda install -n %CONDA_ENV% -q -y -c conda-forge ^ +conda create -n %CONDA_ENV% -q -y -c conda-forge/label/cf201901 +conda install -n %CONDA_ENV% -q -y -c conda-forge/label/cf201901 ^ --file=ci\conda_env_cpp.yml call activate %CONDA_ENV% diff --git a/ci/cpp-msvc-build-main.bat b/ci/cpp-msvc-build-main.bat index d9d7e54..6629060 100644 --- a/ci/cpp-msvc-build-main.bat +++ b/ci/cpp-msvc-build-main.bat @@ -105,7 +105,7 @@ popd call deactivate -conda create -n wheel_test -q -y python=%PYTHON% || exit /B +conda create -n wheel_test -q -y -c conda-forge/label/cf201901 python=%PYTHON% || exit /B call activate wheel_test diff --git a/ci/travis_before_script_c_glib.sh b/ci/travis_before_script_c_glib.sh index e8dd0cd..8507f77 100755 --- a/ci/travis_before_script_c_glib.sh +++ b/ci/travis_before_script_c_glib.sh @@ -23,7 +23,7 @@ source $TRAVIS_BUILD_DIR/ci/travis_env_common.sh source $TRAVIS_BUILD_DIR/ci/travis_install_conda.sh -conda create -n meson -y -q python=3.6 +conda create -n meson -y -q -c conda-forge/label/cf201901 python=3.6 conda activate meson pip install meson @@ -36,7 +36,7 @@ else autoconf-archive \ gtk-doc-tools \ libgirepository1.0-dev - conda install -q -y ninja + conda install -q -y ninja -c conda-forge/label/cf201901 fi gem install test-unit gobject-introspection diff --git a/ci/travis_install_toolchain.sh b/ci/travis_install_toolchain.sh index 82031e8..aa50d59 100755 --- a/ci/travis_install_toolchain.sh +++ b/ci/travis_install_toolchain.sh @@ -27,7 +27,7 @@ if [ ! -e $CPP_TOOLCHAIN ]; then fi # Set up C++ toolchain from conda-forge packages for faster builds -conda create -y -q -p $CPP_TOOLCHAIN \ +conda create -y -q -p $CPP_TOOLCHAIN -c conda-forge/label/cf201901 \ --file=$TRAVIS_BUILD_DIR/ci/conda_env_cpp.yml \ ${CONDA_LLVM} \
svn commit: r31977 - in /dev/arrow/apache-arrow-0.12.0-rc3: ./ apache-arrow-0.12.0.tar.gz apache-arrow-0.12.0.tar.gz.asc apache-arrow-0.12.0.tar.gz.sha256 apache-arrow-0.12.0.tar.gz.sha512
Author: kszucs Date: Tue Jan 15 21:43:51 2019 New Revision: 31977 Log: Apache Arrow 0.12.0 RC3 Added: dev/arrow/apache-arrow-0.12.0-rc3/ dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz (with props) dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz.asc dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz.sha256 dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz.sha512 Added: dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz == Binary file - no diff available. Propchange: dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz -- svn:mime-type = application/octet-stream Added: dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz.asc == --- dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz.asc (added) +++ dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz.asc Tue Jan 15 21:43:51 2019 @@ -0,0 +1,16 @@ +-BEGIN PGP SIGNATURE- + +iQIzBAABCAAdFiEEJl+Aq4T+AxJ+FPARJbzKUiDYQHkFAlw+VAQACgkQJbzKUiDY +QHlOCBAA7gRqX2+oucYiC2QHKtCGf7+YjQ3bHTCfvVH9GEwgYbMztQSajzK1gro3 +Dkbk3W+6fc5zVDvSxA6c8poEajeG3338ySMisMiAcQXR+SNsxYvWNOnxpmJ5YIgS +qEHB+1CgeII++rda65pIlXmrNgfL3hPCezAbrTDntS21b8SzSpuU2GnYxXnp02L9 +Q9WNUFvyANunJ64PEl/dsEGDgXH7AoZwYBOPijtLfiuEQqL7/35Vq+XLzibKwrrq ++FQxZmMAL6J0J4jsdPUJt2YMGOooa0Gp7u1FI15Vx27e1RUWljSEprEbEne/Sucl +fjtl3B/4aeV1wHKlb4EVfjAH9SdIUumI5A6sQF0Ga7KX453Bg0Q2t+RvIwfiSPkd +VQkd5/boOKLNl5mng+Ud/UJBdw7kYxxnDN8ss8rMMcvA0+VhPiqqVDl+DBq4eRk7 +S2MQ4h49a3ueV0eVZ1/yTQOEaXEVKwrZq7S/r4NoPPVL5srY/VL3uo72ndB8+rYt +TJroDZm9Eb2MBtVkYd5KHY1TQUEQICid1k5Q94oZ9dtXJceQ3bvNyCkwGk3WFB+g +LRBnOWEBxdGJC/thHVamcENLLJrZReflqmdpq2ORrGrbyoKF+oj0hf9tUxa+6BGw +ViKUAOmiDwFWxTNvozLjxbqLFzI1fcU5uS6F662WZCchvLwXaZs= +=rSB+ +-END PGP SIGNATURE- Added: dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz.sha256 == --- dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz.sha256 (added) +++ dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz.sha256 Tue Jan 15 21:43:51 2019 @@ -0,0 +1 @@ +8439fa1ef3e430c95c8bbd2c634a45de57f097b86b24c9dfef12c0f48a6f5f92 apache-arrow-0.12.0.tar.gz Added: dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz.sha512 == --- dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz.sha512 (added) +++ dev/arrow/apache-arrow-0.12.0-rc3/apache-arrow-0.12.0.tar.gz.sha512 Tue Jan 15 21:43:51 2019 @@ -0,0 +1 @@ +d53ececed34aafb54cff96fba5fe21d093e3c741d38b414269ac15abdead8fc465ac7cd80772a99091db135dc57659ced8ca19c7804018e698ce2f5868fc4174 apache-arrow-0.12.0.tar.gz
[arrow] annotated tag apache-arrow-0.12.0 updated (80ebd7f -> d38e39e)
This is an automated email from the ASF dual-hosted git repository. kszucs pushed a change to annotated tag apache-arrow-0.12.0 in repository https://gitbox.apache.org/repos/asf/arrow.git. *** WARNING: tag apache-arrow-0.12.0 was modified! *** from 80ebd7f (tag) to d38e39e (tag) tagging d7a81b0ad5eee51178f37dbc2c418adc3da6f03d (commit) replaces apache-arrow-0.11.0 by Krisztián Szűcs on Tue Jan 15 22:12:38 2019 +0100 - Log - [maven-release-plugin] copy for tag apache-arrow-0.12.0 --- discard 0bbaef4 [maven-release-plugin] prepare release apache-arrow-0.12.0 discard 9f46847 [Release] Update versions for 0.12.0 discard 71a5d1d [Release] Update .deb/.rpm changelogs for 0.12.0 discard 9ab2b52 [Release] Update CHANGELOG.md for 0.12.0 add be663c1 ARROW-4247: [Release] Update verify script for 0.12.0 add 5598d2f ARROW-2828: [JS] Refactor Data, Vectors, Visitor, Typings, build, tests, dependencies add 85362b3 ARROW-4249: [Plasma] Clean up client namespace add cf047fc ARROW-4256: [Release] Fix Windows verification script for 0.12 release add a78a3d6 ARROW-4228: [GLib] Add garrow_list_data_type_get_field() add 61b50dc ARROW-4257: [Release] Update release verification script to check binaries on Bintray add 6f4d1a9 ARROW-4246: [Plasma][Python] PlasmaClient.list returns wrong information with CUDA enabled Plasma add 87ac6fd ARROW-4246: [Plasma][Python][Follow-up] Ensure plasma::ObjectTableEntry always has the same size regardless of whether built with CUDA support add 09d3496 ARROW-4260: [Python] NumPy buffer protocol failure add 5a7507c ARROW-4266: [Python][CI] Disable ORC tests in dask integration test add 18c0e82 ARROW-4258: [Python] Safe cast fails from numpy float64 array with nans to integer add 143558e [CI] Temporary fix for conda-forge migration (#3406) new 79594d0 [Release] Update CHANGELOG.md for 0.12.0 new 30ad982 [Release] Update .deb/.rpm changelogs for 0.12.0 new ba02353 [Release] Update versions for 0.12.0 new d7a81b0 [maven-release-plugin] prepare release apache-arrow-0.12.0 This update added new revisions after undoing existing revisions. That is to say, some revisions that were in the old version of the annotated tag are not in the new version. This situation occurs when a user --force pushes a change and generates a repository containing something like this: * -- * -- B -- O -- O -- O (80ebd7f) \ N -- N -- N refs/tags/apache-arrow-0.12.0 (d38e39e) You should already have received notification emails for all of the O revisions, and so the following emails describe only the N revisions from the common base, B. Any revisions marked "omit" are not gone; other references still refer to them. Any revisions marked "discard" are gone forever. The 4 revisions listed above as "new" are entirely new to this repository and will be described in separate emails. The revisions listed as "add" were already present in the repository and have only been added to this reference. Summary of changes: .travis.yml|4 +- CHANGELOG.md | 12 +- c_glib/arrow-glib/composite-data-type.cpp | 17 + c_glib/arrow-glib/composite-data-type.h|6 + c_glib/arrow-glib/version.h.in | 23 + c_glib/doc/arrow-glib/arrow-glib-docs.xml |4 + c_glib/test/test-list-data-type.rb |6 +- ci/appveyor-cpp-build.bat | 15 +- ci/appveyor-cpp-setup.bat |2 +- ci/appveyor-cpp-test-cmake-script.bat |8 +- ci/cpp-msvc-build-main.bat |4 +- ci/travis_before_script_c_glib.sh |4 +- ci/travis_install_toolchain.sh |2 +- ci/travis_script_integration.sh|8 +- ci/travis_script_js.sh |5 +- ci/travis_script_manylinux.sh |2 +- ci/travis_script_python.sh | 15 +- cpp/src/arrow/python/numpy_to_arrow.cc | 12 +- cpp/src/plasma/common.h| 18 +- cpp/src/plasma/io.cc |1 + .../org_apache_arrow_plasma_PlasmaClientJNI.cc |2 + dev/release/verify-release-candidate.bat | 11 +- dev/release/verify-release-candidate.sh| 107 +- .../linux-packages/debian.ubuntu-trusty/changelog |2 +- dev/tasks/linux-packages/debian/changelog |2 +- dev/tasks/linux-packages/yum/arrow.spec.in |2 +- integration/dask/runtest.sh|3 +- integration/integration_test.py|2 +- js/.gitignore
[arrow] 03/04: [Release] Update versions for 0.12.0
This is an automated email from the ASF dual-hosted git repository. kszucs pushed a commit to annotated tag apache-arrow-0.12.0 in repository https://gitbox.apache.org/repos/asf/arrow.git commit ba02353d832905cc8a3d57d88be877434b352679 Author: Krisztián Szűcs AuthorDate: Tue Jan 15 22:04:03 2019 +0100 [Release] Update versions for 0.12.0 --- c_glib/configure.ac | 2 +- c_glib/meson.build| 2 +- cpp/CMakeLists.txt| 2 +- matlab/CMakeLists.txt | 2 +- r/DESCRIPTION | 2 +- ruby/red-arrow-cuda/lib/arrow-cuda/version.rb | 2 +- ruby/red-arrow/lib/arrow/version.rb | 2 +- ruby/red-gandiva/lib/gandiva/version.rb | 2 +- ruby/red-parquet/lib/parquet/version.rb | 2 +- ruby/red-plasma/lib/plasma/version.rb | 2 +- rust/arrow/Cargo.toml | 2 +- rust/parquet/Cargo.toml | 2 +- 12 files changed, 12 insertions(+), 12 deletions(-) diff --git a/c_glib/configure.ac b/c_glib/configure.ac index c63bfff..685b8c1 100644 --- a/c_glib/configure.ac +++ b/c_glib/configure.ac @@ -17,7 +17,7 @@ AC_PREREQ(2.65) -m4_define([arrow_glib_version], 0.12.0-SNAPSHOT) +m4_define([arrow_glib_version], 0.12.0) AC_INIT([arrow-glib], arrow_glib_version, [https://issues.apache.org/jira/browse/ARROW], diff --git a/c_glib/meson.build b/c_glib/meson.build index c2cf36c..4f8e59a 100644 --- a/c_glib/meson.build +++ b/c_glib/meson.build @@ -23,7 +23,7 @@ project('arrow-glib', 'c', 'cpp', 'cpp_std=c++11', ]) -version = '0.12.0-SNAPSHOT' +version = '0.12.0' if version.endswith('-SNAPSHOT') version_numbers = version.split('-')[0].split('.') version_tag = version.split('-')[1] diff --git a/cpp/CMakeLists.txt b/cpp/CMakeLists.txt index 08868af..8c30816 100644 --- a/cpp/CMakeLists.txt +++ b/cpp/CMakeLists.txt @@ -18,7 +18,7 @@ cmake_minimum_required(VERSION 3.2) message(STATUS "Building using CMake version: ${CMAKE_VERSION}") -set(ARROW_VERSION "0.12.0-SNAPSHOT") +set(ARROW_VERSION "0.12.0") string(REGEX MATCH "^[0-9]+\\.[0-9]+\\.[0-9]+" ARROW_BASE_VERSION "${ARROW_VERSION}") diff --git a/matlab/CMakeLists.txt b/matlab/CMakeLists.txt index a6371d1..52673d2 100755 --- a/matlab/CMakeLists.txt +++ b/matlab/CMakeLists.txt @@ -18,7 +18,7 @@ cmake_minimum_required(VERSION 3.2) set(CMAKE_CXX_STANDARD 11) -set(MLARROW_VERSION "0.12.0-SNAPSHOT") +set(MLARROW_VERSION "0.12.0") string(REGEX MATCH "^[0-9]+\\.[0-9]+\\.[0-9]+" MLARROW_BASE_VERSION "${MLARROW_VERSION}") diff --git a/r/DESCRIPTION b/r/DESCRIPTION index 5303a87..768107c 100644 --- a/r/DESCRIPTION +++ b/r/DESCRIPTION @@ -1,6 +1,6 @@ Package: arrow Title: R Integration to 'Apache' 'Arrow' -Version: 0.11.0.9000 +Version: 0.12.0 Authors@R: c( person("Romain", "François", email = "rom...@rstudio.com", role = c("aut", "cre")), person("Javier", "Luraschi", email = "jav...@rstudio.com", role = c("ctb")), diff --git a/ruby/red-arrow-cuda/lib/arrow-cuda/version.rb b/ruby/red-arrow-cuda/lib/arrow-cuda/version.rb index 6426d2d..266254f 100644 --- a/ruby/red-arrow-cuda/lib/arrow-cuda/version.rb +++ b/ruby/red-arrow-cuda/lib/arrow-cuda/version.rb @@ -16,7 +16,7 @@ # under the License. module ArrowCUDA - VERSION = "0.12.0-SNAPSHOT" + VERSION = "0.12.0" module Version numbers, TAG = VERSION.split("-") diff --git a/ruby/red-arrow/lib/arrow/version.rb b/ruby/red-arrow/lib/arrow/version.rb index 8ff0779..363d07d 100644 --- a/ruby/red-arrow/lib/arrow/version.rb +++ b/ruby/red-arrow/lib/arrow/version.rb @@ -16,7 +16,7 @@ # under the License. module Arrow - VERSION = "0.12.0-SNAPSHOT" + VERSION = "0.12.0" module Version numbers, TAG = VERSION.split("-") diff --git a/ruby/red-gandiva/lib/gandiva/version.rb b/ruby/red-gandiva/lib/gandiva/version.rb index dbdaf36..808234f 100644 --- a/ruby/red-gandiva/lib/gandiva/version.rb +++ b/ruby/red-gandiva/lib/gandiva/version.rb @@ -16,7 +16,7 @@ # under the License. module Gandiva - VERSION = "0.12.0-SNAPSHOT" + VERSION = "0.12.0" module Version numbers, TAG = VERSION.split("-") diff --git a/ruby/red-parquet/lib/parquet/version.rb b/ruby/red-parquet/lib/parquet/version.rb index 997a92e..5798240 100644 --- a/ruby/red-parquet/lib/parquet/version.rb +++ b/ruby/red-parquet/lib/parquet/version.rb @@ -16,7 +16,7 @@ # under the License. module Parquet - VERSION = "0.12.0-SNAPSHOT" + VERSION = "0.12.0" module Version numbers, TAG = VERSION.split("-") diff --git a/ruby/red-plasma/lib/plasma/version.rb b/ruby/red-plasma/lib/plasma/version.rb index e88f2de..1ac590f 100644 --- a/ruby/red-plasma/lib/plasma/version.rb +++ b/ruby/red-plasma/lib/plasma/version.rb @@ -16,7 +16,7 @@ # under the License. module Plasma - VERSION = "0.12.0-SNAPSHOT" + VERSION = "0.12.0" module Version numbers, TAG = VERSION.split("-")
[arrow] 02/04: [Release] Update .deb/.rpm changelogs for 0.12.0
This is an automated email from the ASF dual-hosted git repository. kszucs pushed a commit to annotated tag apache-arrow-0.12.0 in repository https://gitbox.apache.org/repos/asf/arrow.git commit 30ad982f003d4b6c27f6eb1eaf9edaf1ce326844 Author: Krisztián Szűcs AuthorDate: Tue Jan 15 22:04:01 2019 +0100 [Release] Update .deb/.rpm changelogs for 0.12.0 --- dev/tasks/linux-packages/debian.ubuntu-trusty/changelog | 6 ++ dev/tasks/linux-packages/debian/changelog | 6 ++ dev/tasks/linux-packages/yum/arrow.spec.in | 3 +++ 3 files changed, 15 insertions(+) diff --git a/dev/tasks/linux-packages/debian.ubuntu-trusty/changelog b/dev/tasks/linux-packages/debian.ubuntu-trusty/changelog index e54e05c..d53004e 100644 --- a/dev/tasks/linux-packages/debian.ubuntu-trusty/changelog +++ b/dev/tasks/linux-packages/debian.ubuntu-trusty/changelog @@ -1,3 +1,9 @@ +apache-arrow (0.12.0-1) unstable; urgency=low + + * New upstream release. + + -- Krisztián Szűcs Tue, 15 Jan 2019 21:04:01 - + apache-arrow (0.11.0-1) unstable; urgency=low * New upstream release. diff --git a/dev/tasks/linux-packages/debian/changelog b/dev/tasks/linux-packages/debian/changelog index a0aff6f..1c8fedd 100644 --- a/dev/tasks/linux-packages/debian/changelog +++ b/dev/tasks/linux-packages/debian/changelog @@ -1,3 +1,9 @@ +apache-arrow (0.12.0-1) unstable; urgency=low + + * New upstream release. + + -- Krisztián Szűcs Tue, 15 Jan 2019 21:04:01 - + apache-arrow (0.11.0-1) unstable; urgency=low * New upstream release. diff --git a/dev/tasks/linux-packages/yum/arrow.spec.in b/dev/tasks/linux-packages/yum/arrow.spec.in index 568477e..db95be3 100644 --- a/dev/tasks/linux-packages/yum/arrow.spec.in +++ b/dev/tasks/linux-packages/yum/arrow.spec.in @@ -397,6 +397,9 @@ Documentation for Apache Parquet GLib. %endif %changelog +* Tue Jan 15 2019 Krisztián Szűcs - 0.12.0-1 +- New upstream release. + * Thu Oct 04 2018 Kouhei Sutou - 0.11.0-1 - New upstream release.
[arrow] 04/04: [maven-release-plugin] prepare release apache-arrow-0.12.0
This is an automated email from the ASF dual-hosted git repository. kszucs pushed a commit to annotated tag apache-arrow-0.12.0 in repository https://gitbox.apache.org/repos/asf/arrow.git commit d7a81b0ad5eee51178f37dbc2c418adc3da6f03d Author: Krisztián Szűcs AuthorDate: Tue Jan 15 22:12:27 2019 +0100 [maven-release-plugin] prepare release apache-arrow-0.12.0 --- java/adapter/jdbc/pom.xml | 2 +- java/flight/pom.xml | 2 +- java/format/pom.xml | 2 +- java/memory/pom.xml | 2 +- java/plasma/pom.xml | 2 +- java/pom.xml | 4 ++-- java/tools/pom.xml| 2 +- java/vector/pom.xml | 2 +- 8 files changed, 9 insertions(+), 9 deletions(-) diff --git a/java/adapter/jdbc/pom.xml b/java/adapter/jdbc/pom.xml index ab0b9b5..add4e07 100644 --- a/java/adapter/jdbc/pom.xml +++ b/java/adapter/jdbc/pom.xml @@ -16,7 +16,7 @@ org.apache.arrow arrow-java-root -0.12.0-SNAPSHOT +0.12.0 ../../pom.xml diff --git a/java/flight/pom.xml b/java/flight/pom.xml index c6de29f..f99194e 100644 --- a/java/flight/pom.xml +++ b/java/flight/pom.xml @@ -11,7 +11,7 @@ org.apache.arrow arrow-java-root -0.12.0-SNAPSHOT +0.12.0 arrow-flight diff --git a/java/format/pom.xml b/java/format/pom.xml index d5ccd5f..7427824 100644 --- a/java/format/pom.xml +++ b/java/format/pom.xml @@ -15,7 +15,7 @@ arrow-java-root org.apache.arrow - 0.12.0-SNAPSHOT + 0.12.0 arrow-format diff --git a/java/memory/pom.xml b/java/memory/pom.xml index 7e2c42d..8f190e5 100644 --- a/java/memory/pom.xml +++ b/java/memory/pom.xml @@ -14,7 +14,7 @@ org.apache.arrow arrow-java-root -0.12.0-SNAPSHOT +0.12.0 arrow-memory Arrow Memory diff --git a/java/plasma/pom.xml b/java/plasma/pom.xml index d50171a..a431ef2 100644 --- a/java/plasma/pom.xml +++ b/java/plasma/pom.xml @@ -14,7 +14,7 @@ org.apache.arrow arrow-java-root -0.12.0-SNAPSHOT +0.12.0 arrow-plasma Arrow Plasma Client diff --git a/java/pom.xml b/java/pom.xml index 0df1178..47b62cc 100644 --- a/java/pom.xml +++ b/java/pom.xml @@ -20,7 +20,7 @@ org.apache.arrow arrow-java-root - 0.12.0-SNAPSHOT + 0.12.0 pom Apache Arrow Java Root POM @@ -45,7 +45,7 @@ scm:git:https://github.com/apache/arrow.git scm:git:https://github.com/apache/arrow.git https://github.com/apache/arrow -apache-arrow-0.10.0 +apache-arrow-0.12.0 diff --git a/java/tools/pom.xml b/java/tools/pom.xml index d01c951..9af5d4d 100644 --- a/java/tools/pom.xml +++ b/java/tools/pom.xml @@ -14,7 +14,7 @@ org.apache.arrow arrow-java-root -0.12.0-SNAPSHOT +0.12.0 arrow-tools Arrow Tools diff --git a/java/vector/pom.xml b/java/vector/pom.xml index fc4eb8f..4daa91e 100644 --- a/java/vector/pom.xml +++ b/java/vector/pom.xml @@ -14,7 +14,7 @@ org.apache.arrow arrow-java-root -0.12.0-SNAPSHOT +0.12.0 arrow-vector Arrow Vectors
[arrow] 01/04: [Release] Update CHANGELOG.md for 0.12.0
This is an automated email from the ASF dual-hosted git repository. kszucs pushed a commit to annotated tag apache-arrow-0.12.0 in repository https://gitbox.apache.org/repos/asf/arrow.git commit 79594d03286f94f016465a37c9d3bb23466a4050 Author: Krisztián Szűcs AuthorDate: Tue Jan 15 22:03:59 2019 +0100 [Release] Update CHANGELOG.md for 0.12.0 --- CHANGELOG.md | 603 +++ 1 file changed, 603 insertions(+) diff --git a/CHANGELOG.md b/CHANGELOG.md index 853806c..fa136aa 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -17,6 +17,608 @@ under the License. --> +# Apache Arrow 0.12.0 (15 January 2019) + +## Bug + +* ARROW-1847 - [Doc] Document the difference between RecordBatch and Table in an FAQ fashion +* ARROW-1994 - [Python] Test against Pandas master +* ARROW-2026 - [Python] Cast all timestamp resolutions to INT96 use\_deprecated\_int96\_timestamps=True +* ARROW-2038 - [Python] Follow-up bug fixes for s3fs Parquet support +* ARROW-2113 - [Python] Incomplete CLASSPATH with "hadoop" contained in it can fool the classpath setting HDFS logic +* ARROW-2591 - [Python] Segmentation fault when writing empty ListType column to Parquet +* ARROW-2592 - [Python] Error reading old Parquet file due to metadata backwards compatibility issue +* ARROW-2708 - [C++] Internal GetValues function in arrow::compute should check for nullptr +* ARROW-2970 - [Python] NumPyConverter::Visit for Binary/String/FixedSizeBinary can overflow +* ARROW-3058 - [Python] Feather reads fail with unintuitive error when conversion from pandas yields ChunkedArray +* ARROW-3186 - [GLib] mesonbuild failures in Travis CI +* ARROW-3202 - [C++] Build does not succeed on Alpine Linux +* ARROW-3225 - [C++/Python] Pandas object conversion of ListType and ListType +* ARROW-3324 - [Parquet] Free more internal resources when writing multiple row groups +* ARROW-3343 - [Java] Java tests fail non-deterministically with memory leak from Flight tests +* ARROW-3405 - [Python] Document CSV reader +* ARROW-3428 - [Python] from\_pandas gives incorrect results when converting floating point to bool +* ARROW-3436 - [C++] Boost version required by Gandiva is too new for Ubuntu 14.04 +* ARROW-3437 - [Gandiva][C++] Configure static linking of libgcc, libstdc++ with LDFLAGS +* ARROW-3438 - [Packaging] Escaped bulletpoints in changelog +* ARROW-3445 - [GLib] Parquet GLib doesn't link Arrow GLib +* ARROW-3449 - [C++] Support CMake 3.2 for "out of the box" builds +* ARROW-3466 - [Python] Crash when importing tensorflow and pyarrow +* ARROW-3467 - Building against external double conversion is broken +* ARROW-3470 - [C++] Row-wise conversion tutorial has fallen out of date +* ARROW-3477 - [C++] Testsuite fails on 32 bit arch +* ARROW-3480 - [Website] Install document for Ubuntu is broken +* ARROW-3485 - [C++] Examples fail with Protobuf error +* ARROW-3494 - [C++] re2 conda-forge package not working in toolchain +* ARROW-3516 - [C++] Use unsigned type for difference of pointers in parallel\_memcpy +* ARROW-3517 - [C++] MinGW 32bit build causes g++ segv +* ARROW-3524 - [C++] Fix compiler warnings from ARROW-3409 on clang-6 +* ARROW-3527 - [R] Unused variables in R-package C++ code +* ARROW-3528 - [R] Typo in R documentation +* ARROW-3535 - [Python] pip install tensorflow install too new numpy in manylinux1 build +* ARROW-3541 - [Rust] Update BufferBuilder to allow for new bit-packed BooleanArray +* ARROW-3544 - [Gandiva] Populate function registry in multiple compilation units to mitigate long compile times in release mode +* ARROW-3549 - [Rust] Replace i64 with usize for some bit utility functions +* ARROW-3573 - [Rust] with\_bitset does not set valid bits correctly +* ARROW-3580 - [Gandiva][C++] Build error with g++ 8.2.0 +* ARROW-3586 - [Python] Segmentation fault when converting empty table to pandas with categoricals +* ARROW-3598 - [Plasma] plasma\_store\_server fails linking with GPU enabled +* ARROW-3613 - [Go] Resize does not correctly update the length +* ARROW-3614 - [R] Handle Type::TIMESTAMP from Arrow to R +* ARROW-3658 - [Rust] validation of offsets buffer is incorrect for \`List\` +* ARROW-3670 - [C++] Use FindBacktrace to find execinfo.h support +* ARROW-3687 - [Rust] Anything measuring array slots should be \`usize\` +* ARROW-3698 - [C++] Segmentation fault when using a large table in Gandiva +* ARROW-3700 - [C++] CSV parser should allow ignoring empty lines +* ARROW-3703 - [Python] DataFrame.to\_parquet crashes if datetime column has time zones +* ARROW-3707 - [C++] test failure with zstd 1.3.7 +* ARROW-3711 - [C++] Don't pass CXX\_FLAGS to C\_FLAGS +* ARROW-3712 - [CI] License check regression (RAT failure) +* ARROW-3715 - [C++] gflags\_ep fails to build with CMake 3.13 +* ARROW-3716 - [R] Missing cases for ChunkedArray conversion +* ARROW-3728 - [Python] Merging Parquet Files - Pandas Meta in Schema Mismatch +* ARROW-3734 - [C++] Linking static zstd library fails on
[arrow] 03/04: [Release] Update versions for 0.12.0
This is an automated email from the ASF dual-hosted git repository. kszucs pushed a commit to annotated tag apache-arrow-0.12.0 in repository https://gitbox.apache.org/repos/asf/arrow.git commit 6fcd910aa01b1c1923779bff29f162ce40caa98d Author: Krisztián Szűcs AuthorDate: Wed Jan 16 04:29:27 2019 +0100 [Release] Update versions for 0.12.0 --- c_glib/configure.ac | 2 +- c_glib/meson.build| 2 +- cpp/CMakeLists.txt| 2 +- matlab/CMakeLists.txt | 2 +- r/DESCRIPTION | 2 +- ruby/red-arrow-cuda/lib/arrow-cuda/version.rb | 2 +- ruby/red-arrow/lib/arrow/version.rb | 2 +- ruby/red-gandiva/lib/gandiva/version.rb | 2 +- ruby/red-parquet/lib/parquet/version.rb | 2 +- ruby/red-plasma/lib/plasma/version.rb | 2 +- rust/arrow/Cargo.toml | 2 +- rust/parquet/Cargo.toml | 2 +- 12 files changed, 12 insertions(+), 12 deletions(-) diff --git a/c_glib/configure.ac b/c_glib/configure.ac index c63bfff..685b8c1 100644 --- a/c_glib/configure.ac +++ b/c_glib/configure.ac @@ -17,7 +17,7 @@ AC_PREREQ(2.65) -m4_define([arrow_glib_version], 0.12.0-SNAPSHOT) +m4_define([arrow_glib_version], 0.12.0) AC_INIT([arrow-glib], arrow_glib_version, [https://issues.apache.org/jira/browse/ARROW], diff --git a/c_glib/meson.build b/c_glib/meson.build index c2cf36c..4f8e59a 100644 --- a/c_glib/meson.build +++ b/c_glib/meson.build @@ -23,7 +23,7 @@ project('arrow-glib', 'c', 'cpp', 'cpp_std=c++11', ]) -version = '0.12.0-SNAPSHOT' +version = '0.12.0' if version.endswith('-SNAPSHOT') version_numbers = version.split('-')[0].split('.') version_tag = version.split('-')[1] diff --git a/cpp/CMakeLists.txt b/cpp/CMakeLists.txt index 08868af..8c30816 100644 --- a/cpp/CMakeLists.txt +++ b/cpp/CMakeLists.txt @@ -18,7 +18,7 @@ cmake_minimum_required(VERSION 3.2) message(STATUS "Building using CMake version: ${CMAKE_VERSION}") -set(ARROW_VERSION "0.12.0-SNAPSHOT") +set(ARROW_VERSION "0.12.0") string(REGEX MATCH "^[0-9]+\\.[0-9]+\\.[0-9]+" ARROW_BASE_VERSION "${ARROW_VERSION}") diff --git a/matlab/CMakeLists.txt b/matlab/CMakeLists.txt index a6371d1..52673d2 100755 --- a/matlab/CMakeLists.txt +++ b/matlab/CMakeLists.txt @@ -18,7 +18,7 @@ cmake_minimum_required(VERSION 3.2) set(CMAKE_CXX_STANDARD 11) -set(MLARROW_VERSION "0.12.0-SNAPSHOT") +set(MLARROW_VERSION "0.12.0") string(REGEX MATCH "^[0-9]+\\.[0-9]+\\.[0-9]+" MLARROW_BASE_VERSION "${MLARROW_VERSION}") diff --git a/r/DESCRIPTION b/r/DESCRIPTION index 5303a87..768107c 100644 --- a/r/DESCRIPTION +++ b/r/DESCRIPTION @@ -1,6 +1,6 @@ Package: arrow Title: R Integration to 'Apache' 'Arrow' -Version: 0.11.0.9000 +Version: 0.12.0 Authors@R: c( person("Romain", "François", email = "rom...@rstudio.com", role = c("aut", "cre")), person("Javier", "Luraschi", email = "jav...@rstudio.com", role = c("ctb")), diff --git a/ruby/red-arrow-cuda/lib/arrow-cuda/version.rb b/ruby/red-arrow-cuda/lib/arrow-cuda/version.rb index 6426d2d..266254f 100644 --- a/ruby/red-arrow-cuda/lib/arrow-cuda/version.rb +++ b/ruby/red-arrow-cuda/lib/arrow-cuda/version.rb @@ -16,7 +16,7 @@ # under the License. module ArrowCUDA - VERSION = "0.12.0-SNAPSHOT" + VERSION = "0.12.0" module Version numbers, TAG = VERSION.split("-") diff --git a/ruby/red-arrow/lib/arrow/version.rb b/ruby/red-arrow/lib/arrow/version.rb index 8ff0779..363d07d 100644 --- a/ruby/red-arrow/lib/arrow/version.rb +++ b/ruby/red-arrow/lib/arrow/version.rb @@ -16,7 +16,7 @@ # under the License. module Arrow - VERSION = "0.12.0-SNAPSHOT" + VERSION = "0.12.0" module Version numbers, TAG = VERSION.split("-") diff --git a/ruby/red-gandiva/lib/gandiva/version.rb b/ruby/red-gandiva/lib/gandiva/version.rb index dbdaf36..808234f 100644 --- a/ruby/red-gandiva/lib/gandiva/version.rb +++ b/ruby/red-gandiva/lib/gandiva/version.rb @@ -16,7 +16,7 @@ # under the License. module Gandiva - VERSION = "0.12.0-SNAPSHOT" + VERSION = "0.12.0" module Version numbers, TAG = VERSION.split("-") diff --git a/ruby/red-parquet/lib/parquet/version.rb b/ruby/red-parquet/lib/parquet/version.rb index 997a92e..5798240 100644 --- a/ruby/red-parquet/lib/parquet/version.rb +++ b/ruby/red-parquet/lib/parquet/version.rb @@ -16,7 +16,7 @@ # under the License. module Parquet - VERSION = "0.12.0-SNAPSHOT" + VERSION = "0.12.0" module Version numbers, TAG = VERSION.split("-") diff --git a/ruby/red-plasma/lib/plasma/version.rb b/ruby/red-plasma/lib/plasma/version.rb index e88f2de..1ac590f 100644 --- a/ruby/red-plasma/lib/plasma/version.rb +++ b/ruby/red-plasma/lib/plasma/version.rb @@ -16,7 +16,7 @@ # under the License. module Plasma - VERSION = "0.12.0-SNAPSHOT" + VERSION = "0.12.0" module Version numbers, TAG = VERSION.split("-")
[arrow] annotated tag apache-arrow-0.12.0 updated (d38e39e -> b1c3298)
This is an automated email from the ASF dual-hosted git repository. kszucs pushed a change to annotated tag apache-arrow-0.12.0 in repository https://gitbox.apache.org/repos/asf/arrow.git. *** WARNING: tag apache-arrow-0.12.0 was modified! *** from d38e39e (tag) to b1c3298 (tag) tagging 8ca41384b5324bfd0ef3d3ed3f728e1d10ed73f0 (commit) replaces apache-arrow-0.11.0 by Krisztián Szűcs on Wed Jan 16 04:36:10 2019 +0100 - Log - [maven-release-plugin] copy for tag apache-arrow-0.12.0 --- discard d7a81b0 [maven-release-plugin] prepare release apache-arrow-0.12.0 discard ba02353 [Release] Update versions for 0.12.0 discard 30ad982 [Release] Update .deb/.rpm changelogs for 0.12.0 discard 79594d0 [Release] Update CHANGELOG.md for 0.12.0 add fbfd1ff ARROW-4269: [Python] Fix serialization in pandas 0.22 add a1a9221 ARROW-4270: [Packaging][Conda] Update xcode version and remove toolchain builds new 6c8c0c1 [Release] Update CHANGELOG.md for 0.12.0 new db508ee [Release] Update .deb/.rpm changelogs for 0.12.0 new 6fcd910 [Release] Update versions for 0.12.0 new 8ca4138 [maven-release-plugin] prepare release apache-arrow-0.12.0 This update added new revisions after undoing existing revisions. That is to say, some revisions that were in the old version of the annotated tag are not in the new version. This situation occurs when a user --force pushes a change and generates a repository containing something like this: * -- * -- B -- O -- O -- O (d38e39e) \ N -- N -- N refs/tags/apache-arrow-0.12.0 (b1c3298) You should already have received notification emails for all of the O revisions, and so the following emails describe only the N revisions from the common base, B. Any revisions marked "omit" are not gone; other references still refer to them. Any revisions marked "discard" are gone forever. The 4 revisions listed above as "new" are entirely new to this repository and will be described in separate emails. The revisions listed as "add" were already present in the repository and have only been added to this reference. Summary of changes: CHANGELOG.md | 4 +- dev/tasks/conda-recipes/appveyor.yml | 21 +++--- dev/tasks/conda-recipes/travis.linux.yml | 7 +- dev/tasks/conda-recipes/travis.osx.yml | 13 ++-- ...inux_c_compilergcccxx_compilergxxpython2.7.yaml | 28 ++- ...inux_c_compilergcccxx_compilergxxpython3.6.yaml | 28 ++- ...inux_c_compilergcccxx_compilergxxpython3.7.yaml | 28 ++- ...lchain_ccxx_compilertoolchain_cxxpython2.7.yaml | 29 --- ...lchain_ccxx_compilertoolchain_cxxpython3.6.yaml | 29 --- ...lchain_ccxx_compilertoolchain_cxxpython3.7.yaml | 29 --- ..._compilerclangcxx_compilerclangxxpython2.7.yaml | 28 ++- ..._compilerclangcxx_compilerclangxxpython3.6.yaml | 28 ++- ..._compilerclangcxx_compilerclangxxpython3.7.yaml | 28 ++- ...lchain_ccxx_compilertoolchain_cxxpython2.7.yaml | 32 ...lchain_ccxx_compilertoolchain_cxxpython3.6.yaml | 32 ...lchain_ccxx_compilertoolchain_cxxpython3.7.yaml | 32 ..._compilervs2015cxx_compilervs2015python3.6.yaml | 20 + ..._compilervs2015cxx_compilervs2015python3.7.yaml | 20 + .../linux-packages/debian.ubuntu-trusty/changelog | 2 +- dev/tasks/linux-packages/debian/changelog | 2 +- dev/tasks/linux-packages/yum/arrow.spec.in | 2 +- dev/tasks/tasks.yml| 88 -- python/pyarrow/serialization.py| 41 +- 23 files changed, 247 insertions(+), 324 deletions(-) delete mode 100644 dev/tasks/conda-recipes/variants/linux_c_compilertoolchain_ccxx_compilertoolchain_cxxpython2.7.yaml delete mode 100644 dev/tasks/conda-recipes/variants/linux_c_compilertoolchain_ccxx_compilertoolchain_cxxpython3.6.yaml delete mode 100644 dev/tasks/conda-recipes/variants/linux_c_compilertoolchain_ccxx_compilertoolchain_cxxpython3.7.yaml delete mode 100644 dev/tasks/conda-recipes/variants/osx_c_compilertoolchain_ccxx_compilertoolchain_cxxpython2.7.yaml delete mode 100644 dev/tasks/conda-recipes/variants/osx_c_compilertoolchain_ccxx_compilertoolchain_cxxpython3.6.yaml delete mode 100644 dev/tasks/conda-recipes/variants/osx_c_compilertoolchain_ccxx_compilertoolchain_cxxpython3.7.yaml
[arrow] Diff for: [GitHub] robertnishihara closed pull request #3410: ARROW-4269: [Python] Fix serialization in pandas 0.22
diff --git a/python/pyarrow/serialization.py b/python/pyarrow/serialization.py index 6bbe1c7bc8..9b261c1bf7 100644 --- a/python/pyarrow/serialization.py +++ b/python/pyarrow/serialization.py @@ -174,26 +174,27 @@ def _deserialize_pandas_series(data): custom_serializer=_pickle_to_buffer, custom_deserializer=_load_pickle_from_buffer) -if hasattr(pd.core.arrays, 'interval'): -context.register_type( -pd.core.arrays.interval.IntervalArray, -'pd.core.arrays.interval.IntervalArray', -custom_serializer=_pickle_to_buffer, -custom_deserializer=_load_pickle_from_buffer) - -if hasattr(pd.core.arrays, 'period'): -context.register_type( -pd.core.arrays.period.PeriodArray, -'pd.core.arrays.period.PeriodArray', -custom_serializer=_pickle_to_buffer, -custom_deserializer=_load_pickle_from_buffer) - -if hasattr(pd.core.arrays, 'datetimes'): -context.register_type( -pd.core.arrays.datetimes.DatetimeArray, -'pd.core.arrays.datetimes.DatetimeArray', -custom_serializer=_pickle_to_buffer, -custom_deserializer=_load_pickle_from_buffer) +if hasattr(pd.core, 'arrays'): +if hasattr(pd.core.arrays, 'interval'): +context.register_type( +pd.core.arrays.interval.IntervalArray, +'pd.core.arrays.interval.IntervalArray', +custom_serializer=_pickle_to_buffer, +custom_deserializer=_load_pickle_from_buffer) + +if hasattr(pd.core.arrays, 'period'): +context.register_type( +pd.core.arrays.period.PeriodArray, +'pd.core.arrays.period.PeriodArray', +custom_serializer=_pickle_to_buffer, +custom_deserializer=_load_pickle_from_buffer) + +if hasattr(pd.core.arrays, 'datetimes'): +context.register_type( +pd.core.arrays.datetimes.DatetimeArray, +'pd.core.arrays.datetimes.DatetimeArray', +custom_serializer=_pickle_to_buffer, +custom_deserializer=_load_pickle_from_buffer) context.register_type( pd.DataFrame, 'pd.DataFrame', With regards, Apache Git Services
[arrow] branch master updated: ARROW-4269: [Python] Fix serialization in pandas 0.22
This is an automated email from the ASF dual-hosted git repository. robertnishihara pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/arrow.git The following commit(s) were added to refs/heads/master by this push: new fbfd1ff ARROW-4269: [Python] Fix serialization in pandas 0.22 fbfd1ff is described below commit fbfd1ff8e656f02105a22e3bbf89745234f84b36 Author: Philipp Moritz AuthorDate: Tue Jan 15 17:17:58 2019 -0800 ARROW-4269: [Python] Fix serialization in pandas 0.22 Author: Philipp Moritz Closes #3410 from pcmoritz/fix-pandas-0.22 and squashes the following commits: 54f13bb3 update dc6ffbde update 67133d0d make change consistent b6cba56a fix serialization in pandas 0.22 --- python/pyarrow/serialization.py | 41 + 1 file changed, 21 insertions(+), 20 deletions(-) diff --git a/python/pyarrow/serialization.py b/python/pyarrow/serialization.py index 6bbe1c7..9b261c1 100644 --- a/python/pyarrow/serialization.py +++ b/python/pyarrow/serialization.py @@ -174,26 +174,27 @@ def _register_custom_pandas_handlers(context): custom_serializer=_pickle_to_buffer, custom_deserializer=_load_pickle_from_buffer) -if hasattr(pd.core.arrays, 'interval'): -context.register_type( -pd.core.arrays.interval.IntervalArray, -'pd.core.arrays.interval.IntervalArray', -custom_serializer=_pickle_to_buffer, -custom_deserializer=_load_pickle_from_buffer) - -if hasattr(pd.core.arrays, 'period'): -context.register_type( -pd.core.arrays.period.PeriodArray, -'pd.core.arrays.period.PeriodArray', -custom_serializer=_pickle_to_buffer, -custom_deserializer=_load_pickle_from_buffer) - -if hasattr(pd.core.arrays, 'datetimes'): -context.register_type( -pd.core.arrays.datetimes.DatetimeArray, -'pd.core.arrays.datetimes.DatetimeArray', -custom_serializer=_pickle_to_buffer, -custom_deserializer=_load_pickle_from_buffer) +if hasattr(pd.core, 'arrays'): +if hasattr(pd.core.arrays, 'interval'): +context.register_type( +pd.core.arrays.interval.IntervalArray, +'pd.core.arrays.interval.IntervalArray', +custom_serializer=_pickle_to_buffer, +custom_deserializer=_load_pickle_from_buffer) + +if hasattr(pd.core.arrays, 'period'): +context.register_type( +pd.core.arrays.period.PeriodArray, +'pd.core.arrays.period.PeriodArray', +custom_serializer=_pickle_to_buffer, +custom_deserializer=_load_pickle_from_buffer) + +if hasattr(pd.core.arrays, 'datetimes'): +context.register_type( +pd.core.arrays.datetimes.DatetimeArray, +'pd.core.arrays.datetimes.DatetimeArray', +custom_serializer=_pickle_to_buffer, +custom_deserializer=_load_pickle_from_buffer) context.register_type( pd.DataFrame, 'pd.DataFrame',
[arrow] branch master updated: ARROW-4270: [Packaging][Conda] Update xcode version and remove toolchain builds
This is an automated email from the ASF dual-hosted git repository. kszucs pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/arrow.git The following commit(s) were added to refs/heads/master by this push: new a1a9221 ARROW-4270: [Packaging][Conda] Update xcode version and remove toolchain builds a1a9221 is described below commit a1a9221cf6522b4c9bc26a9ebf4a459f6217050e Author: Krisztián Szűcs AuthorDate: Wed Jan 16 04:23:49 2019 +0100 ARROW-4270: [Packaging][Conda] Update xcode version and remove toolchain builds Author: Krisztián Szűcs Closes #3411 from kszucs/cf-clang and squashes the following commits: f1a2efca update variant files 34c2fb9e no need of clobber file 92d0fe2d variant and clobber paths a152b373 update osx build b86c13f5 update channels 113fe013 update xcode for conda osx clang builds and remove conda toolchain entries --- dev/tasks/conda-recipes/appveyor.yml | 21 +++--- dev/tasks/conda-recipes/travis.linux.yml | 7 +- dev/tasks/conda-recipes/travis.osx.yml | 13 ++-- ...inux_c_compilergcccxx_compilergxxpython2.7.yaml | 28 ++- ...inux_c_compilergcccxx_compilergxxpython3.6.yaml | 28 ++- ...inux_c_compilergcccxx_compilergxxpython3.7.yaml | 28 ++- ...lchain_ccxx_compilertoolchain_cxxpython2.7.yaml | 29 --- ...lchain_ccxx_compilertoolchain_cxxpython3.6.yaml | 29 --- ...lchain_ccxx_compilertoolchain_cxxpython3.7.yaml | 29 --- ..._compilerclangcxx_compilerclangxxpython2.7.yaml | 28 ++- ..._compilerclangcxx_compilerclangxxpython3.6.yaml | 28 ++- ..._compilerclangcxx_compilerclangxxpython3.7.yaml | 28 ++- ...lchain_ccxx_compilertoolchain_cxxpython2.7.yaml | 32 ...lchain_ccxx_compilertoolchain_cxxpython3.6.yaml | 32 ...lchain_ccxx_compilertoolchain_cxxpython3.7.yaml | 32 ..._compilervs2015cxx_compilervs2015python3.6.yaml | 20 + ..._compilervs2015cxx_compilervs2015python3.7.yaml | 20 + dev/tasks/tasks.yml| 88 -- 18 files changed, 220 insertions(+), 300 deletions(-) diff --git a/dev/tasks/conda-recipes/appveyor.yml b/dev/tasks/conda-recipes/appveyor.yml index 3d3ba43..b3a4743 100644 --- a/dev/tasks/conda-recipes/appveyor.yml +++ b/dev/tasks/conda-recipes/appveyor.yml @@ -35,25 +35,26 @@ install: - cmd: set PYTHONUNBUFFERED=1 - # Add our channels. - - cmd: conda.exe config --set show_channel_urls true - - cmd: conda.exe config --remove channels defaults - - cmd: conda.exe config --add channels defaults - - cmd: conda.exe config --add channels conda-forge - # Configure the VM. - - cmd: conda.exe install -n root --quiet --yes conda-forge-ci-setup=1 - - cmd: run_conda_forge_build_setup + - cmd: conda.exe install -n root -c conda-forge --quiet --yes conda-forge-ci-setup=2 # Skip .NET project specific build phase. build: off test_script: + # Clone arrow - git clone -b {{ arrow.branch }} {{ arrow.remote }} arrow || exit /B - git -C arrow checkout {{ arrow.head }} || exit /B - - pushd arrow\dev\tasks\conda-recipes - - conda.exe build --output-folder . -m {{ variant_config_file }} parquet-cpp arrow-cpp pyarrow + + # Configure conda + - cmd: setup_conda_rc .\ .\ variants\{{ config }}.yaml + - cmd: run_conda_forge_build_setup + + # Build the recipes + - conda.exe build --output-folder . -m variants\{{ config }}.yaml parquet-cpp arrow-cpp pyarrow + + # Rename artifacts - pushd win-64 - for %%f in (*.tar.bz2) do ( set %%g=%%~nf diff --git a/dev/tasks/conda-recipes/travis.linux.yml b/dev/tasks/conda-recipes/travis.linux.yml index f0c4c77..cf2895f 100644 --- a/dev/tasks/conda-recipes/travis.linux.yml +++ b/dev/tasks/conda-recipes/travis.linux.yml @@ -38,6 +38,7 @@ install: MINICONDA_FILE="Miniconda3-latest-Linux-x86_64.sh" curl -L -O "${MINICONDA_URL}/${MINICONDA_FILE}" bash $MINICONDA_FILE -b + # Install conda build dependency - | echo "" @@ -49,16 +50,16 @@ before_script: - git clone -b {{ arrow.branch }} {{ arrow.remote }} arrow - git -C arrow checkout {{ arrow.head }} - pushd arrow/dev/tasks/conda-recipes + # Configure conda - - setup_conda_rc ./ ./ {{ variant_config_file }} + - setup_conda_rc ./ ./ variants/{{ config }}.yaml - source run_conda_forge_build_setup script: - # Don't need to run make_build_number, no build number decrementation happens, it's always 0 - | conda build --croot $TRAVIS_HOME/conda_build_root \ --output-folder . \ --m {{ variant_config_file }} \ +-m variants/{{ config }}.yaml \ parquet-cpp arrow-cpp pyarrow deploy: diff --git a/dev/tasks/conda-recipes/travis.osx.yml b/dev/tasks/conda-recipes/travis.osx.yml index 23fd6e1..31b8b76 100644 --- a/dev/tasks/conda-recipes/travis.osx.yml +++
[arrow] 02/04: [Release] Update .deb/.rpm changelogs for 0.12.0
This is an automated email from the ASF dual-hosted git repository. kszucs pushed a commit to annotated tag apache-arrow-0.12.0 in repository https://gitbox.apache.org/repos/asf/arrow.git commit db508eee05ec50ddf786e24b94d025c3852f8230 Author: Krisztián Szűcs AuthorDate: Wed Jan 16 04:29:25 2019 +0100 [Release] Update .deb/.rpm changelogs for 0.12.0 --- dev/tasks/linux-packages/debian.ubuntu-trusty/changelog | 6 ++ dev/tasks/linux-packages/debian/changelog | 6 ++ dev/tasks/linux-packages/yum/arrow.spec.in | 3 +++ 3 files changed, 15 insertions(+) diff --git a/dev/tasks/linux-packages/debian.ubuntu-trusty/changelog b/dev/tasks/linux-packages/debian.ubuntu-trusty/changelog index e54e05c..0aa6dd8 100644 --- a/dev/tasks/linux-packages/debian.ubuntu-trusty/changelog +++ b/dev/tasks/linux-packages/debian.ubuntu-trusty/changelog @@ -1,3 +1,9 @@ +apache-arrow (0.12.0-1) unstable; urgency=low + + * New upstream release. + + -- Krisztián Szűcs Wed, 16 Jan 2019 03:29:25 - + apache-arrow (0.11.0-1) unstable; urgency=low * New upstream release. diff --git a/dev/tasks/linux-packages/debian/changelog b/dev/tasks/linux-packages/debian/changelog index a0aff6f..c3a1d58 100644 --- a/dev/tasks/linux-packages/debian/changelog +++ b/dev/tasks/linux-packages/debian/changelog @@ -1,3 +1,9 @@ +apache-arrow (0.12.0-1) unstable; urgency=low + + * New upstream release. + + -- Krisztián Szűcs Wed, 16 Jan 2019 03:29:25 - + apache-arrow (0.11.0-1) unstable; urgency=low * New upstream release. diff --git a/dev/tasks/linux-packages/yum/arrow.spec.in b/dev/tasks/linux-packages/yum/arrow.spec.in index 568477e..9391ea5 100644 --- a/dev/tasks/linux-packages/yum/arrow.spec.in +++ b/dev/tasks/linux-packages/yum/arrow.spec.in @@ -397,6 +397,9 @@ Documentation for Apache Parquet GLib. %endif %changelog +* Wed Jan 16 2019 Krisztián Szűcs - 0.12.0-1 +- New upstream release. + * Thu Oct 04 2018 Kouhei Sutou - 0.11.0-1 - New upstream release.
[arrow] 04/04: [maven-release-plugin] prepare release apache-arrow-0.12.0
This is an automated email from the ASF dual-hosted git repository. kszucs pushed a commit to annotated tag apache-arrow-0.12.0 in repository https://gitbox.apache.org/repos/asf/arrow.git commit 8ca41384b5324bfd0ef3d3ed3f728e1d10ed73f0 Author: Krisztián Szűcs AuthorDate: Wed Jan 16 04:35:59 2019 +0100 [maven-release-plugin] prepare release apache-arrow-0.12.0 --- java/adapter/jdbc/pom.xml | 2 +- java/flight/pom.xml | 2 +- java/format/pom.xml | 2 +- java/memory/pom.xml | 2 +- java/plasma/pom.xml | 2 +- java/pom.xml | 4 ++-- java/tools/pom.xml| 2 +- java/vector/pom.xml | 2 +- 8 files changed, 9 insertions(+), 9 deletions(-) diff --git a/java/adapter/jdbc/pom.xml b/java/adapter/jdbc/pom.xml index ab0b9b5..add4e07 100644 --- a/java/adapter/jdbc/pom.xml +++ b/java/adapter/jdbc/pom.xml @@ -16,7 +16,7 @@ org.apache.arrow arrow-java-root -0.12.0-SNAPSHOT +0.12.0 ../../pom.xml diff --git a/java/flight/pom.xml b/java/flight/pom.xml index c6de29f..f99194e 100644 --- a/java/flight/pom.xml +++ b/java/flight/pom.xml @@ -11,7 +11,7 @@ org.apache.arrow arrow-java-root -0.12.0-SNAPSHOT +0.12.0 arrow-flight diff --git a/java/format/pom.xml b/java/format/pom.xml index d5ccd5f..7427824 100644 --- a/java/format/pom.xml +++ b/java/format/pom.xml @@ -15,7 +15,7 @@ arrow-java-root org.apache.arrow - 0.12.0-SNAPSHOT + 0.12.0 arrow-format diff --git a/java/memory/pom.xml b/java/memory/pom.xml index 7e2c42d..8f190e5 100644 --- a/java/memory/pom.xml +++ b/java/memory/pom.xml @@ -14,7 +14,7 @@ org.apache.arrow arrow-java-root -0.12.0-SNAPSHOT +0.12.0 arrow-memory Arrow Memory diff --git a/java/plasma/pom.xml b/java/plasma/pom.xml index d50171a..a431ef2 100644 --- a/java/plasma/pom.xml +++ b/java/plasma/pom.xml @@ -14,7 +14,7 @@ org.apache.arrow arrow-java-root -0.12.0-SNAPSHOT +0.12.0 arrow-plasma Arrow Plasma Client diff --git a/java/pom.xml b/java/pom.xml index 0df1178..47b62cc 100644 --- a/java/pom.xml +++ b/java/pom.xml @@ -20,7 +20,7 @@ org.apache.arrow arrow-java-root - 0.12.0-SNAPSHOT + 0.12.0 pom Apache Arrow Java Root POM @@ -45,7 +45,7 @@ scm:git:https://github.com/apache/arrow.git scm:git:https://github.com/apache/arrow.git https://github.com/apache/arrow -apache-arrow-0.10.0 +apache-arrow-0.12.0 diff --git a/java/tools/pom.xml b/java/tools/pom.xml index d01c951..9af5d4d 100644 --- a/java/tools/pom.xml +++ b/java/tools/pom.xml @@ -14,7 +14,7 @@ org.apache.arrow arrow-java-root -0.12.0-SNAPSHOT +0.12.0 arrow-tools Arrow Tools diff --git a/java/vector/pom.xml b/java/vector/pom.xml index fc4eb8f..4daa91e 100644 --- a/java/vector/pom.xml +++ b/java/vector/pom.xml @@ -14,7 +14,7 @@ org.apache.arrow arrow-java-root -0.12.0-SNAPSHOT +0.12.0 arrow-vector Arrow Vectors
[arrow] 01/04: [Release] Update CHANGELOG.md for 0.12.0
This is an automated email from the ASF dual-hosted git repository. kszucs pushed a commit to annotated tag apache-arrow-0.12.0 in repository https://gitbox.apache.org/repos/asf/arrow.git commit 6c8c0c1211ce1b4999184560d6fe4d83b25745e7 Author: Krisztián Szűcs AuthorDate: Wed Jan 16 04:29:23 2019 +0100 [Release] Update CHANGELOG.md for 0.12.0 --- CHANGELOG.md | 605 +++ 1 file changed, 605 insertions(+) diff --git a/CHANGELOG.md b/CHANGELOG.md index 853806c..5cacdfd 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -17,6 +17,610 @@ under the License. --> +# Apache Arrow 0.12.0 (16 January 2019) + +## Bug + +* ARROW-1847 - [Doc] Document the difference between RecordBatch and Table in an FAQ fashion +* ARROW-1994 - [Python] Test against Pandas master +* ARROW-2026 - [Python] Cast all timestamp resolutions to INT96 use\_deprecated\_int96\_timestamps=True +* ARROW-2038 - [Python] Follow-up bug fixes for s3fs Parquet support +* ARROW-2113 - [Python] Incomplete CLASSPATH with "hadoop" contained in it can fool the classpath setting HDFS logic +* ARROW-2591 - [Python] Segmentation fault when writing empty ListType column to Parquet +* ARROW-2592 - [Python] Error reading old Parquet file due to metadata backwards compatibility issue +* ARROW-2708 - [C++] Internal GetValues function in arrow::compute should check for nullptr +* ARROW-2970 - [Python] NumPyConverter::Visit for Binary/String/FixedSizeBinary can overflow +* ARROW-3058 - [Python] Feather reads fail with unintuitive error when conversion from pandas yields ChunkedArray +* ARROW-3186 - [GLib] mesonbuild failures in Travis CI +* ARROW-3202 - [C++] Build does not succeed on Alpine Linux +* ARROW-3225 - [C++/Python] Pandas object conversion of ListType and ListType +* ARROW-3324 - [Parquet] Free more internal resources when writing multiple row groups +* ARROW-3343 - [Java] Java tests fail non-deterministically with memory leak from Flight tests +* ARROW-3405 - [Python] Document CSV reader +* ARROW-3428 - [Python] from\_pandas gives incorrect results when converting floating point to bool +* ARROW-3436 - [C++] Boost version required by Gandiva is too new for Ubuntu 14.04 +* ARROW-3437 - [Gandiva][C++] Configure static linking of libgcc, libstdc++ with LDFLAGS +* ARROW-3438 - [Packaging] Escaped bulletpoints in changelog +* ARROW-3445 - [GLib] Parquet GLib doesn't link Arrow GLib +* ARROW-3449 - [C++] Support CMake 3.2 for "out of the box" builds +* ARROW-3466 - [Python] Crash when importing tensorflow and pyarrow +* ARROW-3467 - Building against external double conversion is broken +* ARROW-3470 - [C++] Row-wise conversion tutorial has fallen out of date +* ARROW-3477 - [C++] Testsuite fails on 32 bit arch +* ARROW-3480 - [Website] Install document for Ubuntu is broken +* ARROW-3485 - [C++] Examples fail with Protobuf error +* ARROW-3494 - [C++] re2 conda-forge package not working in toolchain +* ARROW-3516 - [C++] Use unsigned type for difference of pointers in parallel\_memcpy +* ARROW-3517 - [C++] MinGW 32bit build causes g++ segv +* ARROW-3524 - [C++] Fix compiler warnings from ARROW-3409 on clang-6 +* ARROW-3527 - [R] Unused variables in R-package C++ code +* ARROW-3528 - [R] Typo in R documentation +* ARROW-3535 - [Python] pip install tensorflow install too new numpy in manylinux1 build +* ARROW-3541 - [Rust] Update BufferBuilder to allow for new bit-packed BooleanArray +* ARROW-3544 - [Gandiva] Populate function registry in multiple compilation units to mitigate long compile times in release mode +* ARROW-3549 - [Rust] Replace i64 with usize for some bit utility functions +* ARROW-3573 - [Rust] with\_bitset does not set valid bits correctly +* ARROW-3580 - [Gandiva][C++] Build error with g++ 8.2.0 +* ARROW-3586 - [Python] Segmentation fault when converting empty table to pandas with categoricals +* ARROW-3598 - [Plasma] plasma\_store\_server fails linking with GPU enabled +* ARROW-3613 - [Go] Resize does not correctly update the length +* ARROW-3614 - [R] Handle Type::TIMESTAMP from Arrow to R +* ARROW-3658 - [Rust] validation of offsets buffer is incorrect for \`List\` +* ARROW-3670 - [C++] Use FindBacktrace to find execinfo.h support +* ARROW-3687 - [Rust] Anything measuring array slots should be \`usize\` +* ARROW-3698 - [C++] Segmentation fault when using a large table in Gandiva +* ARROW-3700 - [C++] CSV parser should allow ignoring empty lines +* ARROW-3703 - [Python] DataFrame.to\_parquet crashes if datetime column has time zones +* ARROW-3707 - [C++] test failure with zstd 1.3.7 +* ARROW-3711 - [C++] Don't pass CXX\_FLAGS to C\_FLAGS +* ARROW-3712 - [CI] License check regression (RAT failure) +* ARROW-3715 - [C++] gflags\_ep fails to build with CMake 3.13 +* ARROW-3716 - [R] Missing cases for ChunkedArray conversion +* ARROW-3728 - [Python] Merging Parquet Files - Pandas Meta in Schema Mismatch +* ARROW-3734 - [C++] Linking static zstd library fails on
svn commit: r31980 - in /dev/arrow/apache-arrow-0.12.0-rc4: ./ apache-arrow-0.12.0.tar.gz apache-arrow-0.12.0.tar.gz.asc apache-arrow-0.12.0.tar.gz.sha256 apache-arrow-0.12.0.tar.gz.sha512
Author: kszucs Date: Wed Jan 16 04:01:43 2019 New Revision: 31980 Log: Apache Arrow 0.12.0 RC4 Added: dev/arrow/apache-arrow-0.12.0-rc4/ dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz (with props) dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz.asc dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz.sha256 dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz.sha512 Added: dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz == Binary file - no diff available. Propchange: dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz -- svn:mime-type = application/octet-stream Added: dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz.asc == --- dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz.asc (added) +++ dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz.asc Wed Jan 16 04:01:43 2019 @@ -0,0 +1,16 @@ +-BEGIN PGP SIGNATURE- + +iQIzBAABCAAdFiEEJl+Aq4T+AxJ+FPARJbzKUiDYQHkFAlw+rJAACgkQJbzKUiDY +QHlWAw//WipvdnrkzZjT55xos+S4lkMNSAqKVHXw8Fc+P5ybSRUBe1y/l+eInr2r ++xVGmDux+tPnRz05gH8AjVD/sgbSQdJfy3Cvl79y7AJJVENwqvWtHTdmFyB8XoPd +Qji7F+UzQ+RA0ghsC90k5f5lohiA23vVjCvLg/dKNk1ImrkvZ5KEdQIkbUi30US/ +EdJ1JABEfmd9E1hhwogZzPEZ96IHt/xXhwoW6Hz7SqYz24yNYsFS0YfyYqGNjaZf +S4oxMqylguKLhC12T1cLnrNjQJsmtnqd5rEWjuJBnr27zbrn1weKrpyPLONvVx5L +ynYEDWKz7gMylxHxcNN668tY4WBQOgg3CM9mboFZzQVIpTqwxX2DEHTxSjf2MgEb +jE+BrTyGUaVfudkIG0Q53hQuKUraKpqISLOeXRp+j6EohWTQO19MbsU10YTBz6Ym +z/XPql9wpFAlAX6CAMKwEE8tUFsU7tXVs1OefwJ4i+GAGDGKnrRcfw2vhyJaHydj +6qgx3QQxERjdCjKDwkT9tFs7sbS1AMOHWZ+B/6h2YdGFDRoGRYHxnY5uDRRxYB0b +4bmAs8F2x767UwYiIieJmgpFBq2i+STb2hc5MQWKlRmYrBQujvigLoXrvOYFsZP2 +YeNhjImQ3/kCs9rlTiLxANPUWgTYyc736MNggDPxUVrcPY8imIQ= +=MWsq +-END PGP SIGNATURE- Added: dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz.sha256 == --- dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz.sha256 (added) +++ dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz.sha256 Wed Jan 16 04:01:43 2019 @@ -0,0 +1 @@ +34dae7e4dde9274e9a52610683e78a80f3ca312258ad9e9f2c0973cf44247a98 apache-arrow-0.12.0.tar.gz Added: dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz.sha512 == --- dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz.sha512 (added) +++ dev/arrow/apache-arrow-0.12.0-rc4/apache-arrow-0.12.0.tar.gz.sha512 Wed Jan 16 04:01:43 2019 @@ -0,0 +1 @@ +a92125e285a66ec590c0e8defbee272e40c0dda41f5fc9f2cc1daf28cfad22bec6473b8cb251a899937a50d9d7167ed54b049825710ad78bc8b78a998cfe5b97 apache-arrow-0.12.0.tar.gz
[arrow] Diff for: [GitHub] kszucs closed pull request #3402: WIP ARROW-4260: [Python] NumPy buffer protocol failure
diff --git a/ci/appveyor-cpp-build.bat b/ci/appveyor-cpp-build.bat index 387dd55d18..6681f5673f 100644 --- a/ci/appveyor-cpp-build.bat +++ b/ci/appveyor-cpp-build.bat @@ -97,7 +97,7 @@ if "%JOB%" == "Build_Debug" ( conda create -n arrow -q -y ^ python=%PYTHON% ^ - six pytest setuptools numpy pandas cython hypothesis ^ + --file=ci\conda_env_python.yml ^ thrift-cpp=0.11.0 boost-cpp ^ -c conda-forge diff --git a/ci/conda_env_python.yml b/ci/conda_env_python.yml index b51f5c32f3..0a5b57167f 100644 --- a/ci/conda_env_python.yml +++ b/ci/conda_env_python.yml @@ -18,8 +18,9 @@ cython cloudpickle hypothesis -numpy +numpy=1.15.4 pandas pytest setuptools setuptools_scm +six \ No newline at end of file With regards, Apache Git Services
[arrow] Diff for: [GitHub] kou closed pull request #3401: ARROW-4246: [Plasma][Python][Follow-up] Ensure plasma::ObjectTableEntry always has the same size regardless of whether built with CUDA support
diff --git a/cpp/src/plasma/common.h b/cpp/src/plasma/common.h index 8e90fb301c..dfbd90c3aa 100644 --- a/cpp/src/plasma/common.h +++ b/cpp/src/plasma/common.h @@ -72,6 +72,12 @@ enum class ObjectState : int { PLASMA_SEALED }; +namespace internal { + +struct CudaIpcPlaceholder {}; + +} // namespace internal + /// This type is used by the Plasma store. It is here because it is exposed to /// the eviction policy. struct ObjectTableEntry { @@ -106,10 +112,10 @@ struct ObjectTableEntry { unsigned char digest[kDigestSize]; #ifdef PLASMA_CUDA - /// Put CUDA related members at the last to create Python bindings easily. - /// IPC GPU handle to share with clients. std::shared_ptr<::arrow::cuda::CudaIpcMemHandle> ipc_handle; +#else + std::shared_ptr ipc_handle; #endif }; diff --git a/python/pyarrow/_plasma.pyx b/python/pyarrow/_plasma.pyx index cfaa39c96e..4f64f202ce 100644 --- a/python/pyarrow/_plasma.pyx +++ b/python/pyarrow/_plasma.pyx @@ -45,6 +45,8 @@ PLASMA_WAIT_TIMEOUT = 2 ** 30 cdef extern from "plasma/common.h" nogil: +cdef cppclass CCudaIpcPlaceholder" plasma::internal::CudaIpcPlaceholder": +pass cdef cppclass CUniqueID" plasma::UniqueID": @@ -79,6 +81,7 @@ cdef extern from "plasma/common.h" nogil: int64_t create_time int64_t construct_duration CObjectState state +shared_ptr[CCudaIpcPlaceholder] ipc_handle ctypedef unordered_map[CUniqueID, unique_ptr[CObjectTableEntry]] \ CObjectTable" plasma::ObjectTable" With regards, Apache Git Services
[arrow] branch master updated: ARROW-4246: [Plasma][Python][Follow-up] Ensure plasma::ObjectTableEntry always has the same size regardless of whether built with CUDA support
This is an automated email from the ASF dual-hosted git repository. kou pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/arrow.git The following commit(s) were added to refs/heads/master by this push: new 87ac6fd ARROW-4246: [Plasma][Python][Follow-up] Ensure plasma::ObjectTableEntry always has the same size regardless of whether built with CUDA support 87ac6fd is described below commit 87ac6fddaf21d0b4ee8b8090533ff293db0da1b4 Author: Wes McKinney AuthorDate: Tue Jan 15 17:02:12 2019 +0900 ARROW-4246: [Plasma][Python][Follow-up] Ensure plasma::ObjectTableEntry always has the same size regardless of whether built with CUDA support This ensures that the struct size remains constant whether `PLASMA_CUDA` is on or off. How this is handled on the Cython side will have to be done in a separate patch Author: Wes McKinney Closes #3401 from wesm/plasma-cuda-fix and squashes the following commits: ef203421 Ensure plasma::ObjectTableEntry always has the same size regardless of whether Plasma has CUDA support --- cpp/src/plasma/common.h| 10 -- python/pyarrow/_plasma.pyx | 3 +++ 2 files changed, 11 insertions(+), 2 deletions(-) diff --git a/cpp/src/plasma/common.h b/cpp/src/plasma/common.h index 8e90fb3..dfbd90c 100644 --- a/cpp/src/plasma/common.h +++ b/cpp/src/plasma/common.h @@ -72,6 +72,12 @@ enum class ObjectState : int { PLASMA_SEALED }; +namespace internal { + +struct CudaIpcPlaceholder {}; + +} // namespace internal + /// This type is used by the Plasma store. It is here because it is exposed to /// the eviction policy. struct ObjectTableEntry { @@ -106,10 +112,10 @@ struct ObjectTableEntry { unsigned char digest[kDigestSize]; #ifdef PLASMA_CUDA - /// Put CUDA related members at the last to create Python bindings easily. - /// IPC GPU handle to share with clients. std::shared_ptr<::arrow::cuda::CudaIpcMemHandle> ipc_handle; +#else + std::shared_ptr ipc_handle; #endif }; diff --git a/python/pyarrow/_plasma.pyx b/python/pyarrow/_plasma.pyx index cfaa39c..4f64f20 100644 --- a/python/pyarrow/_plasma.pyx +++ b/python/pyarrow/_plasma.pyx @@ -45,6 +45,8 @@ PLASMA_WAIT_TIMEOUT = 2 ** 30 cdef extern from "plasma/common.h" nogil: +cdef cppclass CCudaIpcPlaceholder" plasma::internal::CudaIpcPlaceholder": +pass cdef cppclass CUniqueID" plasma::UniqueID": @@ -79,6 +81,7 @@ cdef extern from "plasma/common.h" nogil: int64_t create_time int64_t construct_duration CObjectState state +shared_ptr[CCudaIpcPlaceholder] ipc_handle ctypedef unordered_map[CUniqueID, unique_ptr[CObjectTableEntry]] \ CObjectTable" plasma::ObjectTable"