commit:     edf6876c43828c4321e0f9b70398f4cc8f51da55
Author:     Michał Górny <mgorny <AT> gentoo <DOT> org>
AuthorDate: Wed Jul 10 09:51:50 2024 +0000
Commit:     Michał Górny <mgorny <AT> gentoo <DOT> org>
CommitDate: Wed Jul 10 10:53:03 2024 +0000
URL:        https://gitweb.gentoo.org/repo/gentoo.git/commit/?id=edf6876c

dev-python/tables: Add a patch for numpy-2 support

Signed-off-by: Michał Górny <mgorny <AT> gentoo.org>

 dev-python/tables/files/tables-3.9.2-numpy-2.patch | 264 +++++++++++++++++++++
 dev-python/tables/tables-3.9.2-r1.ebuild           |  80 +++++++
 2 files changed, 344 insertions(+)

diff --git a/dev-python/tables/files/tables-3.9.2-numpy-2.patch 
b/dev-python/tables/files/tables-3.9.2-numpy-2.patch
new file mode 100644
index 000000000000..a3d56d13f1a8
--- /dev/null
+++ b/dev-python/tables/files/tables-3.9.2-numpy-2.patch
@@ -0,0 +1,264 @@
+diff --git a/setup.py b/setup.py
+index 48dc6ca5..e8957c1f 100755
+--- a/setup.py
++++ b/setup.py
+@@ -736,7 +736,10 @@ if __name__ == "__main__":
+ 
+     # -----------------------------------------------------------------
+ 
+-    def_macros = [("NDEBUG", 1)]
++    def_macros = [
++        ("NDEBUG", 1),
++        ("NPY_TARGET_VERSION", "NPY_1_20_API_VERSION"),
++    ]
+ 
+     # Define macros for Windows platform
+     if os.name == "nt":
+diff --git a/src/utils.c b/src/utils.c
+index 15fce02d..b28dbc90 100644
+--- a/src/utils.c
++++ b/src/utils.c
+@@ -765,8 +765,8 @@ hid_t create_ieee_complex64(const char *byteorder) {
+     return float_id;
+   }
+ 
+-  H5Tinsert(complex_id, "r", HOFFSET(npy_complex64, real), float_id);
+-  H5Tinsert(complex_id, "i", HOFFSET(npy_complex64, imag), float_id);
++  H5Tinsert(complex_id, "r", 0, float_id);
++  H5Tinsert(complex_id, "i", 4, float_id);
+   H5Tclose(float_id);
+   return complex_id;
+ }
+@@ -790,8 +790,8 @@ hid_t create_ieee_complex128(const char *byteorder) {
+     return float_id;
+   }
+ 
+-  H5Tinsert(complex_id, "r", HOFFSET(npy_complex128, real), float_id);
+-  H5Tinsert(complex_id, "i", HOFFSET(npy_complex128, imag), float_id);
++  H5Tinsert(complex_id, "r", 0, float_id);
++  H5Tinsert(complex_id, "i", 8, float_id);
+   H5Tclose(float_id);
+   return complex_id;
+ }
+@@ -822,8 +822,8 @@ hid_t create_ieee_complex192(const char *byteorder) {
+     return err;
+   }
+ 
+-  H5Tinsert(complex_id, "r", HOFFSET(npy_complex192, real), float_id);
+-  H5Tinsert(complex_id, "i", HOFFSET(npy_complex192, imag), float_id);
++  H5Tinsert(complex_id, "r", 0, float_id);
++  H5Tinsert(complex_id, "i", 12, float_id);
+   H5Tclose(float_id);
+   return complex_id;
+ }
+@@ -854,8 +854,8 @@ hid_t create_ieee_complex256(const char *byteorder) {
+     return err;
+   }
+ 
+-  H5Tinsert(complex_id, "r", HOFFSET(npy_complex256, real), float_id);
+-  H5Tinsert(complex_id, "i", HOFFSET(npy_complex256, imag), float_id);
++  H5Tinsert(complex_id, "r", 0, float_id);
++  H5Tinsert(complex_id, "i", 16, float_id);
+   H5Tclose(float_id);
+   return complex_id;
+ }
+diff --git a/tables/atom.py b/tables/atom.py
+index 56ab6423..5d4fba2b 100644
+--- a/tables/atom.py
++++ b/tables/atom.py
+@@ -276,15 +276,15 @@ class Atom(metaclass=MetaAtom):
+             >>> atom1 = StringAtom(itemsize=10)  # same as ``atom2``
+             >>> atom2 = Atom.from_kind('string', 10)  # same as ``atom1``
+             >>> atom3 = IntAtom()
+-            >>> atom1 == 'foo'
++            >>> bool(atom1 == 'foo')
+             False
+-            >>> atom1 == atom2
++            >>> bool(atom1 == atom2)
+             True
+-            >>> atom2 != atom1
++            >>> bool(atom2 != atom1)
+             False
+-            >>> atom1 == atom3
++            >>> bool(atom1 == atom3)
+             False
+-            >>> atom3 != atom2
++            >>> bool(atom3 != atom2)
+             True
+ 
+     """
+diff --git a/tables/index.py b/tables/index.py
+index e8c8caf7..28ff37e2 100644
+--- a/tables/index.py
++++ b/tables/index.py
+@@ -581,7 +581,8 @@ class Index(NotLoggedMixin, Group, indexesextension.Index):
+                 # Add a second offset in this case
+                 # First normalize the number of rows
+                 offset2 = (nrow % self.nslicesblock) * slicesize // lbucket
+-                idx += offset2
++                assert offset2 < 2**(indsize*8)
++                idx += np.asarray(offset2).astype(idx.dtype)
+         # Add the last row at the beginning of arr & idx (if needed)
+         if (indsize == 8 and nelementsILR > 0):
+             # It is possible that the values in LR are already sorted.
+@@ -622,11 +623,11 @@ class Index(NotLoggedMixin, Group, 
indexesextension.Index):
+             show_stats("Entering final_idx32", tref)
+         # Do an upcast first in order to add the offset.
+         idx = idx.astype('uint64')
+-        idx += offset
++        idx += np.asarray(offset).astype(idx.dtype)
+         # The next partition is valid up to table sizes of
+         # 2**30 * 2**18 = 2**48 bytes, that is, 256 Tera-elements,
+         # which should be a safe figure, at least for a while.
+-        idx //= self.lbucket
++        idx //= np.asarray(self.lbucket).astype(idx.dtype)
+         # After the division, we can downsize the indexes to 'uint32'
+         idx = idx.astype('uint32')
+         if profile:
+@@ -2002,7 +2003,7 @@ class Index(NotLoggedMixin, Group, 
indexesextension.Index):
+                 else:
+                     self.indicesLR._read_index_slice(start, stop, idx)
+                 if indsize == 8:
+-                    idx //= lbucket
++                    idx //= np.asarray(lbucket).astype(idx.dtype)
+                 elif indsize == 2:
+                     # The chunkmap size cannot be never larger than 'int_'
+                     idx = idx.astype("int_")
+diff --git a/tables/tests/common.py b/tables/tests/common.py
+index 31378a88..1992f39a 100644
+--- a/tables/tests/common.py
++++ b/tables/tests/common.py
+@@ -205,7 +205,7 @@ def allequal(a, b, flavor="numpy"):
+     return result
+ 
+ 
+-def areArraysEqual(arr1, arr2):
++def areArraysEqual(arr1, arr2, *, check_type=True):
+     """Are both `arr1` and `arr2` equal arrays?
+ 
+     Arguments can be regular NumPy arrays, chararray arrays or
+@@ -217,8 +217,8 @@ def areArraysEqual(arr1, arr2):
+     t1 = type(arr1)
+     t2 = type(arr2)
+ 
+-    if not ((hasattr(arr1, 'dtype') and arr1.dtype == arr2.dtype) or
+-            issubclass(t1, t2) or issubclass(t2, t1)):
++    if check_type and not ((hasattr(arr1, 'dtype') and arr1.dtype == 
arr2.dtype) or
++                           issubclass(t1, t2) or issubclass(t2, t1)):
+         return False
+ 
+     return np.all(arr1 == arr2)
+diff --git a/tables/tests/test_expression.py b/tables/tests/test_expression.py
+index 018d4208..d9c0e990 100644
+--- a/tables/tests/test_expression.py
++++ b/tables/tests/test_expression.py
+@@ -265,9 +265,12 @@ class MixedContainersTestCase(common.TempFileMixin, 
common.PyTablesTestCase):
+         if common.verbose:
+             print("Computed expression:", repr(r1), r1.dtype)
+             print("Should look like:", repr(r2), r2.dtype)
+-        self.assertTrue(
+-            r1.shape == r2.shape and r1.dtype == r2.dtype and r1 == r2,
+-            "Evaluate is returning a wrong value.")
++        msg = f"Evaluate is returning a wrong value: {expr_str}\n{r1=}\n{r2=}"
++        self.assertEqual(r1.shape, r2.shape, msg=msg)
++        # In something like 2 * np.in16(3) + np.int16(2) the result is still a
++        # np.int16 in NumPy 2.0, so we shouldn't actually check this:
++        # self.assertEqual(r1.dtype, r2.dtype, msg=msg)
++        self.assertEqual(r1, r2, msg=msg)
+ 
+     def test01a_out(self):
+         """Checking expressions with mixed objects (`out` param)"""
+@@ -305,8 +308,9 @@ class MixedContainersTestCase(common.TempFileMixin, 
common.PyTablesTestCase):
+             if common.verbose:
+                 print("Computed expression:", repr(r1), r1.dtype)
+                 print("Should look like:", repr(r2), r2.dtype)
+-            self.assertTrue(common.areArraysEqual(r1, r2),
+-                            "Evaluate is returning a wrong value.")
++            msg = f"Evaluate is returning a wrong value: 
{expr_str}\n{r1=}\n{r2=}"
++            # On NumPy 2 type promotion is different so don't check type here
++            self.assertTrue(common.areArraysEqual(r1, r2, check_type=False), 
msg=msg)
+ 
+     def test02a_sss(self):
+         """Checking mixed objects and start, stop, step (I)"""
+diff --git a/tables/tests/test_indexvalues.py 
b/tables/tests/test_indexvalues.py
+index fac33af8..85ca38c3 100644
+--- a/tables/tests/test_indexvalues.py
++++ b/tables/tests/test_indexvalues.py
+@@ -2296,6 +2296,11 @@ class SelectValuesTestCase(common.TempFileMixin, 
common.PyTablesTestCase):
+         self.assertFalse(t1var3.index.dirty)
+         self.assertFalse(t1var4.index.dirty)
+ 
++        # TODO: IT IS DIRTY BECAUSE THIS FIXES THINGS FOR FINSV2aTestCase,
++        # which otherwise fails a test a few lines below!
++        for col in table1.colinstances.values():
++            col.reindex()
++
+         # Do some selections and check the results
+         # First selection: string
+         # Convert the limits to the appropriate type
+@@ -2318,11 +2323,15 @@ class SelectValuesTestCase(common.TempFileMixin, 
common.PyTablesTestCase):
+         # Second selection: bool
+         results1 = [p["var2"] for p in table1.where('t1var2 == True')]
+         results2 = [p["var2"] for p in table2 if p["var2"] is True]
+-        if common.verbose:
+-            print("Length results:", len(results1))
+-            print("Should be:", len(results2))
+-        self.assertEqual(len(results1), len(results2))
+-        self.assertEqual(results1, results2)
++        t2var1_vals = [p["var1"] for p in table2]
++        t2var2_vals = [p["var2"] for p in table2]
++        msg = (
++            f"Incorrect results for t1var2[n] == True where\n"
++            
f"t2var1_vals={repr(t2var1_vals)}\nt2var2_vals={repr(t2var2_vals)}\n"
++            f"\n{results1=}\n{results2=}"
++        )
++        self.assertEqual(len(results1), len(results2), msg=msg)
++        self.assertEqual(results1, results2, msg=msg)
+ 
+         # Third selection: int
+         # Convert the limits to the appropriate type
+@@ -3228,7 +3237,9 @@ class LastRowReuseBuffers(common.PyTablesTestCase):
+ 
+ 
+ normal_tests = (
+-    "SV1aTestCase", "SV2aTestCase", "SV3aTestCase",
++    "SV1aTestCase",
++    "SV2aTestCase",
++    "SV3aTestCase",
+ )
+ 
+ heavy_tests = (
+diff --git a/tables/utils.py b/tables/utils.py
+index e11e5ba7..7d786e32 100644
+--- a/tables/utils.py
++++ b/tables/utils.py
+@@ -25,6 +25,11 @@ byteorders = {
+ SizeType = np.int64
+ 
+ 
++copy_if_needed = (
++    None if np.lib.NumpyVersion(np.__version__) >= "2.0.0" else False
++)
++
++
+ def correct_byteorder(ptype, byteorder):
+     """Fix the byteorder depending on the PyTables types."""
+ 
+@@ -78,7 +83,7 @@ def idx2long(index):
+ # with atom from a generic python type.  If copy is stated as True, it
+ # is assured that it will return a copy of the object and never the same
+ # object or a new one sharing the same memory.
+-def convert_to_np_atom(arr, atom, copy=False):
++def convert_to_np_atom(arr, atom, copy=copy_if_needed):
+     """Convert a generic object into a NumPy object compliant with atom."""
+ 
+     # First, convert the object into a NumPy array
+@@ -112,7 +117,7 @@ def convert_to_np_atom2(object, atom):
+ 
+     # Check whether the object needs to be copied to make the operation
+     # safe to in-place conversion.
+-    copy = atom.type in ['time64']
++    copy = True if atom.type in ['time64'] else copy_if_needed
+     nparr = convert_to_np_atom(object, atom, copy)
+     # Finally, check the byteorder and change it if needed
+     byteorder = byteorders[nparr.dtype.byteorder]

diff --git a/dev-python/tables/tables-3.9.2-r1.ebuild 
b/dev-python/tables/tables-3.9.2-r1.ebuild
new file mode 100644
index 000000000000..bdbc563361d3
--- /dev/null
+++ b/dev-python/tables/tables-3.9.2-r1.ebuild
@@ -0,0 +1,80 @@
+# Copyright 1999-2024 Gentoo Authors
+# Distributed under the terms of the GNU General Public License v2
+
+EAPI=8
+
+DISTUTILS_EXT=1
+DISTUTILS_USE_PEP517=setuptools
+PYTHON_COMPAT=( pypy3 python3_{10..12} )
+PYTHON_REQ_USE="threads(+)"
+
+inherit distutils-r1 prefix pypi
+
+DESCRIPTION="Hierarchical datasets for Python"
+HOMEPAGE="
+       https://www.pytables.org/
+       https://github.com/PyTables/PyTables/
+       https://pypi.org/project/tables/
+"
+
+LICENSE="BSD"
+SLOT="0"
+KEYWORDS="~amd64 ~arm ~arm64 ~ia64 ~loong ~ppc64 ~riscv ~sparc ~x86 
~amd64-linux ~x86-linux"
+IUSE="+cpudetection examples test"
+RESTRICT="!test? ( test )"
+
+DEPEND="
+       app-arch/bzip2:0=
+       app-arch/lz4:0=
+       >=app-arch/zstd-1.0.0:=
+       >=dev-libs/c-blosc-1.11.1:0=
+       >=dev-libs/c-blosc2-2.11.0:=
+       dev-libs/lzo:2=
+       >=dev-python/numpy-1.19.0:=[${PYTHON_USEDEP}]
+       >=sci-libs/hdf5-1.8.4:=
+"
+RDEPEND="
+       ${DEPEND}
+       >=dev-python/numexpr-2.6.2[${PYTHON_USEDEP}]
+       dev-python/packaging[${PYTHON_USEDEP}]
+       cpudetection? ( dev-python/py-cpuinfo[${PYTHON_USEDEP}] )
+"
+BDEPEND="
+       >=dev-python/cython-0.21[${PYTHON_USEDEP}]
+       virtual/pkgconfig
+       cpudetection? ( dev-python/py-cpuinfo[${PYTHON_USEDEP}] )
+       test? (
+               ${RDEPEND}
+       )
+"
+
+python_prepare_all() {
+       local PATCHES=(
+               # https://github.com/PyTables/PyTables/pull/1176
+               "${FILESDIR}/${P}-numpy-2.patch"
+       )
+
+       rm -r c-blosc/{blosc,internal-complibs} || die
+
+       distutils-r1_python_prepare_all
+
+       sed -i -e '/blosc2/d' pyproject.toml || die
+       hprefixify -w '/prefixes =/' setup.py
+
+       export PYTABLES_NO_EMBEDDED_LIBS=1
+       export USE_PKGCONFIG=TRUE
+}
+
+python_test() {
+       cd "${BUILD_DIR}/install$(python_get_sitedir)" || die
+       "${EPYTHON}" tables/tests/test_all.py -v || die
+}
+
+python_install_all() {
+       distutils-r1_python_install_all
+
+       if use examples; then
+               dodoc -r contrib examples
+               docompress -x /usr/share/doc/${PF}/{contrib,examples}
+       fi
+}

Reply via email to