This is an automated email from the ASF dual-hosted git repository.

alamb pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/arrow-rs.git


The following commit(s) were added to refs/heads/master by this push:
     new c396dfb50 Update versions and CHANGELOG for `16.0.0` (#1826)
c396dfb50 is described below

commit c396dfb5035d22e57717b6dd365486b76eb611bc
Author: Andrew Lamb <[email protected]>
AuthorDate: Fri Jun 10 13:27:03 2022 -0400

    Update versions and CHANGELOG for `16.0.0` (#1826)
    
    * Update versions to 16.0.0
    
    * Update changelog
    
    * Updates
    
    * updates
    
    * Update for latest
    
    * polish
---
 CHANGELOG-old.md                             |  72 +++++++++++++++++
 CHANGELOG.md                                 | 117 ++++++++++++++-------------
 arrow-flight/Cargo.toml                      |   4 +-
 arrow-flight/README.md                       |   2 +-
 arrow-pyarrow-integration-testing/Cargo.toml |   4 +-
 arrow/Cargo.toml                             |   2 +-
 arrow/README.md                              |   2 +-
 dev/release/README.md                        |   2 +-
 dev/release/update_change_log.sh             |   4 +-
 integration-testing/Cargo.toml               |   2 +-
 parquet/Cargo.toml                           |   6 +-
 parquet_derive/Cargo.toml                    |   4 +-
 parquet_derive/README.md                     |   4 +-
 parquet_derive_test/Cargo.toml               |   6 +-
 14 files changed, 154 insertions(+), 77 deletions(-)

diff --git a/CHANGELOG-old.md b/CHANGELOG-old.md
index 9ec964d86..518697ce0 100644
--- a/CHANGELOG-old.md
+++ b/CHANGELOG-old.md
@@ -17,6 +17,78 @@
   under the License.
 -->
 
+
+## [15.0.0](https://github.com/apache/arrow-rs/tree/15.0.0) (2022-05-27)
+
+[Full Changelog](https://github.com/apache/arrow-rs/compare/14.0.0...15.0.0)
+
+**Breaking changes:**
+
+- Change `ArrayDataBuilder::null_bit_buffer` to accept `Option<Buffer>` rather 
than `Buffer` [\#1739](https://github.com/apache/arrow-rs/pull/1739) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
+- Remove `null_count` from `ArrayData::try_new()` 
[\#1721](https://github.com/apache/arrow-rs/pull/1721) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
+- Change parquet writers to use standard `std:io::Write` rather custom 
`ParquetWriter` trait \(\#1717\) \(\#1163\) 
[\#1719](https://github.com/apache/arrow-rs/pull/1719) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
+- Add explicit column mask for selection in parquet: `ProjectionMask` 
\(\#1701\) [\#1716](https://github.com/apache/arrow-rs/pull/1716) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
+- Add type\_ids in Union datatype 
[\#1703](https://github.com/apache/arrow-rs/pull/1703) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([viirya](https://github.com/viirya))
+- Fix Parquet Reader's Arrow Schema Inference 
[\#1682](https://github.com/apache/arrow-rs/pull/1682) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([tustvold](https://github.com/tustvold))
+
+**Implemented enhancements:**
+
+- Rename the `string` kernel to `concatenate_elements` 
[\#1747](https://github.com/apache/arrow-rs/issues/1747) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- `ArrayDataBuilder::null_bit_buffer` should accept `Option<Buffer>` as input 
type [\#1737](https://github.com/apache/arrow-rs/issues/1737) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Fix schema comparison for non\_canonical\_map when running flight test 
[\#1730](https://github.com/apache/arrow-rs/issues/1730) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Add support in aggregate kernel for `BinaryArray` 
[\#1724](https://github.com/apache/arrow-rs/issues/1724) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Fix incorrect null\_count in `generate_unions_case` integration test 
[\#1712](https://github.com/apache/arrow-rs/issues/1712) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Keep type ids in Union datatype to follow Arrow spec and integrate with 
other implementations [\#1690](https://github.com/apache/arrow-rs/issues/1690) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Support Reading Alternative List Representations to Arrow From Parquet 
[\#1680](https://github.com/apache/arrow-rs/issues/1680) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- Speed up the offsets checking 
[\#1675](https://github.com/apache/arrow-rs/issues/1675) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Separate Parquet -\> Arrow Schema Conversion From ArrayBuilder 
[\#1655](https://github.com/apache/arrow-rs/issues/1655) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- Add `leaf_columns` argument to `ArrowReader::get_record_reader_by_columns` 
[\#1653](https://github.com/apache/arrow-rs/issues/1653) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- Implement `string_concat` kernel  
[\#1540](https://github.com/apache/arrow-rs/issues/1540) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Improve Unit Test Coverage of ArrayReaderBuilder 
[\#1484](https://github.com/apache/arrow-rs/issues/1484) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+
+**Fixed bugs:**
+
+- Parquet write failure \(from record batches\) when data is nested two levels 
deep  [\#1744](https://github.com/apache/arrow-rs/issues/1744) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- IPC reader may break on projection 
[\#1735](https://github.com/apache/arrow-rs/issues/1735) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Latest nightly fails to build with feature simd 
[\#1734](https://github.com/apache/arrow-rs/issues/1734) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Trying to write parquet file in parallel results in corrupt file 
[\#1717](https://github.com/apache/arrow-rs/issues/1717) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- Roundtrip failure when using DELTA\_BINARY\_PACKED 
[\#1708](https://github.com/apache/arrow-rs/issues/1708) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- `ArrayData::try_new` cannot always return expected error. 
[\#1707](https://github.com/apache/arrow-rs/issues/1707) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+-  "out of order projection is not supported" after Fix Parquet Arrow Schema 
Inference [\#1701](https://github.com/apache/arrow-rs/issues/1701) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- Rust is not interoperability with C++ for IPC schemas with dictionaries 
[\#1694](https://github.com/apache/arrow-rs/issues/1694) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Incorrect Repeated Field Schema Inference 
[\#1681](https://github.com/apache/arrow-rs/issues/1681) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- Parquet Treats Embedded Arrow Schema as Authoritative 
[\#1663](https://github.com/apache/arrow-rs/issues/1663) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- parquet\_to\_arrow\_schema\_by\_columns Incorrectly Handles Nested Types 
[\#1654](https://github.com/apache/arrow-rs/issues/1654) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- Inconsistent Arrow Schema When Projecting Nested Parquet File 
[\#1652](https://github.com/apache/arrow-rs/issues/1652) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- StructArrayReader Cannot Handle Nested Lists 
[\#1651](https://github.com/apache/arrow-rs/issues/1651) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- Bug \(`substring` kernel\): The null buffer is not aligned when `offset != 
0` [\#1639](https://github.com/apache/arrow-rs/issues/1639) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+
+**Documentation updates:**
+
+- Parquet command line tool does not install "globally" 
[\#1710](https://github.com/apache/arrow-rs/issues/1710) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- Improve integration test document to follow Arrow C++ repo CI 
[\#1742](https://github.com/apache/arrow-rs/pull/1742) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([viirya](https://github.com/viirya))
+
+**Merged pull requests:**
+
+- Test for list array equality with different offsets 
[\#1756](https://github.com/apache/arrow-rs/pull/1756) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([alamb](https://github.com/alamb))
+- Rename `string_concat` to `concat_elements_utf8` 
[\#1754](https://github.com/apache/arrow-rs/pull/1754) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([alamb](https://github.com/alamb))
+- Rename the `string` kernel to `concat_elements`. 
[\#1752](https://github.com/apache/arrow-rs/pull/1752) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
+- Support writing nested lists to parquet 
[\#1746](https://github.com/apache/arrow-rs/pull/1746) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
+- Pin nightly version to bypass packed\_simd build error 
[\#1743](https://github.com/apache/arrow-rs/pull/1743) 
([viirya](https://github.com/viirya))
+- Fix projection in IPC reader 
[\#1736](https://github.com/apache/arrow-rs/pull/1736) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([iyupeng](https://github.com/iyupeng))
+- `cargo install` installs not globally 
[\#1732](https://github.com/apache/arrow-rs/pull/1732) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([kazuk](https://github.com/kazuk))
+- Fix schema comparison for non\_canonical\_map when running flight test 
[\#1731](https://github.com/apache/arrow-rs/pull/1731) 
([viirya](https://github.com/viirya))
+- Add `min_binary` and `max_binary` aggregate kernels 
[\#1725](https://github.com/apache/arrow-rs/pull/1725) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
+- Fix parquet benchmarks 
[\#1723](https://github.com/apache/arrow-rs/pull/1723) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
+- Fix BitReader::get\_batch zero extension \(\#1708\) 
[\#1722](https://github.com/apache/arrow-rs/pull/1722) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
+- Implementation string concat 
[\#1720](https://github.com/apache/arrow-rs/pull/1720) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([Ismail-Maj](https://github.com/Ismail-Maj))
+- Check the length of `null_bit_buffer` in `ArrayData::try_new()` 
[\#1714](https://github.com/apache/arrow-rs/pull/1714) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
+- Fix incorrect null\_count in `generate_unions_case` integration test 
[\#1713](https://github.com/apache/arrow-rs/pull/1713) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([viirya](https://github.com/viirya))
+- Fix: Null buffer accounts for `offset` in `substring` kernel. 
[\#1704](https://github.com/apache/arrow-rs/pull/1704) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
+- Minor: Refine `OffsetSizeTrait` to extend `num::Integer`  
[\#1702](https://github.com/apache/arrow-rs/pull/1702) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
+- Fix StructArrayReader handling nested lists \(\#1651\)  
[\#1700](https://github.com/apache/arrow-rs/pull/1700) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
+- Speed up the offsets checking 
[\#1684](https://github.com/apache/arrow-rs/pull/1684) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
+
 ## [14.0.0](https://github.com/apache/arrow-rs/tree/14.0.0) (2022-05-13)
 
 [Full Changelog](https://github.com/apache/arrow-rs/compare/13.0.0...14.0.0)
diff --git a/CHANGELOG.md b/CHANGELOG.md
index 14d2a3c6e..549d4da1a 100644
--- a/CHANGELOG.md
+++ b/CHANGELOG.md
@@ -19,77 +19,82 @@
 
 # Changelog
 
-## [15.0.0](https://github.com/apache/arrow-rs/tree/15.0.0) (2022-05-27)
+## [16.0.0](https://github.com/apache/arrow-rs/tree/16.0.0) (2022-06-10)
 
-[Full Changelog](https://github.com/apache/arrow-rs/compare/14.0.0...15.0.0)
+[Full Changelog](https://github.com/apache/arrow-rs/compare/15.0.0...16.0.0)
 
 **Breaking changes:**
 
-- Change `ArrayDataBuilder::null_bit_buffer` to accept `Option<Buffer>` rather 
than `Buffer` [\#1739](https://github.com/apache/arrow-rs/pull/1739) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
-- Remove `null_count` from `ArrayData::try_new()` 
[\#1721](https://github.com/apache/arrow-rs/pull/1721) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
-- Change parquet writers to use standard `std:io::Write` rather custom 
`ParquetWriter` trait \(\#1717\) \(\#1163\) 
[\#1719](https://github.com/apache/arrow-rs/pull/1719) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
-- Add explicit column mask for selection in parquet: `ProjectionMask` 
\(\#1701\) [\#1716](https://github.com/apache/arrow-rs/pull/1716) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
-- Add type\_ids in Union datatype 
[\#1703](https://github.com/apache/arrow-rs/pull/1703) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([viirya](https://github.com/viirya))
-- Fix Parquet Reader's Arrow Schema Inference 
[\#1682](https://github.com/apache/arrow-rs/pull/1682) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([tustvold](https://github.com/tustvold))
+- Seal `ArrowNativeType` and `OffsetSizeTrait` for safety \(\#1028\) 
[\#1819](https://github.com/apache/arrow-rs/pull/1819) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([tustvold](https://github.com/tustvold))
+- Improve API for `csv::infer_file_schema` by removing redundant ref  
[\#1776](https://github.com/apache/arrow-rs/pull/1776) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([tustvold](https://github.com/tustvold))
 
 **Implemented enhancements:**
 
-- Rename the `string` kernel to `concatenate_elements` 
[\#1747](https://github.com/apache/arrow-rs/issues/1747) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
-- `ArrayDataBuilder::null_bit_buffer` should accept `Option<Buffer>` as input 
type [\#1737](https://github.com/apache/arrow-rs/issues/1737) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
-- Fix schema comparison for non\_canonical\_map when running flight test 
[\#1730](https://github.com/apache/arrow-rs/issues/1730) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
-- Add support in aggregate kernel for `BinaryArray` 
[\#1724](https://github.com/apache/arrow-rs/issues/1724) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
-- Fix incorrect null\_count in `generate_unions_case` integration test 
[\#1712](https://github.com/apache/arrow-rs/issues/1712) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
-- Keep type ids in Union datatype to follow Arrow spec and integrate with 
other implementations [\#1690](https://github.com/apache/arrow-rs/issues/1690) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
-- Support Reading Alternative List Representations to Arrow From Parquet 
[\#1680](https://github.com/apache/arrow-rs/issues/1680) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
-- Speed up the offsets checking 
[\#1675](https://github.com/apache/arrow-rs/issues/1675) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
-- Separate Parquet -\> Arrow Schema Conversion From ArrayBuilder 
[\#1655](https://github.com/apache/arrow-rs/issues/1655) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
-- Add `leaf_columns` argument to `ArrowReader::get_record_reader_by_columns` 
[\#1653](https://github.com/apache/arrow-rs/issues/1653) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
-- Implement `string_concat` kernel  
[\#1540](https://github.com/apache/arrow-rs/issues/1540) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
-- Improve Unit Test Coverage of ArrayReaderBuilder 
[\#1484](https://github.com/apache/arrow-rs/issues/1484) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- List equality method should work on empty offset `ListArray` 
[\#1817](https://github.com/apache/arrow-rs/issues/1817) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Command line tool for convert CSV to Parquet 
[\#1797](https://github.com/apache/arrow-rs/issues/1797) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- IPC writer should write validity buffer for `UnionArray` in V4 IPC message 
[\#1793](https://github.com/apache/arrow-rs/issues/1793) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Add function for row alignment with page mask 
[\#1790](https://github.com/apache/arrow-rs/issues/1790) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- Rust IPC Read should be able to read V4 UnionType Array 
[\#1788](https://github.com/apache/arrow-rs/issues/1788) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- `combine_option_bitmap` should accept arbitrary number of input arrays. 
[\#1780](https://github.com/apache/arrow-rs/issues/1780) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Add `substring_by_char` kernels for slicing on character boundaries 
[\#1768](https://github.com/apache/arrow-rs/issues/1768) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Support reading `PageIndex` from column metadata 
[\#1761](https://github.com/apache/arrow-rs/issues/1761) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- Support casting from `DataType::Utf8` to `DataType::Boolean` 
[\#1740](https://github.com/apache/arrow-rs/issues/1740) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Make current position available in `FileWriter`. 
[\#1691](https://github.com/apache/arrow-rs/issues/1691) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- Support writing parquet to `stdout` 
[\#1687](https://github.com/apache/arrow-rs/issues/1687) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
 
 **Fixed bugs:**
 
-- Parquet write failure \(from record batches\) when data is nested two levels 
deep  [\#1744](https://github.com/apache/arrow-rs/issues/1744) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
-- IPC reader may break on projection 
[\#1735](https://github.com/apache/arrow-rs/issues/1735) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
-- Latest nightly fails to build with feature simd 
[\#1734](https://github.com/apache/arrow-rs/issues/1734) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
-- Trying to write parquet file in parallel results in corrupt file 
[\#1717](https://github.com/apache/arrow-rs/issues/1717) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
-- Roundtrip failure when using DELTA\_BINARY\_PACKED 
[\#1708](https://github.com/apache/arrow-rs/issues/1708) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
-- `ArrayData::try_new` cannot always return expected error. 
[\#1707](https://github.com/apache/arrow-rs/issues/1707) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
--  "out of order projection is not supported" after Fix Parquet Arrow Schema 
Inference [\#1701](https://github.com/apache/arrow-rs/issues/1701) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
-- Rust is not interoperability with C++ for IPC schemas with dictionaries 
[\#1694](https://github.com/apache/arrow-rs/issues/1694) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
-- Incorrect Repeated Field Schema Inference 
[\#1681](https://github.com/apache/arrow-rs/issues/1681) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
-- Parquet Treats Embedded Arrow Schema as Authoritative 
[\#1663](https://github.com/apache/arrow-rs/issues/1663) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
-- parquet\_to\_arrow\_schema\_by\_columns Incorrectly Handles Nested Types 
[\#1654](https://github.com/apache/arrow-rs/issues/1654) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
-- Inconsistent Arrow Schema When Projecting Nested Parquet File 
[\#1652](https://github.com/apache/arrow-rs/issues/1652) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
-- StructArrayReader Cannot Handle Nested Lists 
[\#1651](https://github.com/apache/arrow-rs/issues/1651) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
-- Bug \(`substring` kernel\): The null buffer is not aligned when `offset != 
0` [\#1639](https://github.com/apache/arrow-rs/issues/1639) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Incorrect Offset Validation for Sliced List Array Children 
[\#1814](https://github.com/apache/arrow-rs/issues/1814) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Parquet Snappy Codec overwrites Existing Data in Decompression Buffer 
[\#1806](https://github.com/apache/arrow-rs/issues/1806) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
+- `flight_data_to_arrow_batch` does not support `RecordBatch`es with no 
columns [\#1783](https://github.com/apache/arrow-rs/issues/1783) 
[[arrow-flight](https://github.com/apache/arrow-rs/labels/arrow-flight)]
+- parquet does not compile with `features=["zstd"]` 
[\#1630](https://github.com/apache/arrow-rs/issues/1630) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
 
 **Documentation updates:**
 
-- Parquet command line tool does not install "globally" 
[\#1710](https://github.com/apache/arrow-rs/issues/1710) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)]
-- Improve integration test document to follow Arrow C++ repo CI 
[\#1742](https://github.com/apache/arrow-rs/pull/1742) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([viirya](https://github.com/viirya))
+- Update arrow module docs 
[\#1840](https://github.com/apache/arrow-rs/pull/1840) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([tustvold](https://github.com/tustvold))
+- Update safety disclaimer 
[\#1837](https://github.com/apache/arrow-rs/pull/1837) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([tustvold](https://github.com/tustvold))
+- Update ballista readme link 
[\#1765](https://github.com/apache/arrow-rs/pull/1765) 
([tustvold](https://github.com/tustvold))
+- Move changelog archive to `CHANGELOG-old.md` 
[\#1759](https://github.com/apache/arrow-rs/pull/1759) 
([alamb](https://github.com/alamb))
+
+**Closed issues:**
+
+- `DataType::Decimal` Non-Compliant? 
[\#1779](https://github.com/apache/arrow-rs/issues/1779) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Further simplify the offset validation 
[\#1770](https://github.com/apache/arrow-rs/issues/1770) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Best way to convert arrow to Rust native type 
[\#1760](https://github.com/apache/arrow-rs/issues/1760) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
+- Why `Parquet` is a part of `Arrow`? 
[\#1715](https://github.com/apache/arrow-rs/issues/1715) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)]
 
 **Merged pull requests:**
 
-- Test for list array equality with different offsets 
[\#1756](https://github.com/apache/arrow-rs/pull/1756) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([alamb](https://github.com/alamb))
-- Rename `string_concat` to `concat_elements_utf8` 
[\#1754](https://github.com/apache/arrow-rs/pull/1754) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([alamb](https://github.com/alamb))
-- Rename the `string` kernel to `concat_elements`. 
[\#1752](https://github.com/apache/arrow-rs/pull/1752) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
-- Support writing nested lists to parquet 
[\#1746](https://github.com/apache/arrow-rs/pull/1746) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
-- Pin nightly version to bypass packed\_simd build error 
[\#1743](https://github.com/apache/arrow-rs/pull/1743) 
([viirya](https://github.com/viirya))
-- Fix projection in IPC reader 
[\#1736](https://github.com/apache/arrow-rs/pull/1736) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([iyupeng](https://github.com/iyupeng))
-- `cargo install` installs not globally 
[\#1732](https://github.com/apache/arrow-rs/pull/1732) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([kazuk](https://github.com/kazuk))
-- Fix schema comparison for non\_canonical\_map when running flight test 
[\#1731](https://github.com/apache/arrow-rs/pull/1731) 
([viirya](https://github.com/viirya))
-- Add `min_binary` and `max_binary` aggregate kernels 
[\#1725](https://github.com/apache/arrow-rs/pull/1725) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
-- Fix parquet benchmarks 
[\#1723](https://github.com/apache/arrow-rs/pull/1723) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
-- Fix BitReader::get\_batch zero extension \(\#1708\) 
[\#1722](https://github.com/apache/arrow-rs/pull/1722) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
-- Implementation string concat 
[\#1720](https://github.com/apache/arrow-rs/pull/1720) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([Ismail-Maj](https://github.com/Ismail-Maj))
-- Check the length of `null_bit_buffer` in `ArrayData::try_new()` 
[\#1714](https://github.com/apache/arrow-rs/pull/1714) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
-- Fix incorrect null\_count in `generate_unions_case` integration test 
[\#1713](https://github.com/apache/arrow-rs/pull/1713) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([viirya](https://github.com/viirya))
-- Fix: Null buffer accounts for `offset` in `substring` kernel. 
[\#1704](https://github.com/apache/arrow-rs/pull/1704) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
-- Minor: Refine `OffsetSizeTrait` to extend `num::Integer`  
[\#1702](https://github.com/apache/arrow-rs/pull/1702) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
-- Fix StructArrayReader handling nested lists \(\#1651\)  
[\#1700](https://github.com/apache/arrow-rs/pull/1700) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
-- Speed up the offsets checking 
[\#1684](https://github.com/apache/arrow-rs/pull/1684) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
-
-For older versions, see [CHANGELOG-old](CHANGELOG-old.md) and 
[apache/arrow/CHANGELOG.md](https://github.com/apache/arrow/blob/master/CHANGELOG.md)
+- Make equals\_datatype method public, enabling other modules 
[\#1838](https://github.com/apache/arrow-rs/pull/1838) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([nl5887](https://github.com/nl5887))
+- \[Minor\] Clarify `PageIterator` Documentation 
[\#1831](https://github.com/apache/arrow-rs/pull/1831) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([Ted-Jiang](https://github.com/Ted-Jiang))
+- Update MIRI pin [\#1828](https://github.com/apache/arrow-rs/pull/1828) 
([tustvold](https://github.com/tustvold))
+- Change to use `resolver v2`, test more feature flag combinations in CI, fix 
errors \(\#1630\) [\#1822](https://github.com/apache/arrow-rs/pull/1822) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([tustvold](https://github.com/tustvold))
+- Add ScalarBuffer abstraction \(\#1811\) 
[\#1820](https://github.com/apache/arrow-rs/pull/1820) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([tustvold](https://github.com/tustvold))
+- Fix list equal for empty offset list array 
[\#1818](https://github.com/apache/arrow-rs/pull/1818) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([viirya](https://github.com/viirya))
+- Fix Decimal and List ArrayData Validation \(\#1813\) \(\#1814\) 
[\#1816](https://github.com/apache/arrow-rs/pull/1816) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([tustvold](https://github.com/tustvold))
+- Don't overwrite existing data on snappy decompress \(\#1806\) 
[\#1807](https://github.com/apache/arrow-rs/pull/1807) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
+- Rename `arrow/benches/string_kernels.rs` to 
`arrow/benches/substring_kernels.rs` 
[\#1805](https://github.com/apache/arrow-rs/pull/1805) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
+- Add public API for decoding parquet footer 
[\#1804](https://github.com/apache/arrow-rs/pull/1804) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
+- Add AsyncFileReader trait 
[\#1803](https://github.com/apache/arrow-rs/pull/1803) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
+- add parquet-fromcsv \(\#1\) 
[\#1798](https://github.com/apache/arrow-rs/pull/1798) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([kazuk](https://github.com/kazuk))
+- Use IPC row count info in IPC reader 
[\#1796](https://github.com/apache/arrow-rs/pull/1796) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([viirya](https://github.com/viirya))
+- Fix typos in the Memory and Buffers section of the docs home 
[\#1795](https://github.com/apache/arrow-rs/pull/1795) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([datapythonista](https://github.com/datapythonista))
+- Write validity buffer for UnionArray in V4 IPC message 
[\#1794](https://github.com/apache/arrow-rs/pull/1794) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([viirya](https://github.com/viirya))
+- feat:Add function for row alignment with page mask 
[\#1791](https://github.com/apache/arrow-rs/pull/1791) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([Ted-Jiang](https://github.com/Ted-Jiang))
+- Read and skip validity buffer of UnionType Array for V4 ipc message 
[\#1789](https://github.com/apache/arrow-rs/pull/1789) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
[[arrow-flight](https://github.com/apache/arrow-rs/labels/arrow-flight)] 
([viirya](https://github.com/viirya))
+- Add `Substring_by_char` 
[\#1784](https://github.com/apache/arrow-rs/pull/1784) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
+- Add `ParquetFileArrowReader::try_new` 
[\#1782](https://github.com/apache/arrow-rs/pull/1782) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
+- Arbitrary size combine option bitmap 
[\#1781](https://github.com/apache/arrow-rs/pull/1781) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([Ismail-Maj](https://github.com/Ismail-Maj))
+- Implement `ChunkReader` for `Bytes`, deprecate `SliceableCursor` 
[\#1775](https://github.com/apache/arrow-rs/pull/1775) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
+- Access metadata of flushed row groups on write \(\#1691\) 
[\#1774](https://github.com/apache/arrow-rs/pull/1774) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
+- Simplify ParquetFileArrowReader Metadata API 
[\#1773](https://github.com/apache/arrow-rs/pull/1773) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([tustvold](https://github.com/tustvold))
+- MINOR: Unpin nightly version as packed\_simd releases new version 
[\#1771](https://github.com/apache/arrow-rs/pull/1771) 
([viirya](https://github.com/viirya))
+- Update comfy-table requirement from 5.0 to 6.0 
[\#1769](https://github.com/apache/arrow-rs/pull/1769) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([dependabot[bot]](https://github.com/apps/dependabot))
+- Optionally disable `validate_decimal_precision` check in 
`DecimalBuilder.append_value` for interop test 
[\#1767](https://github.com/apache/arrow-rs/pull/1767) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([viirya](https://github.com/viirya))
+- Minor: Clean up the code of MutableArrayData 
[\#1763](https://github.com/apache/arrow-rs/pull/1763) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([HaoYang670](https://github.com/HaoYang670))
+- Support reading PageIndex from parquet metadata, prepare for skipping pages 
at reading [\#1762](https://github.com/apache/arrow-rs/pull/1762) 
[[parquet](https://github.com/apache/arrow-rs/labels/parquet)] 
([Ted-Jiang](https://github.com/Ted-Jiang))
+- Support casting `Utf8` to `Boolean` 
[\#1738](https://github.com/apache/arrow-rs/pull/1738) 
[[arrow](https://github.com/apache/arrow-rs/labels/arrow)] 
([MazterQyou](https://github.com/MazterQyou))
+
+
 
 \* *This Changelog was automatically generated by 
[github_changelog_generator](https://github.com/github-changelog-generator/github-changelog-generator)*
diff --git a/arrow-flight/Cargo.toml b/arrow-flight/Cargo.toml
index 19215cd3d..f64ca7d9e 100644
--- a/arrow-flight/Cargo.toml
+++ b/arrow-flight/Cargo.toml
@@ -18,7 +18,7 @@
 [package]
 name = "arrow-flight"
 description = "Apache Arrow Flight"
-version = "15.0.0"
+version = "16.0.0"
 edition = "2021"
 rust-version = "1.57"
 authors = ["Apache Arrow <[email protected]>"]
@@ -27,7 +27,7 @@ repository = "https://github.com/apache/arrow-rs";
 license = "Apache-2.0"
 
 [dependencies]
-arrow = { path = "../arrow", version = "15.0.0" }
+arrow = { path = "../arrow", version = "16.0.0" }
 base64 = "0.13"
 tonic = "0.7"
 bytes = "1"
diff --git a/arrow-flight/README.md b/arrow-flight/README.md
index 45b081799..a951699f4 100644
--- a/arrow-flight/README.md
+++ b/arrow-flight/README.md
@@ -27,7 +27,7 @@ Add this to your Cargo.toml:
 
 ```toml
 [dependencies]
-arrow-flight = "15.0.0"
+arrow-flight = "16.0.0"
 ```
 
 Apache Arrow Flight is a gRPC based protocol for exchanging Arrow data between 
processes. See the blog post [Introducing Apache Arrow Flight: A Framework for 
Fast Data 
Transport](https://arrow.apache.org/blog/2019/10/13/introducing-arrow-flight/) 
for more information.
diff --git a/arrow-pyarrow-integration-testing/Cargo.toml 
b/arrow-pyarrow-integration-testing/Cargo.toml
index cba15fc61..58ba72609 100644
--- a/arrow-pyarrow-integration-testing/Cargo.toml
+++ b/arrow-pyarrow-integration-testing/Cargo.toml
@@ -18,7 +18,7 @@
 [package]
 name = "arrow-pyarrow-integration-testing"
 description = ""
-version = "15.0.0"
+version = "16.0.0"
 homepage = "https://github.com/apache/arrow-rs";
 repository = "https://github.com/apache/arrow-rs";
 authors = ["Apache Arrow <[email protected]>"]
@@ -32,7 +32,7 @@ name = "arrow_pyarrow_integration_testing"
 crate-type = ["cdylib"]
 
 [dependencies]
-arrow = { path = "../arrow", version = "15.0.0", features = ["pyarrow"] }
+arrow = { path = "../arrow", version = "16.0.0", features = ["pyarrow"] }
 pyo3 = { version = "0.16", features = ["extension-module"] }
 
 [package.metadata.maturin]
diff --git a/arrow/Cargo.toml b/arrow/Cargo.toml
index fd60078f1..ebcdd9e7a 100644
--- a/arrow/Cargo.toml
+++ b/arrow/Cargo.toml
@@ -17,7 +17,7 @@
 
 [package]
 name = "arrow"
-version = "15.0.0"
+version = "16.0.0"
 description = "Rust implementation of Apache Arrow"
 homepage = "https://github.com/apache/arrow-rs";
 repository = "https://github.com/apache/arrow-rs";
diff --git a/arrow/README.md b/arrow/README.md
index 729f5a29b..67de57ff0 100644
--- a/arrow/README.md
+++ b/arrow/README.md
@@ -32,7 +32,7 @@ This crate is tested with the latest stable version of Rust. 
We do not currently
 
 The arrow crate follows the [SemVer 
standard](https://doc.rust-lang.org/cargo/reference/semver.html) defined by 
Cargo and works well within the Rust crate ecosystem.
 
-However, for historical reasons, this crate uses versions with major numbers 
greater than `0.x` (e.g. `15.0.0`), unlike many other crates in the Rust 
ecosystem which spend extended time releasing versions `0.x` to signal planned 
ongoing API changes. Minor arrow releases contain only compatible changes, 
while major releases may contain breaking API changes.
+However, for historical reasons, this crate uses versions with major numbers 
greater than `0.x` (e.g. `16.0.0`), unlike many other crates in the Rust 
ecosystem which spend extended time releasing versions `0.x` to signal planned 
ongoing API changes. Minor arrow releases contain only compatible changes, 
while major releases may contain breaking API changes.
 
 ## Features
 
diff --git a/dev/release/README.md b/dev/release/README.md
index 96c730a63..912b60dae 100644
--- a/dev/release/README.md
+++ b/dev/release/README.md
@@ -61,7 +61,7 @@ CHANGELOG_GITHUB_TOKEN=<TOKEN> 
./dev/release/update_change_log.sh
 git commit -a -m 'Create changelog'
 
 # update versions
-sed -i '' -e 's/14.0.0/15.0.0/g' `find . -name 'Cargo.toml' -or -name '*.md' | 
grep -v CHANGELOG.md`
+sed -i '' -e 's/14.0.0/16.0.0/g' `find . -name 'Cargo.toml' -or -name '*.md' | 
grep -v CHANGELOG.md`
 git commit -a -m 'Update version'
 ```
 
diff --git a/dev/release/update_change_log.sh b/dev/release/update_change_log.sh
index 43b4ee606..316f10c25 100755
--- a/dev/release/update_change_log.sh
+++ b/dev/release/update_change_log.sh
@@ -29,8 +29,8 @@
 
 set -e
 
-SINCE_TAG="14.0.0"
-FUTURE_RELEASE="15.0.0"
+SINCE_TAG="15.0.0"
+FUTURE_RELEASE="16.0.0"
 
 SOURCE_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
 SOURCE_TOP_DIR="$(cd "${SOURCE_DIR}/../../" && pwd)"
diff --git a/integration-testing/Cargo.toml b/integration-testing/Cargo.toml
index 7215c31d2..253c9a25e 100644
--- a/integration-testing/Cargo.toml
+++ b/integration-testing/Cargo.toml
@@ -18,7 +18,7 @@
 [package]
 name = "arrow-integration-testing"
 description = "Binaries used in the Arrow integration tests"
-version = "15.0.0"
+version = "16.0.0"
 homepage = "https://github.com/apache/arrow-rs";
 repository = "https://github.com/apache/arrow-rs";
 authors = ["Apache Arrow <[email protected]>"]
diff --git a/parquet/Cargo.toml b/parquet/Cargo.toml
index ced22fc52..66e2797b2 100644
--- a/parquet/Cargo.toml
+++ b/parquet/Cargo.toml
@@ -17,7 +17,7 @@
 
 [package]
 name = "parquet"
-version = "15.0.0"
+version = "16.0.0"
 license = "Apache-2.0"
 description = "Apache Parquet implementation in Rust"
 homepage = "https://github.com/apache/arrow-rs";
@@ -42,7 +42,7 @@ zstd = { version = "0.11.1", optional = true, 
default-features = false }
 chrono = { version = "0.4", default-features = false, features = ["alloc"] }
 num = "0.4"
 num-bigint = "0.4"
-arrow = { path = "../arrow", version = "15.0.0", optional = true, 
default-features = false, features = ["ipc"] }
+arrow = { path = "../arrow", version = "16.0.0", optional = true, 
default-features = false, features = ["ipc"] }
 base64 = { version = "0.13", optional = true }
 clap = { version = "3", optional = true, features = ["derive", "env"] }
 serde_json = { version = "1.0", optional = true }
@@ -61,7 +61,7 @@ flate2 = "1.0"
 lz4 = "1.23"
 zstd = "0.11"
 serde_json = { version = "1.0", features = ["preserve_order"] }
-arrow = { path = "../arrow", version = "15.0.0", default-features = false, 
features = ["ipc", "test_utils", "prettyprint"] }
+arrow = { path = "../arrow", version = "16.0.0", default-features = false, 
features = ["ipc", "test_utils", "prettyprint"] }
 
 [package.metadata.docs.rs]
 all-features = true
diff --git a/parquet_derive/Cargo.toml b/parquet_derive/Cargo.toml
index bab5d230a..680074d08 100644
--- a/parquet_derive/Cargo.toml
+++ b/parquet_derive/Cargo.toml
@@ -17,7 +17,7 @@
 
 [package]
 name = "parquet_derive"
-version = "15.0.0"
+version = "16.0.0"
 license = "Apache-2.0"
 description = "Derive macros for the Rust implementation of Apache Parquet"
 homepage = "https://github.com/apache/arrow-rs";
@@ -35,4 +35,4 @@ proc-macro = true
 proc-macro2 = "1.0"
 quote = "1.0"
 syn = { version = "1.0", features = ["full", "extra-traits"] }
-parquet = { path = "../parquet", version = "15.0.0" }
+parquet = { path = "../parquet", version = "16.0.0" }
diff --git a/parquet_derive/README.md b/parquet_derive/README.md
index 61ccf3093..4f390b0cd 100644
--- a/parquet_derive/README.md
+++ b/parquet_derive/README.md
@@ -32,8 +32,8 @@ Add this to your Cargo.toml:
 
 ```toml
 [dependencies]
-parquet = "15.0.0"
-parquet_derive = "15.0.0"
+parquet = "16.0.0"
+parquet_derive = "16.0.0"
 ```
 
 and this to your crate root:
diff --git a/parquet_derive_test/Cargo.toml b/parquet_derive_test/Cargo.toml
index 53daf2977..7bf6db673 100644
--- a/parquet_derive_test/Cargo.toml
+++ b/parquet_derive_test/Cargo.toml
@@ -17,7 +17,7 @@
 
 [package]
 name = "parquet_derive_test"
-version = "15.0.0"
+version = "16.0.0"
 license = "Apache-2.0"
 description = "Integration test package for parquet-derive"
 homepage = "https://github.com/apache/arrow-rs";
@@ -29,6 +29,6 @@ publish = false
 rust-version = "1.57"
 
 [dependencies]
-parquet = { path = "../parquet", version = "15.0.0" }
-parquet_derive = { path = "../parquet_derive", version = "15.0.0" }
+parquet = { path = "../parquet", version = "16.0.0" }
+parquet_derive = { path = "../parquet_derive", version = "16.0.0" }
 chrono = "0.4.19"

Reply via email to