This is an automated email from the ASF dual-hosted git repository.

westonpace pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow.git


The following commit(s) were added to refs/heads/main by this push:
     new 3e123ad6e5 GH-38960: [C++] Fix spelling (acero) (#38961)
3e123ad6e5 is described below

commit 3e123ad6e503b486cc95503827099ec11b17132c
Author: Josh Soref <[email protected]>
AuthorDate: Tue Dec 5 13:57:41 2023 -0500

    GH-38960: [C++] Fix spelling (acero) (#38961)
    
    
    
    ### Rationale for this change
    
    ### What changes are included in this PR?
    
    Spelling fixes to cpp/src/arrow/acero/
    
    ### Are these changes tested?
    
    ### Are there any user-facing changes?
    
    * Closes: #38960
    
    Authored-by: Josh Soref <[email protected]>
    Signed-off-by: Weston Pace <[email protected]>
---
 cpp/src/arrow/acero/accumulation_queue.h    | 2 +-
 cpp/src/arrow/acero/aggregate_internal.h    | 2 +-
 cpp/src/arrow/acero/asof_join_node.cc       | 4 ++--
 cpp/src/arrow/acero/asof_join_node_test.cc  | 4 ++--
 cpp/src/arrow/acero/expression_benchmark.cc | 4 ++--
 cpp/src/arrow/acero/hash_aggregate_test.cc  | 2 +-
 cpp/src/arrow/acero/partition_util.h        | 4 ++--
 cpp/src/arrow/acero/pivot_longer_node.cc    | 2 +-
 cpp/src/arrow/acero/sorted_merge_node.cc    | 2 +-
 cpp/src/arrow/acero/swiss_join.cc           | 8 ++++----
 cpp/src/arrow/acero/swiss_join_internal.h   | 4 ++--
 11 files changed, 19 insertions(+), 19 deletions(-)

diff --git a/cpp/src/arrow/acero/accumulation_queue.h 
b/cpp/src/arrow/acero/accumulation_queue.h
index 285790207f..a27b8b399c 100644
--- a/cpp/src/arrow/acero/accumulation_queue.h
+++ b/cpp/src/arrow/acero/accumulation_queue.h
@@ -82,7 +82,7 @@ class SequencingQueue {
     /// This method will be called on each batch in order.  Calls to this 
method
     /// will be serialized and it will not be called reentrantly.  This makes 
it
     /// safe to do things that rely on order but minimal time should be spent 
here
-    /// to avoid becoming a bottlneck.
+    /// to avoid becoming a bottleneck.
     ///
     /// \return a follow-up task that will be scheduled.  The follow-up 
task(s) are
     ///         is not guaranteed to run in any particular order.  If nullopt 
is
diff --git a/cpp/src/arrow/acero/aggregate_internal.h 
b/cpp/src/arrow/acero/aggregate_internal.h
index 72537a7f7e..5730d99f93 100644
--- a/cpp/src/arrow/acero/aggregate_internal.h
+++ b/cpp/src/arrow/acero/aggregate_internal.h
@@ -224,7 +224,7 @@ class ScalarAggregateNode : public ExecNode, public 
TracedNode {
   // Field indices corresponding to the segment-keys
   const std::vector<int> segment_field_ids_;
   // Holds the value of segment keys of the most recent input batch
-  // The values are updated everytime an input batch is processed
+  // The values are updated every time an input batch is processed
   std::vector<Datum> segmenter_values_;
 
   const std::vector<std::vector<int>> target_fieldsets_;
diff --git a/cpp/src/arrow/acero/asof_join_node.cc 
b/cpp/src/arrow/acero/asof_join_node.cc
index 4a3b6b199c..2609905a0b 100644
--- a/cpp/src/arrow/acero/asof_join_node.cc
+++ b/cpp/src/arrow/acero/asof_join_node.cc
@@ -472,7 +472,7 @@ class BackpressureController : public BackpressureControl {
 };
 
 class InputState {
-  // InputState correponds to an input
+  // InputState corresponds to an input
   // Input record batches are queued up in InputState until processed and
   // turned into output record batches.
 
@@ -1453,7 +1453,7 @@ class AsofJoinNode : public ExecNode {
   bool must_hash_;
   bool may_rehash_;
   // InputStates
-  // Each input state correponds to an input table
+  // Each input state corresponds to an input table
   std::vector<std::unique_ptr<InputState>> state_;
   std::mutex gate_;
   TolType tolerance_;
diff --git a/cpp/src/arrow/acero/asof_join_node_test.cc 
b/cpp/src/arrow/acero/asof_join_node_test.cc
index df3172b2a0..e400cc0316 100644
--- a/cpp/src/arrow/acero/asof_join_node_test.cc
+++ b/cpp/src/arrow/acero/asof_join_node_test.cc
@@ -604,7 +604,7 @@ struct BasicTest {
     auto r0_types = init_types(all_types, [](T& t) { return t->byte_width() > 
1; });
     auto r1_types = init_types(all_types, [](T& t) { return t->byte_width() > 
1; });
 
-    // sample a limited number of type-combinations to keep the runnning time 
reasonable
+    // sample a limited number of type-combinations to keep the running time 
reasonable
     // the scoped-traces below help reproduce a test failure, should it happen
     auto start_time = std::chrono::system_clock::now();
     auto seed = start_time.time_since_epoch().count();
@@ -1279,7 +1279,7 @@ TRACED_TEST(AsofJoinTest, TestUnsupportedOntype, {
                                field("r0_v0", float32())}));
 })
 
-TRACED_TEST(AsofJoinTest, TestUnsupportedBytype, {
+TRACED_TEST(AsofJoinTest, TestUnsupportedByType, {
   DoRunInvalidTypeTest(schema({field("time", int64()), field("key", 
list(int32())),
                                field("l_v0", float64())}),
                        schema({field("time", int64()), field("key", 
list(int32())),
diff --git a/cpp/src/arrow/acero/expression_benchmark.cc 
b/cpp/src/arrow/acero/expression_benchmark.cc
index 9799446ed6..a57dd6b9e3 100644
--- a/cpp/src/arrow/acero/expression_benchmark.cc
+++ b/cpp/src/arrow/acero/expression_benchmark.cc
@@ -107,7 +107,7 @@ static void 
ExecuteScalarExpressionOverhead(benchmark::State& state, Expression
 }
 
 /// \brief Baseline benchmarks are implemented in pure C++ without arrow for 
performance
-/// comparision.
+/// comparison.
 template <typename BenchmarkType>
 void ExecuteScalarExpressionBaseline(benchmark::State& state) {
   const auto rows_per_batch = static_cast<int32_t>(state.range(0));
@@ -193,7 +193,7 @@ BENCHMARK_CAPTURE(BindAndEvaluate, nested_scalar,
 struct ComplexExpressionBaseline {
  public:
   ComplexExpressionBaseline(size_t input_size) {
-    /* hack - cuts off a few elemets if the input size is not a multiple of 64 
for
+    /* hack - cuts off a few elements if the input size is not a multiple of 
64 for
      * simplicity. We can't use std::vector<bool> here since it slows down 
things
      * massively */
     less_20.resize(input_size / 64);
diff --git a/cpp/src/arrow/acero/hash_aggregate_test.cc 
b/cpp/src/arrow/acero/hash_aggregate_test.cc
index 02e67927cc..a4874f3581 100644
--- a/cpp/src/arrow/acero/hash_aggregate_test.cc
+++ b/cpp/src/arrow/acero/hash_aggregate_test.cc
@@ -261,7 +261,7 @@ Result<Datum> MakeGroupByOutput(const 
std::vector<ExecBatch>& output_batches,
     return struct_arr;
   }
 
-  // The exec plan may reorder the output rows.  The tests are all setup to 
expect ouptut
+  // The exec plan may reorder the output rows.  The tests are all setup to 
expect output
   // in ascending order of keys.  So we need to sort the result by the key 
columns.  To do
   // that we create a table using the key columns, calculate the sort indices 
from that
   // table (sorting on all fields) and then use those indices to calculate our 
result.
diff --git a/cpp/src/arrow/acero/partition_util.h 
b/cpp/src/arrow/acero/partition_util.h
index 27cde61d58..1413a8326a 100644
--- a/cpp/src/arrow/acero/partition_util.h
+++ b/cpp/src/arrow/acero/partition_util.h
@@ -33,11 +33,11 @@ class PartitionSort {
  public:
   /// \brief Bucket sort rows on partition ids in O(num_rows) time.
   ///
-  /// Include in the output exclusive cummulative sum of bucket sizes.
+  /// Include in the output exclusive cumulative sum of bucket sizes.
   /// This corresponds to ranges in the sorted array containing all row ids for
   /// each of the partitions.
   ///
-  /// prtn_ranges must be initailized and have at least prtn_ranges + 1 
elements
+  /// prtn_ranges must be initialized and have at least prtn_ranges + 1 
elements
   /// when this method returns prtn_ranges[i] will contains the total number of
   /// elements in partitions 0 through i.  prtn_ranges[0] will be 0.
   ///
diff --git a/cpp/src/arrow/acero/pivot_longer_node.cc 
b/cpp/src/arrow/acero/pivot_longer_node.cc
index e54f00a20b..ea5ca44baa 100644
--- a/cpp/src/arrow/acero/pivot_longer_node.cc
+++ b/cpp/src/arrow/acero/pivot_longer_node.cc
@@ -135,7 +135,7 @@ class PivotLongerNode : public ExecNode, public TracedNode {
     for (std::size_t i = 0; i < measurement_types.size(); i++) {
       if (!measurement_types[i]) {
         return Status::Invalid(
-            "All row templates had nullopt for the meausrement column at index 
", i, " (",
+            "All row templates had nullopt for the measurement column at index 
", i, " (",
             options.measurement_field_names[i], ")");
       }
       fields.push_back(
diff --git a/cpp/src/arrow/acero/sorted_merge_node.cc 
b/cpp/src/arrow/acero/sorted_merge_node.cc
index f3b934eda1..4d4565a6bb 100644
--- a/cpp/src/arrow/acero/sorted_merge_node.cc
+++ b/cpp/src/arrow/acero/sorted_merge_node.cc
@@ -95,7 +95,7 @@ class BackpressureController : public BackpressureControl {
   std::atomic<int32_t>& backpressure_counter_;
 };
 
-/// InputState correponds to an input. Input record batches are queued up in 
InputState
+/// InputState corresponds to an input. Input record batches are queued up in 
InputState
 /// until processed and turned into output record batches.
 class InputState {
  public:
diff --git a/cpp/src/arrow/acero/swiss_join.cc 
b/cpp/src/arrow/acero/swiss_join.cc
index 3f11b89af3..2f79ed299b 100644
--- a/cpp/src/arrow/acero/swiss_join.cc
+++ b/cpp/src/arrow/acero/swiss_join.cc
@@ -1433,16 +1433,16 @@ void SwissTableForJoinBuild::PrtnMerge(int prtn_id) {
     if (!no_payload_) {
       // Count sort payloads on key id
       //
-      // Start by computing inclusive cummulative sum of counters.
+      // Start by computing inclusive cumulative sum of counters.
       //
       uint32_t sum = 0;
       for (int64_t i = 0; i < num_keys; ++i) {
         sum += counters[i];
         counters[i] = sum;
       }
-      // Now use cummulative sum of counters to obtain the target position in
+      // Now use cumulative sum of counters to obtain the target position in
       // the sorted order for each row. At the end of this process the counters
-      // will contain exclusive cummulative sum (instead of inclusive that is
+      // will contain exclusive cumulative sum (instead of inclusive that is
       // there at the beginning).
       //
       source_payload_ids.resize(prtn_state.key_ids.size());
@@ -1458,7 +1458,7 @@ void SwissTableForJoinBuild::PrtnMerge(int prtn_id) {
       }
     } else {
       // When there is no payload to process, we just need to compute exclusive
-      // cummulative sum of counters and add the base payload id to all of 
them.
+      // cumulative sum of counters and add the base payload id to all of them.
       //
       uint32_t sum = 0;
       for (int64_t i = 0; i < num_keys; ++i) {
diff --git a/cpp/src/arrow/acero/swiss_join_internal.h 
b/cpp/src/arrow/acero/swiss_join_internal.h
index 88b80f06f5..6403b7a655 100644
--- a/cpp/src/arrow/acero/swiss_join_internal.h
+++ b/cpp/src/arrow/acero/swiss_join_internal.h
@@ -156,7 +156,7 @@ class RowArrayMerge {
   // All input sources must be initialized, but they can contain zero rows.
   //
   // Output in vector the first target row id for each source (exclusive
-  // cummulative sum of number of rows in sources). This output is optional,
+  // cumulative sum of number of rows in sources). This output is optional,
   // caller can pass in nullptr to indicate that it is not needed.
   //
   static Status PrepareForMerge(RowArray* target, const 
std::vector<RowArray*>& sources,
@@ -235,7 +235,7 @@ class SwissTableMerge {
   // All input sources must be initialized, but they can be empty.
   //
   // Output in a vector the first target group id for each source (exclusive
-  // cummulative sum of number of groups in sources). This output is optional,
+  // cumulative sum of number of groups in sources). This output is optional,
   // caller can pass in nullptr to indicate that it is not needed.
   //
   static Status PrepareForMerge(SwissTable* target,

Reply via email to