sanjibansg commented on code in PR #12755:
URL: https://github.com/apache/arrow/pull/12755#discussion_r851495892
##########
cpp/src/arrow/compute/exec/expression_benchmark.cc:
##########
@@ -69,6 +70,26 @@ static void SimplifyFilterWithGuarantee(benchmark::State&
state, Expression filt
}
}
+static void ExecuteScalarExpressionOverhead(benchmark::State& state,
Expression expr) {
+ const auto rows_per_batch = static_cast<int32_t>(state.range(0));
+ const auto num_batches = 10000000 / rows_per_batch;
Review Comment:
I basically wanted to benchmark the execution with different batch sizes but
with the same total data amount. So, if we want to benchmark on batches with
10 rows each, then we should have 100000 batches, thus in total we will have
10000000 data points. Keeping the total data fix, the benchmarking tries to see
how the execution gets affected when we execute the same amount of data but
with batches of different sizes.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]