zanmato1984 commented on code in PR #48252:
URL: https://github.com/apache/arrow/pull/48252#discussion_r2576067421
##########
cpp/src/arrow/csv/fuzz.cc:
##########
@@ -42,10 +42,11 @@ Status FuzzCsvReader(const uint8_t* data, int64_t size) {
auto read_options = ReadOptions::Defaults();
// Make chunking more likely
- read_options.block_size = 4096;
+ read_options.block_size = 1000;
auto parse_options = ParseOptions::Defaults();
auto convert_options = ConvertOptions::Defaults();
convert_options.auto_dict_encode = true;
+ convert_options.auto_dict_max_cardinality = 50;
Review Comment:
Why do we need these changes?
##########
cpp/src/arrow/csv/generate_fuzz_corpus.cc:
##########
@@ -0,0 +1,204 @@
+// Licensed to the Apache Software Foundation (ASF) under one
+// or more contributor license agreements. See the NOTICE file
+// distributed with this work for additional information
+// regarding copyright ownership. The ASF licenses this file
+// to you under the Apache License, Version 2.0 (the
+// "License"); you may not use this file except in compliance
+// with the License. You may obtain a copy of the License at
+//
+// http://www.apache.org/licenses/LICENSE-2.0
+//
+// Unless required by applicable law or agreed to in writing,
+// software distributed under the License is distributed on an
+// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+// KIND, either express or implied. See the License for the
+// specific language governing permissions and limitations
+// under the License.
+
+// A command line executable that generates a bunch of valid IPC files
+// containing example record batches. Those are used as fuzzing seeds
+// to make fuzzing more efficient.
+
+#include <cstdlib>
+#include <functional>
+#include <iostream>
+#include <memory>
+#include <string>
+#include <vector>
+
+#include "arrow/array.h"
+#include "arrow/array/util.h"
+#include "arrow/compute/cast.h"
+#include "arrow/csv/options.h"
+#include "arrow/csv/writer.h"
+#include "arrow/io/file.h"
+#include "arrow/io/memory.h"
+#include "arrow/ipc/writer.h"
+#include "arrow/json/from_string.h"
+#include "arrow/record_batch.h"
+#include "arrow/result.h"
+#include "arrow/testing/random.h"
+#include "arrow/util/io_util.h"
+
+namespace arrow::csv {
+
+using ::arrow::internal::CreateDir;
+using ::arrow::internal::PlatformFilename;
+using ::arrow::json::ArrayFromJSONString;
+
+Result<std::shared_ptr<Buffer>> WriteRecordBatch(
+ const std::shared_ptr<RecordBatch>& batch, const WriteOptions& options) {
+ ARROW_ASSIGN_OR_RAISE(auto sink, io::BufferOutputStream::Create(1024));
+ ARROW_ASSIGN_OR_RAISE(auto writer, MakeCSVWriter(sink.get(),
batch->schema(), options));
+ RETURN_NOT_OK(writer->WriteRecordBatch(*batch));
+ RETURN_NOT_OK(writer->Close());
+ return sink->Finish();
+}
+
+Result<std::shared_ptr<RecordBatch>> MakeBatch(
+ std::function<Result<std::shared_ptr<Array>>(int64_t length, double
null_probability)>
+ array_factory,
+ int64_t length) {
+ ArrayVector columns;
+ FieldVector fields;
+
+ struct ColumnSpec {
+ std::string name;
+ double null_probability;
+ };
+ for (auto spec : {ColumnSpec{"with_nulls", 0.2}, ColumnSpec{"without_nulls",
0.0}}) {
+ ARROW_ASSIGN_OR_RAISE(std::shared_ptr<Array> column,
+ array_factory(length, spec.null_probability));
+ columns.push_back(column);
+ fields.push_back(field(spec.name, column->type()));
+ }
+ return RecordBatch::Make(schema(std::move(fields)), length,
std::move(columns));
+}
+
+Result<RecordBatchVector> Batches() {
+ ::arrow::random::RandomArrayGenerator gen(/*seed=*/42);
+ RecordBatchVector batches;
+
+ auto append_batch = [&](auto array_factory, int64_t length) -> Status {
+ ARROW_ASSIGN_OR_RAISE(auto batch, MakeBatch(array_factory, length));
+ batches.push_back(batch);
+ return Status::OK();
+ };
+
+ // Ideally, we should exercise all possible inference kinds (see
inference_internal.h)
+ auto make_nulls = [&](int64_t length, double null_probability) {
+ return MakeArrayOfNull(null(), length);
+ };
+ auto make_ints = [&](int64_t length, double null_probability) {
+ return gen.Int64(length, /*min=*/-1'000'000, /*max=*/1'000'000,
null_probability);
+ };
+ auto make_floats = [&](int64_t length, double null_probability) {
+ return gen.Float64(length, /*min=*/-100.0, /*max=*/100.0,
null_probability);
+ };
+ auto make_booleans = [&](int64_t length, double null_probability) {
+ return gen.Boolean(length, /*true_probability=*/0.8, null_probability);
+ };
+ auto make_dates = [&](int64_t length, double null_probability) {
+ return gen.Date64(length, /*min=*/1, /*max=*/365 * 60, null_probability);
+ };
+ auto make_times = [&](int64_t length, double null_probability) {
+ return gen.Int32(length, /*min=*/0, /*max=*/86399, null_probability)
+ ->View(time32(TimeUnit::SECOND));
+ };
+
+ std::string timezone;
+ auto make_timestamps = [&](int64_t length, double null_probability) {
+ return gen.Int64(length, /*min=*/1, /*max=*/1764079190, null_probability)
+ ->View(timestamp(TimeUnit::SECOND, timezone));
+ };
+ auto make_timestamps_ns = [&](int64_t length, double null_probability) {
+ return gen
+ .Int64(length, /*min=*/1, /*max=*/1764079190LL * 1'000'000'000,
null_probability)
+ ->View(timestamp(TimeUnit::NANO, timezone));
+ };
+
+ auto make_strings = [&](int64_t length, double null_probability) {
+ return gen.String(length, /*min_length=*/3, /*max_length=*/15,
null_probability);
+ };
+ auto make_string_with_repeats = [&](int64_t length, double null_probability)
{
+ // `unique` should be less than `auto_dict_max_cardinality` in fuzz target
+ return gen.StringWithRepeats(length, /*unique=*/10, /*min_length=*/3,
+ /*max_length=*/15, null_probability);
+ };
+
+ RETURN_NOT_OK(append_batch(make_nulls, /*length=*/2000));
+ RETURN_NOT_OK(append_batch(make_ints, /*length=*/500));
+ RETURN_NOT_OK(append_batch(make_floats, /*length=*/150));
+ RETURN_NOT_OK(append_batch(make_booleans, /*length=*/500));
+
+ RETURN_NOT_OK(append_batch(make_dates, /*length=*/200));
+ RETURN_NOT_OK(append_batch(make_times, /*length=*/400));
+ timezone = "";
+ RETURN_NOT_OK(append_batch(make_timestamps, /*length=*/200));
+ RETURN_NOT_OK(append_batch(make_timestamps_ns, /*length=*/100));
+ // Will generate timestamps with a "Z" suffix
+ timezone = "UTC";
+ RETURN_NOT_OK(append_batch(make_timestamps, /*length=*/200));
+ RETURN_NOT_OK(append_batch(make_timestamps_ns, /*length=*/100));
+ // Will generate timestamps with a "+0100" or "+0200" suffix
+ timezone = "Europe/Paris";
+ RETURN_NOT_OK(append_batch(make_timestamps, /*length=*/200));
+ RETURN_NOT_OK(append_batch(make_timestamps_ns, /*length=*/100));
+
+ RETURN_NOT_OK(append_batch(make_strings, /*length=*/300));
+ RETURN_NOT_OK(append_batch(make_string_with_repeats, /*length=*/300));
+ // XXX Cannot add non-UTF8 binary as the CSV writer doesn't support writing
it
+
+ return batches;
+}
+
+Status DoMain(const std::string& out_dir) {
+ ARROW_ASSIGN_OR_RAISE(auto dir_fn, PlatformFilename::FromString(out_dir));
+ RETURN_NOT_OK(CreateDir(dir_fn));
+
+ int sample_num = 1;
+ auto sample_name = [&]() -> std::string {
+ return "csv-file-" + std::to_string(sample_num++);
+ };
+
+ ARROW_ASSIGN_OR_RAISE(auto batches, Batches());
+
+ auto options = WriteOptions::Defaults();
+ RETURN_NOT_OK(options.Validate());
+
+ for (const auto& batch : batches) {
+ RETURN_NOT_OK(batch->ValidateFull());
+ ARROW_ASSIGN_OR_RAISE(auto buffer, WriteRecordBatch(batch, options));
+
+ ARROW_ASSIGN_OR_RAISE(auto sample_fn, dir_fn.Join(sample_name()));
+ std::cerr << sample_fn.ToString() << std::endl;
Review Comment:
Why use standard error rater than standard out?
##########
cpp/build-support/fuzzing/generate_corpuses.sh:
##########
@@ -71,7 +71,7 @@ rm -rf ${PANDAS_DIR}
git clone --depth=1 https://github.com/pandas-dev/pandas ${PANDAS_DIR}
rm -rf ${CORPUS_DIR}
-mkdir -p ${CORPUS_DIR}
Review Comment:
So before this change, we merely create an empty directory?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]