Skip to content

Commit

Permalink
sort results
Browse files Browse the repository at this point in the history
  • Loading branch information
jihoonson committed Dec 19, 2024
1 parent ed3cdc0 commit 942821f
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
2 changes: 1 addition & 1 deletion integration_tests/src/main/python/orc_write_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -370,7 +370,7 @@ def test_async_writer(spark_tmp_path, hold_gpu):
gen_list = [('_c' + str(i), gen) for i, gen in enumerate(orc_gen)]
assert_gpu_and_cpu_writes_are_equal_collect(
lambda spark, path: gen_df(spark, gen_list, length=num_rows).write.orc(path),
lambda spark, path: spark.read.orc(path),
lambda spark, path: spark.read.orc(path).orderBy([('_c' + str(i)) for i in range(num_cols)]),
data_path,
conf={"spark.rapids.sql.asyncWrite.queryOutput.enabled": "true",
"spark.rapids.sql.batchSizeBytes": 4 * num_cols * 100, # 100 rows per batch
Expand Down
2 changes: 1 addition & 1 deletion integration_tests/src/main/python/parquet_write_test.py
Original file line number Diff line number Diff line change
Expand Up @@ -706,7 +706,7 @@ def test_async_writer(spark_tmp_path, hold_gpu):
gen_list = [('_c' + str(i), gen) for i, gen in enumerate(parquet_gen)]
assert_gpu_and_cpu_writes_are_equal_collect(
lambda spark, path: gen_df(spark, gen_list, length=num_rows).write.parquet(path),
lambda spark, path: spark.read.parquet(path),
lambda spark, path: spark.read.parquet(path).orderBy([('_c' + str(i)) for i in range(num_cols)]),
data_path,
copy_and_update(
writer_confs,
Expand Down

0 comments on commit 942821f

Please sign in to comment.