Skip to content

Commit

Permalink
Trigger failures by removing the configs
Browse files Browse the repository at this point in the history
  • Loading branch information
linliu-code committed Dec 17, 2024
1 parent a776c57 commit f34f380
Showing 1 changed file with 0 additions and 8 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,6 @@
import org.apache.hudi.common.engine.TaskContextSupplier;
import org.apache.hudi.common.table.HoodieTableConfig;
import org.apache.hudi.common.util.Option;
import org.apache.hudi.common.util.StringUtils;
import org.apache.hudi.exception.HoodieIOException;
import org.apache.hudi.io.storage.row.HoodieRowParquetConfig;
import org.apache.hudi.io.storage.row.HoodieRowParquetWriteSupport;
Expand All @@ -36,7 +35,6 @@
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FSDataOutputStream;
import org.apache.parquet.hadoop.metadata.CompressionCodecName;
import org.apache.spark.sql.HoodieDataTypeUtils;
import org.apache.spark.sql.HoodieInternalRowUtils;
import org.apache.spark.sql.types.StructType;

Expand All @@ -58,12 +56,6 @@ protected HoodieFileWriter newParquetFileWriter(
TaskContextSupplier taskContextSupplier) throws IOException {
String writeSchema = config.getStringOrDefault(
WRITE_SCHEMA_OVERRIDE, config.getString(AVRO_SCHEMA_STRING));
if (!StringUtils.isNullOrEmpty(writeSchema)) {
// The parquet write legacy format property needs to be overridden
// if there is a decimal field of small precision, to maintain the compatibility.
HoodieDataTypeUtils.tryOverrideParquetWriteLegacyFormatProperty(
config.getProps(), new Schema.Parser().parse(writeSchema));
}
boolean populateMetaFields = config.getBooleanOrDefault(HoodieTableConfig.POPULATE_META_FIELDS);
String compressionCodecName = config.getStringOrDefault(HoodieStorageConfig.PARQUET_COMPRESSION_CODEC_NAME);
// Support PARQUET_COMPRESSION_CODEC_NAME is ""
Expand Down

0 comments on commit f34f380

Please sign in to comment.