[HUDI-3290] Different file formats for the partition metadata file. (#5179)
* [HUDI-3290] Different file formats for the partition metadata file. Partition metadata files are stored in each partition to help identify the base path of a table. These files are saved in the properties file format. Some query engines do not work when non Parquet/ORC files are found in a partition. Added a new table config 'hoodie.partition.metafile.use.data.format' which when enabled (default false for backward compatibility) ensures that partition metafiles will be saved in the same format as the base files of a dataset. For new datasets, the config can be set via hudi-cli. Deltastreamer has a new parameter --partition-metafile-use-data-format which will create a table with this setting. * Code review comments - Adding a new command to migrate from text to base file formats for meta file. - Reimplementing readFromFS() to first read the text format, then base format - Avoid extra exists() checks in readFromFS() - Added unit tests, enabled parquet format across hoodie-hadoop-mr - Code cleanup, restructuring, naming consistency. * Wiring in all the other Spark code paths to respect this config - Turned on parquet meta format for COW data source tests - Removed the deltastreamer command line to keep it shorter * populate HoodiePartitionMetadata#format after readFromFS() Co-authored-by: Vinoth Chandar <vinoth@apache.org> Co-authored-by: Raymond Xu <2701446+xushiyan@users.noreply.github.com>
This commit is contained in:
@@ -34,6 +34,7 @@ import org.apache.hudi.common.table.log.block.HoodieLogBlock;
|
||||
import org.apache.hudi.common.table.log.block.HoodieParquetDataBlock;
|
||||
import org.apache.hudi.common.testutils.HoodieTestUtils;
|
||||
import org.apache.hudi.common.testutils.SchemaTestUtil;
|
||||
import org.apache.hudi.common.util.Option;
|
||||
import org.apache.hudi.hadoop.utils.HoodieHiveUtils;
|
||||
|
||||
import org.apache.avro.Schema;
|
||||
@@ -185,7 +186,7 @@ public class InputFormatTestUtil {
|
||||
public static void setupSnapshotScanMode(JobConf jobConf) {
|
||||
setupSnapshotScanMode(jobConf, false);
|
||||
}
|
||||
|
||||
|
||||
private static void setupSnapshotScanMode(JobConf jobConf, boolean includePending) {
|
||||
setUpScanMode(jobConf);
|
||||
String includePendingCommitsName =
|
||||
@@ -467,8 +468,8 @@ public class InputFormatTestUtil {
|
||||
new LocalFileSystem(lfs),
|
||||
"0",
|
||||
new Path(basePath.toAbsolutePath().toString()),
|
||||
new Path(partitionPath.toAbsolutePath().toString())
|
||||
);
|
||||
new Path(partitionPath.toAbsolutePath().toString()),
|
||||
Option.of(HoodieFileFormat.PARQUET));
|
||||
|
||||
partitionMetadata.trySave((int) (Math.random() * 1000));
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user