From 6a5cfb45b9f4784757d2f2d49495d72e23375309 Mon Sep 17 00:00:00 2001 From: Sagar Sumit Date: Sat, 26 Feb 2022 00:21:53 +0530 Subject: [PATCH] [MINOR] Fix table type in input format test (#4912) --- .../hadoop/functional/TestHoodieCombineHiveInputFormat.java | 2 +- .../org/apache/hudi/hadoop/testutils/InputFormatTestUtil.java | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/hudi-hadoop-mr/src/test/java/org/apache/hudi/hadoop/functional/TestHoodieCombineHiveInputFormat.java b/hudi-hadoop-mr/src/test/java/org/apache/hudi/hadoop/functional/TestHoodieCombineHiveInputFormat.java index 32d7f1e3b..ec6ea0a8b 100644 --- a/hudi-hadoop-mr/src/test/java/org/apache/hudi/hadoop/functional/TestHoodieCombineHiveInputFormat.java +++ b/hudi-hadoop-mr/src/test/java/org/apache/hudi/hadoop/functional/TestHoodieCombineHiveInputFormat.java @@ -104,7 +104,7 @@ public class TestHoodieCombineHiveInputFormat extends HoodieCommonTestHarness { final int numRecords = 1000; // Create 3 partitions, each partition holds one parquet file and 1000 records List partitionDirs = InputFormatTestUtil - .prepareMultiPartitionedParquetTable(tempDir, schema, 3, numRecords, commitTime); + .prepareMultiPartitionedParquetTable(tempDir, schema, 3, numRecords, commitTime, HoodieTableType.MERGE_ON_READ); InputFormatTestUtil.commit(tempDir, commitTime); TableDesc tblDesc = Utilities.defaultTd; diff --git a/hudi-hadoop-mr/src/test/java/org/apache/hudi/hadoop/testutils/InputFormatTestUtil.java b/hudi-hadoop-mr/src/test/java/org/apache/hudi/hadoop/testutils/InputFormatTestUtil.java index 352ed0d77..836ad5712 100644 --- a/hudi-hadoop-mr/src/test/java/org/apache/hudi/hadoop/testutils/InputFormatTestUtil.java +++ b/hudi-hadoop-mr/src/test/java/org/apache/hudi/hadoop/testutils/InputFormatTestUtil.java @@ -246,9 +246,9 @@ public class InputFormatTestUtil { } public static List prepareMultiPartitionedParquetTable(java.nio.file.Path basePath, Schema schema, - int numberPartitions, int numberOfRecordsPerPartition, String commitNumber) throws IOException { + int numberPartitions, int numberOfRecordsPerPartition, String commitNumber, HoodieTableType tableType) throws IOException { List result = new ArrayList<>(); - HoodieTestUtils.init(HoodieTestUtils.getDefaultHadoopConf(), basePath.toString()); + HoodieTestUtils.init(HoodieTestUtils.getDefaultHadoopConf(), basePath.toString(), tableType, HoodieFileFormat.PARQUET); for (int i = 0; i < numberPartitions; i++) { java.nio.file.Path partitionPath = basePath.resolve(Paths.get(2016 + i + "", "05", "01")); setupPartition(basePath, partitionPath);