1
0

[HUDI-1775] Add option for compaction parallelism (#2785)

This commit is contained in:
Danny Chan
2021-04-09 13:46:19 +08:00
committed by GitHub
parent 08e82c469c
commit 6786581c48
6 changed files with 58 additions and 11 deletions

View File

@@ -449,6 +449,10 @@ public class TestWriteCopyOnWrite {
@Test
public void testIndexStateBootstrap() throws Exception {
// reset the config option
conf.setBoolean(FlinkOptions.INDEX_BOOTSTRAP_ENABLED, true);
funcWrapper = new StreamWriteFunctionWrapper<>(tempFile.getAbsolutePath(), conf);
// open the function and ingest data
funcWrapper.openFunction();
for (RowData rowData : TestData.DATA_SET_INSERT) {

View File

@@ -169,6 +169,7 @@ public class HoodieDataSourceITCase extends AbstractTestBase {
options.put(FlinkOptions.TABLE_TYPE.key(), FlinkOptions.TABLE_TYPE_MERGE_ON_READ);
options.put(FlinkOptions.QUERY_TYPE.key(), FlinkOptions.QUERY_TYPE_READ_OPTIMIZED);
options.put(FlinkOptions.COMPACTION_DELTA_COMMITS.key(), "1");
options.put(FlinkOptions.COMPACTION_TASKS.key(), "1");
String hoodieTableDDL = TestConfigurations.getCreateHoodieTableDDL("t1", options);
streamTableEnv.executeSql(hoodieTableDDL);
String insertInto = "insert into t1 select * from source";
@@ -180,7 +181,7 @@ public class HoodieDataSourceITCase extends AbstractTestBase {
}
@Test
void testStreamWriteWithCleaning() throws InterruptedException {
void testStreamWriteWithCleaning() {
// create filesystem table named source
// the source generates 4 commits but the cleaning task