[HUDI-1775] Add option for compaction parallelism (#2785)
This commit is contained in:
@@ -449,6 +449,10 @@ public class TestWriteCopyOnWrite {
|
||||
|
||||
@Test
|
||||
public void testIndexStateBootstrap() throws Exception {
|
||||
// reset the config option
|
||||
conf.setBoolean(FlinkOptions.INDEX_BOOTSTRAP_ENABLED, true);
|
||||
funcWrapper = new StreamWriteFunctionWrapper<>(tempFile.getAbsolutePath(), conf);
|
||||
|
||||
// open the function and ingest data
|
||||
funcWrapper.openFunction();
|
||||
for (RowData rowData : TestData.DATA_SET_INSERT) {
|
||||
|
||||
@@ -169,6 +169,7 @@ public class HoodieDataSourceITCase extends AbstractTestBase {
|
||||
options.put(FlinkOptions.TABLE_TYPE.key(), FlinkOptions.TABLE_TYPE_MERGE_ON_READ);
|
||||
options.put(FlinkOptions.QUERY_TYPE.key(), FlinkOptions.QUERY_TYPE_READ_OPTIMIZED);
|
||||
options.put(FlinkOptions.COMPACTION_DELTA_COMMITS.key(), "1");
|
||||
options.put(FlinkOptions.COMPACTION_TASKS.key(), "1");
|
||||
String hoodieTableDDL = TestConfigurations.getCreateHoodieTableDDL("t1", options);
|
||||
streamTableEnv.executeSql(hoodieTableDDL);
|
||||
String insertInto = "insert into t1 select * from source";
|
||||
@@ -180,7 +181,7 @@ public class HoodieDataSourceITCase extends AbstractTestBase {
|
||||
}
|
||||
|
||||
@Test
|
||||
void testStreamWriteWithCleaning() throws InterruptedException {
|
||||
void testStreamWriteWithCleaning() {
|
||||
// create filesystem table named source
|
||||
|
||||
// the source generates 4 commits but the cleaning task
|
||||
|
||||
Reference in New Issue
Block a user