[HUDI-2537] Fix metadata table for flink (#3774)
This commit is contained in:
@@ -41,7 +41,6 @@ import org.apache.hadoop.fs.FileSystem;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.junit.jupiter.api.AfterEach;
|
||||
import org.junit.jupiter.api.BeforeEach;
|
||||
import org.junit.jupiter.api.Disabled;
|
||||
import org.junit.jupiter.api.Test;
|
||||
import org.junit.jupiter.api.io.TempDir;
|
||||
|
||||
@@ -181,7 +180,7 @@ public class TestStreamWriteOperatorCoordinator {
|
||||
assertDoesNotThrow(() -> coordinator.notifyCheckpointComplete(1));
|
||||
}
|
||||
|
||||
@Disabled
|
||||
@Test
|
||||
void testSyncMetadataTable() throws Exception {
|
||||
// reset
|
||||
reset();
|
||||
@@ -193,7 +192,6 @@ public class TestStreamWriteOperatorCoordinator {
|
||||
coordinator = new StreamWriteOperatorCoordinator(conf, context);
|
||||
coordinator.start();
|
||||
coordinator.setExecutor(new MockCoordinatorExecutor(context));
|
||||
coordinator.setMetadataSyncExecutor(new MockCoordinatorExecutor(context));
|
||||
|
||||
final WriteMetadataEvent event0 = WriteMetadataEvent.emptyBootstrap(0);
|
||||
|
||||
@@ -209,7 +207,7 @@ public class TestStreamWriteOperatorCoordinator {
|
||||
assertThat(completedTimeline.lastInstant().get().getTimestamp(), is("0000000000000"));
|
||||
|
||||
// test metadata table compaction
|
||||
// write another 4 commits
|
||||
// write another 3 commits
|
||||
for (int i = 1; i < 4; i++) {
|
||||
instant = mockWriteWithMetadata();
|
||||
metadataTableMetaClient.reloadActiveTimeline();
|
||||
@@ -247,7 +245,13 @@ public class TestStreamWriteOperatorCoordinator {
|
||||
double failureFraction) {
|
||||
final WriteStatus writeStatus = new WriteStatus(trackSuccessRecords, failureFraction);
|
||||
writeStatus.setPartitionPath(partitionPath);
|
||||
writeStatus.setStat(new HoodieWriteStat());
|
||||
|
||||
HoodieWriteStat writeStat = new HoodieWriteStat();
|
||||
writeStat.setPartitionPath(partitionPath);
|
||||
writeStat.setFileId("fileId123");
|
||||
writeStat.setPath("path123");
|
||||
|
||||
writeStatus.setStat(writeStat);
|
||||
|
||||
return WriteMetadataEvent.builder()
|
||||
.taskID(taskId)
|
||||
|
||||
@@ -142,9 +142,6 @@ public class StreamWriteFunctionWrapper<I> {
|
||||
public void openFunction() throws Exception {
|
||||
this.coordinator.start();
|
||||
this.coordinator.setExecutor(new MockCoordinatorExecutor(coordinatorContext));
|
||||
if (conf.getBoolean(FlinkOptions.METADATA_ENABLED)) {
|
||||
this.coordinator.setMetadataSyncExecutor(new MockCoordinatorExecutor(coordinatorContext));
|
||||
}
|
||||
toHoodieFunction = new RowDataToHoodieFunction<>(TestConfigurations.ROW_TYPE, conf);
|
||||
toHoodieFunction.setRuntimeContext(runtimeContext);
|
||||
toHoodieFunction.open(conf);
|
||||
|
||||
@@ -27,7 +27,6 @@ import org.apache.hudi.utils.TestData;
|
||||
import org.apache.flink.configuration.Configuration;
|
||||
import org.apache.hadoop.fs.FileStatus;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.junit.jupiter.api.Disabled;
|
||||
import org.junit.jupiter.api.Test;
|
||||
import org.junit.jupiter.api.io.TempDir;
|
||||
import org.junit.jupiter.params.ParameterizedTest;
|
||||
@@ -90,7 +89,8 @@ public class TestFileIndex {
|
||||
assertTrue(fileStatuses[0].getPath().toString().endsWith(HoodieFileFormat.PARQUET.getFileExtension()));
|
||||
}
|
||||
|
||||
@Disabled
|
||||
@ParameterizedTest
|
||||
@ValueSource(booleans = {true, false})
|
||||
void testFileListingEmptyTable(boolean enableMetadata) {
|
||||
Configuration conf = TestConfigurations.getDefaultConf(tempFile.getAbsolutePath());
|
||||
conf.setBoolean(FlinkOptions.METADATA_ENABLED, enableMetadata);
|
||||
|
||||
Reference in New Issue
Block a user