[HUDI-4290] Fix fetchLatestBaseFiles to filter replaced filegroups (#5941)
* [HUDI-4290] Fix fetchLatestBaseFiles to filter replaced filegroups * Separate out incremental sync fsview test with clustering
This commit is contained in:
@@ -18,8 +18,6 @@
|
|||||||
|
|
||||||
package org.apache.hudi.common.table.view;
|
package org.apache.hudi.common.table.view;
|
||||||
|
|
||||||
import org.apache.hadoop.fs.FileStatus;
|
|
||||||
import org.apache.hadoop.fs.Path;
|
|
||||||
import org.apache.hudi.common.bootstrap.index.BootstrapIndex;
|
import org.apache.hudi.common.bootstrap.index.BootstrapIndex;
|
||||||
import org.apache.hudi.common.fs.FSUtils;
|
import org.apache.hudi.common.fs.FSUtils;
|
||||||
import org.apache.hudi.common.model.BootstrapBaseFileMapping;
|
import org.apache.hudi.common.model.BootstrapBaseFileMapping;
|
||||||
@@ -41,6 +39,9 @@ import org.apache.hudi.common.util.Option;
|
|||||||
import org.apache.hudi.common.util.ValidationUtils;
|
import org.apache.hudi.common.util.ValidationUtils;
|
||||||
import org.apache.hudi.common.util.collection.Pair;
|
import org.apache.hudi.common.util.collection.Pair;
|
||||||
import org.apache.hudi.exception.HoodieIOException;
|
import org.apache.hudi.exception.HoodieIOException;
|
||||||
|
|
||||||
|
import org.apache.hadoop.fs.FileStatus;
|
||||||
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.log4j.LogManager;
|
import org.apache.log4j.LogManager;
|
||||||
import org.apache.log4j.Logger;
|
import org.apache.log4j.Logger;
|
||||||
|
|
||||||
@@ -973,6 +974,7 @@ public abstract class AbstractTableFileSystemView implements SyncableFileSystemV
|
|||||||
*/
|
*/
|
||||||
public Stream<HoodieBaseFile> fetchLatestBaseFiles(final String partitionPath) {
|
public Stream<HoodieBaseFile> fetchLatestBaseFiles(final String partitionPath) {
|
||||||
return fetchAllStoredFileGroups(partitionPath)
|
return fetchAllStoredFileGroups(partitionPath)
|
||||||
|
.filter(fg -> !isFileGroupReplaced(fg))
|
||||||
.map(fg -> Pair.of(fg.getFileGroupId(), getLatestBaseFile(fg)))
|
.map(fg -> Pair.of(fg.getFileGroupId(), getLatestBaseFile(fg)))
|
||||||
.filter(p -> p.getValue().isPresent())
|
.filter(p -> p.getValue().isPresent())
|
||||||
.map(p -> addBootstrapBaseFileIfPresent(p.getKey(), p.getValue().get()));
|
.map(p -> addBootstrapBaseFileIfPresent(p.getKey(), p.getValue().get()));
|
||||||
|
|||||||
@@ -0,0 +1,181 @@
|
|||||||
|
/*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing,
|
||||||
|
* software distributed under the License is distributed on an
|
||||||
|
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||||
|
* KIND, either express or implied. See the License for the
|
||||||
|
* specific language governing permissions and limitations
|
||||||
|
* under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hudi.common.table.view;
|
||||||
|
|
||||||
|
import org.apache.hudi.common.fs.FSUtils;
|
||||||
|
import org.apache.hudi.common.model.HoodieCommitMetadata;
|
||||||
|
import org.apache.hudi.common.model.HoodieFileGroup;
|
||||||
|
import org.apache.hudi.common.model.WriteOperationType;
|
||||||
|
import org.apache.hudi.common.table.timeline.HoodieActiveTimeline;
|
||||||
|
import org.apache.hudi.common.table.timeline.HoodieInstant;
|
||||||
|
import org.apache.hudi.common.table.timeline.HoodieTimeline;
|
||||||
|
import org.apache.hudi.common.testutils.HoodieCommonTestHarness;
|
||||||
|
import org.apache.hudi.common.testutils.HoodieTestUtils;
|
||||||
|
import org.apache.hudi.common.util.CommitUtils;
|
||||||
|
import org.apache.hudi.common.util.Option;
|
||||||
|
|
||||||
|
import org.junit.jupiter.api.AfterEach;
|
||||||
|
import org.junit.jupiter.api.BeforeEach;
|
||||||
|
import org.junit.jupiter.api.Test;
|
||||||
|
|
||||||
|
import java.io.File;
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.nio.charset.StandardCharsets;
|
||||||
|
import java.util.ArrayList;
|
||||||
|
import java.util.Collections;
|
||||||
|
import java.util.HashMap;
|
||||||
|
import java.util.List;
|
||||||
|
import java.util.Map;
|
||||||
|
import java.util.Set;
|
||||||
|
import java.util.UUID;
|
||||||
|
import java.util.stream.Collectors;
|
||||||
|
import java.util.stream.Stream;
|
||||||
|
|
||||||
|
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||||
|
import static org.junit.jupiter.api.Assertions.assertFalse;
|
||||||
|
|
||||||
|
public class TestHoodieTableFSViewWithClustering extends HoodieCommonTestHarness {
|
||||||
|
|
||||||
|
private static final String TEST_WRITE_TOKEN = "1-0-1";
|
||||||
|
private static final String BOOTSTRAP_SOURCE_PATH = "/usr/warehouse/hive/data/tables/src1/";
|
||||||
|
|
||||||
|
private IncrementalTimelineSyncFileSystemView fsView;
|
||||||
|
private TableFileSystemView.BaseFileOnlyView roView;
|
||||||
|
|
||||||
|
@BeforeEach
|
||||||
|
public void setup() throws IOException {
|
||||||
|
metaClient = HoodieTestUtils.init(tempDir.toAbsolutePath().toString(), getTableType(), BOOTSTRAP_SOURCE_PATH, false);
|
||||||
|
basePath = metaClient.getBasePathV2().toString();
|
||||||
|
refreshFsView();
|
||||||
|
}
|
||||||
|
|
||||||
|
protected void refreshFsView() throws IOException {
|
||||||
|
super.refreshFsView();
|
||||||
|
closeFsView();
|
||||||
|
fsView = (IncrementalTimelineSyncFileSystemView) getFileSystemView(metaClient.getActiveTimeline().filterCompletedAndCompactionInstants());
|
||||||
|
roView = fsView;
|
||||||
|
}
|
||||||
|
|
||||||
|
private void closeFsView() {
|
||||||
|
if (null != fsView) {
|
||||||
|
fsView.close();
|
||||||
|
fsView = null;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@AfterEach
|
||||||
|
public void close() {
|
||||||
|
closeFsView();
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testReplaceFileIdIsExcludedInView() throws IOException {
|
||||||
|
String partitionPath1 = "2020/06/27";
|
||||||
|
String partitionPath2 = "2020/07/14";
|
||||||
|
new File(basePath + "/" + partitionPath1).mkdirs();
|
||||||
|
new File(basePath + "/" + partitionPath2).mkdirs();
|
||||||
|
|
||||||
|
// create 2 fileId in partition1 - fileId1 is replaced later on.
|
||||||
|
String fileId1 = UUID.randomUUID().toString();
|
||||||
|
String fileId2 = UUID.randomUUID().toString();
|
||||||
|
|
||||||
|
// create 2 fileId in partition2 - fileId3, fileId4 is replaced later on.
|
||||||
|
String fileId3 = UUID.randomUUID().toString();
|
||||||
|
String fileId4 = UUID.randomUUID().toString();
|
||||||
|
|
||||||
|
assertFalse(roView.getLatestBaseFiles(partitionPath1)
|
||||||
|
.anyMatch(dfile -> dfile.getFileId().equals(fileId1) || dfile.getFileId().equals(fileId2)),
|
||||||
|
"No commit, should not find any data file");
|
||||||
|
assertFalse(roView.getLatestBaseFiles(partitionPath2)
|
||||||
|
.anyMatch(dfile -> dfile.getFileId().equals(fileId3) || dfile.getFileId().equals(fileId4)),
|
||||||
|
"No commit, should not find any data file");
|
||||||
|
assertFalse(fsView.fetchLatestBaseFiles(partitionPath1)
|
||||||
|
.anyMatch(dfile -> dfile.getFileId().equals(fileId1) || dfile.getFileId().equals(fileId2)),
|
||||||
|
"No commit, should not find any data file");
|
||||||
|
assertFalse(fsView.fetchLatestBaseFiles(partitionPath2)
|
||||||
|
.anyMatch(dfile -> dfile.getFileId().equals(fileId3) || dfile.getFileId().equals(fileId4)),
|
||||||
|
"No commit, should not find any data file");
|
||||||
|
|
||||||
|
// Only one commit
|
||||||
|
String commitTime1 = "1";
|
||||||
|
String fileName1 = FSUtils.makeBaseFileName(commitTime1, TEST_WRITE_TOKEN, fileId1);
|
||||||
|
String fileName2 = FSUtils.makeBaseFileName(commitTime1, TEST_WRITE_TOKEN, fileId2);
|
||||||
|
String fileName3 = FSUtils.makeBaseFileName(commitTime1, TEST_WRITE_TOKEN, fileId3);
|
||||||
|
String fileName4 = FSUtils.makeBaseFileName(commitTime1, TEST_WRITE_TOKEN, fileId4);
|
||||||
|
new File(basePath + "/" + partitionPath1 + "/" + fileName1).createNewFile();
|
||||||
|
new File(basePath + "/" + partitionPath1 + "/" + fileName2).createNewFile();
|
||||||
|
new File(basePath + "/" + partitionPath2 + "/" + fileName3).createNewFile();
|
||||||
|
new File(basePath + "/" + partitionPath2 + "/" + fileName4).createNewFile();
|
||||||
|
|
||||||
|
Map<String, List<String>> partitionToReplaceFileIds = new HashMap<>();
|
||||||
|
List<String> replacedFileIdsP1 = new ArrayList<>();
|
||||||
|
replacedFileIdsP1.add(fileId1);
|
||||||
|
partitionToReplaceFileIds.put(partitionPath1, replacedFileIdsP1);
|
||||||
|
List<String> replacedFileIdsP2 = new ArrayList<>();
|
||||||
|
replacedFileIdsP2.add(fileId3);
|
||||||
|
replacedFileIdsP2.add(fileId4);
|
||||||
|
partitionToReplaceFileIds.put(partitionPath2, replacedFileIdsP2);
|
||||||
|
HoodieCommitMetadata commitMetadata =
|
||||||
|
CommitUtils.buildMetadata(Collections.emptyList(), partitionToReplaceFileIds, Option.empty(), WriteOperationType.INSERT_OVERWRITE, "", HoodieTimeline.REPLACE_COMMIT_ACTION);
|
||||||
|
|
||||||
|
HoodieActiveTimeline commitTimeline = metaClient.getActiveTimeline();
|
||||||
|
HoodieInstant instant1 = new HoodieInstant(true, HoodieTimeline.REPLACE_COMMIT_ACTION, commitTime1);
|
||||||
|
saveAsComplete(commitTimeline, instant1, Option.of(commitMetadata.toJsonString().getBytes(StandardCharsets.UTF_8)));
|
||||||
|
refreshFsView();
|
||||||
|
assertEquals(0, roView.getLatestBaseFiles(partitionPath1)
|
||||||
|
.filter(dfile -> dfile.getFileId().equals(fileId1)).count());
|
||||||
|
assertEquals(fileName2, roView.getLatestBaseFiles(partitionPath1)
|
||||||
|
.filter(dfile -> dfile.getFileId().equals(fileId2)).findFirst().get().getFileName());
|
||||||
|
assertEquals(0, roView.getLatestBaseFiles(partitionPath2)
|
||||||
|
.filter(dfile -> dfile.getFileId().equals(fileId3)).count());
|
||||||
|
assertEquals(0, roView.getLatestBaseFiles(partitionPath2)
|
||||||
|
.filter(dfile -> dfile.getFileId().equals(fileId4)).count());
|
||||||
|
assertEquals(0, fsView.fetchLatestBaseFiles(partitionPath1)
|
||||||
|
.filter(dfile -> dfile.getFileId().equals(fileId1)).count());
|
||||||
|
assertEquals(fileName2, fsView.fetchLatestBaseFiles(partitionPath1)
|
||||||
|
.filter(dfile -> dfile.getFileId().equals(fileId2)).findFirst().get().getFileName());
|
||||||
|
assertEquals(0, fsView.fetchLatestBaseFiles(partitionPath2)
|
||||||
|
.filter(dfile -> dfile.getFileId().equals(fileId3)).count());
|
||||||
|
assertEquals(0, fsView.fetchLatestBaseFiles(partitionPath2)
|
||||||
|
.filter(dfile -> dfile.getFileId().equals(fileId4)).count());
|
||||||
|
|
||||||
|
// ensure replacedFileGroupsBefore works with all instants
|
||||||
|
List<HoodieFileGroup> replacedOnInstant1 = fsView.getReplacedFileGroupsBeforeOrOn("0", partitionPath1).collect(Collectors.toList());
|
||||||
|
assertEquals(0, replacedOnInstant1.size());
|
||||||
|
|
||||||
|
List<HoodieFileGroup> allReplaced = fsView.getReplacedFileGroupsBeforeOrOn("2", partitionPath1).collect(Collectors.toList());
|
||||||
|
allReplaced.addAll(fsView.getReplacedFileGroupsBeforeOrOn("2", partitionPath2).collect(Collectors.toList()));
|
||||||
|
assertEquals(3, allReplaced.size());
|
||||||
|
Set<String> allReplacedFileIds = allReplaced.stream().map(fg -> fg.getFileGroupId().getFileId()).collect(Collectors.toSet());
|
||||||
|
Set<String> actualReplacedFileIds = Stream.of(fileId1, fileId3, fileId4).collect(Collectors.toSet());
|
||||||
|
assertEquals(actualReplacedFileIds, allReplacedFileIds);
|
||||||
|
}
|
||||||
|
|
||||||
|
private static void saveAsComplete(HoodieActiveTimeline timeline, HoodieInstant inflight, Option<byte[]> data) {
|
||||||
|
if (inflight.getAction().equals(HoodieTimeline.COMPACTION_ACTION)) {
|
||||||
|
timeline.transitionCompactionInflightToComplete(inflight, data);
|
||||||
|
} else {
|
||||||
|
HoodieInstant requested = new HoodieInstant(HoodieInstant.State.REQUESTED, inflight.getAction(), inflight.getTimestamp());
|
||||||
|
timeline.createNewInstant(requested);
|
||||||
|
timeline.transitionRequestedToInflight(requested, Option.empty());
|
||||||
|
timeline.saveAsComplete(inflight, data);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -18,9 +18,6 @@
|
|||||||
|
|
||||||
package org.apache.hudi.common.table.view;
|
package org.apache.hudi.common.table.view;
|
||||||
|
|
||||||
import org.apache.hadoop.fs.FileStatus;
|
|
||||||
import org.apache.hadoop.fs.Path;
|
|
||||||
import org.apache.hadoop.fs.permission.FsAction;
|
|
||||||
import org.apache.hudi.avro.model.HoodieClusteringPlan;
|
import org.apache.hudi.avro.model.HoodieClusteringPlan;
|
||||||
import org.apache.hudi.avro.model.HoodieCompactionPlan;
|
import org.apache.hudi.avro.model.HoodieCompactionPlan;
|
||||||
import org.apache.hudi.avro.model.HoodieFSPermission;
|
import org.apache.hudi.avro.model.HoodieFSPermission;
|
||||||
@@ -61,6 +58,9 @@ import org.apache.hudi.common.util.Option;
|
|||||||
import org.apache.hudi.common.util.collection.ImmutablePair;
|
import org.apache.hudi.common.util.collection.ImmutablePair;
|
||||||
import org.apache.hudi.common.util.collection.Pair;
|
import org.apache.hudi.common.util.collection.Pair;
|
||||||
|
|
||||||
|
import org.apache.hadoop.fs.FileStatus;
|
||||||
|
import org.apache.hadoop.fs.Path;
|
||||||
|
import org.apache.hadoop.fs.permission.FsAction;
|
||||||
import org.apache.log4j.LogManager;
|
import org.apache.log4j.LogManager;
|
||||||
import org.apache.log4j.Logger;
|
import org.apache.log4j.Logger;
|
||||||
import org.junit.jupiter.api.BeforeEach;
|
import org.junit.jupiter.api.BeforeEach;
|
||||||
|
|||||||
Reference in New Issue
Block a user