1
0

[HUDI-1723] Fix path selector listing files with the same mod date (#2845)

This commit is contained in:
Raymond Xu
2021-05-25 07:19:10 -07:00
committed by GitHub
parent e7020748b5
commit afa6bc0b10
4 changed files with 184 additions and 10 deletions

View File

@@ -45,6 +45,8 @@ import java.nio.charset.StandardCharsets;
import java.nio.file.Files;
import java.nio.file.Path;
import java.nio.file.Paths;
import java.nio.file.attribute.FileTime;
import java.time.Instant;
import java.util.HashMap;
import java.util.Map;
@@ -224,6 +226,11 @@ public class FileCreateUtils {
public static void createBaseFile(String basePath, String partitionPath, String instantTime, String fileId, long length)
throws Exception {
createBaseFile(basePath, partitionPath, instantTime, fileId, length, Instant.now().toEpochMilli());
}
public static void createBaseFile(String basePath, String partitionPath, String instantTime, String fileId, long length, long lastModificationTimeMilli)
throws Exception {
Path parentPath = Paths.get(basePath, partitionPath);
Files.createDirectories(parentPath);
Path baseFilePath = parentPath.resolve(baseFileName(instantTime, fileId));
@@ -231,6 +238,7 @@ public class FileCreateUtils {
Files.createFile(baseFilePath);
}
new RandomAccessFile(baseFilePath.toFile(), "rw").setLength(length);
Files.setLastModifiedTime(baseFilePath, FileTime.fromMillis(lastModificationTimeMilli));
}
public static void createLogFile(String basePath, String partitionPath, String instantTime, String fileId, int version)

View File

@@ -121,28 +121,30 @@ public class DFSPathSelector implements Serializable {
eligibleFiles.sort(Comparator.comparingLong(FileStatus::getModificationTime));
// Filter based on checkpoint & input size, if needed
long currentBytes = 0;
long maxModificationTime = Long.MIN_VALUE;
long newCheckpointTime = lastCheckpointTime;
List<FileStatus> filteredFiles = new ArrayList<>();
for (FileStatus f : eligibleFiles) {
if (currentBytes + f.getLen() >= sourceLimit) {
if (currentBytes + f.getLen() >= sourceLimit && f.getModificationTime() > newCheckpointTime) {
// we have enough data, we are done
// Also, we've read up to a file with a newer modification time
// so that some files with the same modification time won't be skipped in next read
break;
}
maxModificationTime = f.getModificationTime();
newCheckpointTime = f.getModificationTime();
currentBytes += f.getLen();
filteredFiles.add(f);
}
// no data to read
if (filteredFiles.isEmpty()) {
return new ImmutablePair<>(Option.empty(), lastCheckpointStr.orElseGet(() -> String.valueOf(Long.MIN_VALUE)));
return new ImmutablePair<>(Option.empty(), String.valueOf(newCheckpointTime));
}
// read the files out.
String pathStr = filteredFiles.stream().map(f -> f.getPath().toString()).collect(Collectors.joining(","));
return new ImmutablePair<>(Option.ofNullable(pathStr), String.valueOf(maxModificationTime));
return new ImmutablePair<>(Option.ofNullable(pathStr), String.valueOf(newCheckpointTime));
} catch (IOException ioe) {
throw new HoodieIOException("Unable to read from source from checkpoint: " + lastCheckpointStr, ioe);
}

View File

@@ -144,27 +144,30 @@ public class DatePartitionPathSelector extends DFSPathSelector {
// Filter based on checkpoint & input size, if needed
long currentBytes = 0;
long newCheckpointTime = lastCheckpointTime;
List<FileStatus> filteredFiles = new ArrayList<>();
for (FileStatus f : sortedEligibleFiles) {
if (currentBytes + f.getLen() >= sourceLimit) {
if (currentBytes + f.getLen() >= sourceLimit && f.getModificationTime() > newCheckpointTime) {
// we have enough data, we are done
// Also, we've read up to a file with a newer modification time
// so that some files with the same modification time won't be skipped in next read
break;
}
newCheckpointTime = f.getModificationTime();
currentBytes += f.getLen();
filteredFiles.add(f);
}
// no data to read
if (filteredFiles.isEmpty()) {
return new ImmutablePair<>(
Option.empty(), lastCheckpointStr.orElseGet(() -> String.valueOf(Long.MIN_VALUE)));
return new ImmutablePair<>(Option.empty(), String.valueOf(newCheckpointTime));
}
// read the files out.
String pathStr = filteredFiles.stream().map(f -> f.getPath().toString()).collect(Collectors.joining(","));
long maxModificationTime = filteredFiles.get(filteredFiles.size() - 1).getModificationTime();
return new ImmutablePair<>(Option.ofNullable(pathStr), String.valueOf(maxModificationTime));
return new ImmutablePair<>(Option.ofNullable(pathStr), String.valueOf(newCheckpointTime));
}
/**

View File

@@ -0,0 +1,161 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
package org.apache.hudi.utilities.sources.helpers;
import org.apache.hudi.common.config.TypedProperties;
import org.apache.hudi.common.util.Option;
import org.apache.hudi.common.util.ReflectionUtils;
import org.apache.hudi.common.util.collection.Pair;
import org.apache.hudi.testutils.HoodieClientTestHarness;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.Path;
import org.junit.jupiter.api.AfterEach;
import org.junit.jupiter.api.BeforeEach;
import org.junit.jupiter.params.ParameterizedTest;
import org.junit.jupiter.params.provider.ValueSource;
import java.nio.file.Paths;
import java.util.Arrays;
import java.util.List;
import java.util.stream.Collectors;
import static org.apache.hudi.common.testutils.FileCreateUtils.createBaseFile;
import static org.apache.hudi.utilities.sources.helpers.DFSPathSelector.Config.ROOT_INPUT_PATH_PROP;
import static org.apache.hudi.utilities.sources.helpers.DatePartitionPathSelector.Config.PARTITIONS_LIST_PARALLELISM;
import static org.junit.jupiter.api.Assertions.assertEquals;
import static org.junit.jupiter.api.Assertions.assertTrue;
public class TestDFSPathSelectorCommonMethods extends HoodieClientTestHarness {
TypedProperties props;
Path inputPath;
@BeforeEach
void setUp() {
initSparkContexts();
initPath();
initFileSystem();
props = new TypedProperties();
props.setProperty(ROOT_INPUT_PATH_PROP, basePath);
props.setProperty(PARTITIONS_LIST_PARALLELISM, "1");
inputPath = new Path(basePath);
}
@AfterEach
public void teardown() throws Exception {
cleanupResources();
}
@ParameterizedTest
@ValueSource(classes = {DFSPathSelector.class, DatePartitionPathSelector.class})
public void listEligibleFilesShouldIgnoreCertainPrefixes(Class<?> clazz) throws Exception {
DFSPathSelector selector = (DFSPathSelector) ReflectionUtils.loadClass(clazz.getName(), props, hadoopConf);
createBaseFile(basePath, "p1", "000", "foo1", 1);
createBaseFile(basePath, "p1", "000", ".foo2", 1);
createBaseFile(basePath, "p1", "000", "_foo3", 1);
List<FileStatus> eligibleFiles = selector.listEligibleFiles(fs, inputPath, 0);
assertEquals(1, eligibleFiles.size());
assertTrue(eligibleFiles.get(0).getPath().getName().startsWith("foo1"));
}
@ParameterizedTest
@ValueSource(classes = {DFSPathSelector.class, DatePartitionPathSelector.class})
public void listEligibleFilesShouldIgnore0LengthFiles(Class<?> clazz) throws Exception {
DFSPathSelector selector = (DFSPathSelector) ReflectionUtils.loadClass(clazz.getName(), props, hadoopConf);
createBaseFile(basePath, "p1", "000", "foo1", 1);
createBaseFile(basePath, "p1", "000", "foo2", 0);
createBaseFile(basePath, "p1", "000", "foo3", 0);
List<FileStatus> eligibleFiles = selector.listEligibleFiles(fs, inputPath, 0);
assertEquals(1, eligibleFiles.size());
assertTrue(eligibleFiles.get(0).getPath().getName().startsWith("foo1"));
}
@ParameterizedTest
@ValueSource(classes = {DFSPathSelector.class, DatePartitionPathSelector.class})
public void listEligibleFilesShouldIgnoreFilesEarlierThanCheckpointTime(Class<?> clazz) throws Exception {
DFSPathSelector selector = (DFSPathSelector) ReflectionUtils.loadClass(clazz.getName(), props, hadoopConf);
createBaseFile(basePath, "p1", "000", "foo1", 1);
createBaseFile(basePath, "p1", "000", "foo2", 1);
createBaseFile(basePath, "p1", "000", "foo3", 1);
List<FileStatus> eligibleFiles = selector.listEligibleFiles(fs, inputPath, Long.MAX_VALUE);
assertEquals(0, eligibleFiles.size());
}
@ParameterizedTest
@ValueSource(classes = {DFSPathSelector.class, DatePartitionPathSelector.class})
public void getNextFilePathsAndMaxModificationTimeShouldRespectSourceLimit(Class<?> clazz) throws Exception {
DFSPathSelector selector = (DFSPathSelector) ReflectionUtils.loadClass(clazz.getName(), props, hadoopConf);
createBaseFile(basePath, "p1", "000", "foo1", 10, 1000);
createBaseFile(basePath, "p1", "000", "foo2", 10, 2000);
createBaseFile(basePath, "p1", "000", "foo3", 10, 3000);
createBaseFile(basePath, "p1", "000", "foo4", 10, 4000);
createBaseFile(basePath, "p1", "000", "foo5", 10, 5000);
Pair<Option<String>, String> nextFilePathsAndCheckpoint = selector
.getNextFilePathsAndMaxModificationTime(jsc, Option.empty(), 30);
List<String> fileNames = Arrays
.stream(nextFilePathsAndCheckpoint.getLeft().get().split(","))
.map(p -> Paths.get(p).toFile().getName())
.sorted().collect(Collectors.toList());
assertEquals(2, fileNames.size());
assertTrue(fileNames.get(0).startsWith("foo1"));
assertTrue(fileNames.get(1).startsWith("foo2"));
String checkpointStr1stRead = nextFilePathsAndCheckpoint.getRight();
assertEquals(2000L, Long.parseLong(checkpointStr1stRead), "should read up to foo2 (inclusive)");
}
@ParameterizedTest
@ValueSource(classes = {DFSPathSelector.class, DatePartitionPathSelector.class})
public void getNextFilePathsAndMaxModificationTimeShouldIgnoreSourceLimitIfSameModTimeFilesPresent(Class<?> clazz) throws Exception {
DFSPathSelector selector = (DFSPathSelector) ReflectionUtils.loadClass(clazz.getName(), props, hadoopConf);
createBaseFile(basePath, "p1", "000", "foo1", 10, 1000);
createBaseFile(basePath, "p1", "000", "foo2", 10, 1000);
createBaseFile(basePath, "p1", "000", "foo3", 10, 1000);
createBaseFile(basePath, "p1", "000", "foo4", 10, 2000);
createBaseFile(basePath, "p1", "000", "foo5", 10, 2000);
Pair<Option<String>, String> nextFilePathsAndCheckpoint = selector
.getNextFilePathsAndMaxModificationTime(jsc, Option.empty(), 20);
List<String> fileNames1stRead = Arrays
.stream(nextFilePathsAndCheckpoint.getLeft().get().split(","))
.map(p -> Paths.get(p).toFile().getName())
.sorted().collect(Collectors.toList());
assertEquals(3, fileNames1stRead.size());
assertTrue(fileNames1stRead.get(0).startsWith("foo1"));
assertTrue(fileNames1stRead.get(1).startsWith("foo2"));
assertTrue(fileNames1stRead.get(2).startsWith("foo3"));
String checkpointStr1stRead = nextFilePathsAndCheckpoint.getRight();
assertEquals(1000L, Long.parseLong(checkpointStr1stRead), "should read up to foo3 (inclusive)");
nextFilePathsAndCheckpoint = selector
.getNextFilePathsAndMaxModificationTime(jsc, Option.of(checkpointStr1stRead), 20);
List<String> fileNames2ndRead = Arrays
.stream(nextFilePathsAndCheckpoint.getLeft().get().split(","))
.map(p -> Paths.get(p).toFile().getName())
.sorted().collect(Collectors.toList());
assertEquals(2, fileNames2ndRead.size());
assertTrue(fileNames2ndRead.get(0).startsWith("foo4"));
assertTrue(fileNames2ndRead.get(1).startsWith("foo5"));
String checkpointStr2ndRead = nextFilePathsAndCheckpoint.getRight();
assertEquals(2000L, Long.parseLong(checkpointStr2ndRead), "should read up to foo5 (inclusive)");
}
}