[HUDI-1837] Add optional instant range to log record scanner for log (#2870)
This commit is contained in:
@@ -35,6 +35,7 @@ import org.apache.hudi.common.table.log.block.HoodieDeleteBlock;
|
||||
import org.apache.hudi.common.table.log.block.HoodieHFileDataBlock;
|
||||
import org.apache.hudi.common.table.log.block.HoodieLogBlock;
|
||||
import org.apache.hudi.common.table.timeline.HoodieTimeline;
|
||||
import org.apache.hudi.common.util.Option;
|
||||
import org.apache.hudi.common.util.SpillableMapUtils;
|
||||
import org.apache.hudi.exception.HoodieException;
|
||||
import org.apache.hudi.exception.HoodieIOException;
|
||||
@@ -88,6 +89,8 @@ public abstract class AbstractHoodieLogRecordScanner {
|
||||
private final boolean reverseReader;
|
||||
// Buffer Size for log file reader
|
||||
private final int bufferSize;
|
||||
// optional instant range for incremental block filtering
|
||||
private final Option<InstantRange> instantRange;
|
||||
// FileSystem
|
||||
private final FileSystem fs;
|
||||
// Total log files read - for metrics
|
||||
@@ -105,8 +108,8 @@ public abstract class AbstractHoodieLogRecordScanner {
|
||||
// Progress
|
||||
private float progress = 0.0f;
|
||||
|
||||
public AbstractHoodieLogRecordScanner(FileSystem fs, String basePath, List<String> logFilePaths, Schema readerSchema,
|
||||
String latestInstantTime, boolean readBlocksLazily, boolean reverseReader, int bufferSize) {
|
||||
protected AbstractHoodieLogRecordScanner(FileSystem fs, String basePath, List<String> logFilePaths, Schema readerSchema,
|
||||
String latestInstantTime, boolean readBlocksLazily, boolean reverseReader, int bufferSize, Option<InstantRange> instantRange) {
|
||||
this.readerSchema = readerSchema;
|
||||
this.latestInstantTime = latestInstantTime;
|
||||
this.hoodieTableMetaClient = HoodieTableMetaClient.builder().setConf(fs.getConf()).setBasePath(basePath).build();
|
||||
@@ -118,6 +121,7 @@ public abstract class AbstractHoodieLogRecordScanner {
|
||||
this.reverseReader = reverseReader;
|
||||
this.fs = fs;
|
||||
this.bufferSize = bufferSize;
|
||||
this.instantRange = instantRange;
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -141,6 +145,7 @@ public abstract class AbstractHoodieLogRecordScanner {
|
||||
totalLogFiles.set(scannedLogFiles.size());
|
||||
// Use the HoodieLogFileReader to iterate through the blocks in the log file
|
||||
HoodieLogBlock r = logFormatReaderWrapper.next();
|
||||
final String instantTime = r.getLogBlockHeader().get(INSTANT_TIME);
|
||||
totalLogBlocks.incrementAndGet();
|
||||
if (r.getBlockType() != CORRUPT_BLOCK
|
||||
&& !HoodieTimeline.compareTimestamps(r.getLogBlockHeader().get(INSTANT_TIME), HoodieTimeline.LESSER_THAN_OR_EQUALS, this.latestInstantTime
|
||||
@@ -149,12 +154,15 @@ public abstract class AbstractHoodieLogRecordScanner {
|
||||
break;
|
||||
}
|
||||
if (r.getBlockType() != CORRUPT_BLOCK && r.getBlockType() != COMMAND_BLOCK) {
|
||||
String instantTime = r.getLogBlockHeader().get(INSTANT_TIME);
|
||||
if (!completedInstantsTimeline.containsOrBeforeTimelineStarts(instantTime)
|
||||
|| inflightInstantsTimeline.containsInstant(instantTime)) {
|
||||
// hit an uncommitted block possibly from a failed write, move to the next one and skip processing this one
|
||||
continue;
|
||||
}
|
||||
if (instantRange.isPresent() && !instantRange.get().isInRange(instantTime)) {
|
||||
// filter the log block by instant range
|
||||
continue;
|
||||
}
|
||||
}
|
||||
switch (r.getBlockType()) {
|
||||
case HFILE_DATA_BLOCK:
|
||||
@@ -392,6 +400,10 @@ public abstract class AbstractHoodieLogRecordScanner {
|
||||
|
||||
public abstract Builder withBufferSize(int bufferSize);
|
||||
|
||||
public Builder withInstantRange(Option<InstantRange> instantRange) {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
|
||||
public abstract AbstractHoodieLogRecordScanner build();
|
||||
}
|
||||
}
|
||||
|
||||
@@ -24,6 +24,7 @@ import org.apache.hudi.common.model.HoodieRecordPayload;
|
||||
import org.apache.hudi.common.util.DefaultSizeEstimator;
|
||||
import org.apache.hudi.common.util.HoodieRecordSizeEstimator;
|
||||
import org.apache.hudi.common.util.HoodieTimer;
|
||||
import org.apache.hudi.common.util.Option;
|
||||
import org.apache.hudi.common.util.SpillableMapUtils;
|
||||
import org.apache.hudi.common.util.collection.ExternalSpillableMap;
|
||||
import org.apache.hudi.exception.HoodieIOException;
|
||||
@@ -69,18 +70,11 @@ public class HoodieMergedLogRecordScanner extends AbstractHoodieLogRecordScanner
|
||||
public final HoodieTimer timer = new HoodieTimer();
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
public HoodieMergedLogRecordScanner(FileSystem fs, String basePath, List<String> logFilePaths, Schema readerSchema,
|
||||
protected HoodieMergedLogRecordScanner(FileSystem fs, String basePath, List<String> logFilePaths, Schema readerSchema,
|
||||
String latestInstantTime, Long maxMemorySizeInBytes, boolean readBlocksLazily,
|
||||
boolean reverseReader, int bufferSize, String spillableMapBasePath) {
|
||||
this(fs, basePath, logFilePaths, readerSchema, latestInstantTime, maxMemorySizeInBytes, readBlocksLazily,
|
||||
reverseReader, bufferSize, spillableMapBasePath, true);
|
||||
}
|
||||
|
||||
@SuppressWarnings("unchecked")
|
||||
public HoodieMergedLogRecordScanner(FileSystem fs, String basePath, List<String> logFilePaths, Schema readerSchema,
|
||||
String latestInstantTime, Long maxMemorySizeInBytes, boolean readBlocksLazily,
|
||||
boolean reverseReader, int bufferSize, String spillableMapBasePath, boolean autoScan) {
|
||||
super(fs, basePath, logFilePaths, readerSchema, latestInstantTime, readBlocksLazily, reverseReader, bufferSize);
|
||||
boolean reverseReader, int bufferSize, String spillableMapBasePath,
|
||||
Option<InstantRange> instantRange, boolean autoScan) {
|
||||
super(fs, basePath, logFilePaths, readerSchema, latestInstantTime, readBlocksLazily, reverseReader, bufferSize, instantRange);
|
||||
try {
|
||||
// Store merged records for all versions for this log file, set the in-memory footprint to maxInMemoryMapSize
|
||||
this.records = new ExternalSpillableMap<>(maxMemorySizeInBytes, spillableMapBasePath, new DefaultSizeEstimator(),
|
||||
@@ -163,17 +157,21 @@ public class HoodieMergedLogRecordScanner extends AbstractHoodieLogRecordScanner
|
||||
* Builder used to build {@code HoodieUnMergedLogRecordScanner}.
|
||||
*/
|
||||
public static class Builder extends AbstractHoodieLogRecordScanner.Builder {
|
||||
private FileSystem fs;
|
||||
private String basePath;
|
||||
private List<String> logFilePaths;
|
||||
private Schema readerSchema;
|
||||
private String latestInstantTime;
|
||||
private boolean readBlocksLazily;
|
||||
private boolean reverseReader;
|
||||
private int bufferSize;
|
||||
protected FileSystem fs;
|
||||
protected String basePath;
|
||||
protected List<String> logFilePaths;
|
||||
protected Schema readerSchema;
|
||||
protected String latestInstantTime;
|
||||
protected boolean readBlocksLazily;
|
||||
protected boolean reverseReader;
|
||||
protected int bufferSize;
|
||||
// specific configurations
|
||||
private Long maxMemorySizeInBytes;
|
||||
private String spillableMapBasePath;
|
||||
protected Long maxMemorySizeInBytes;
|
||||
protected String spillableMapBasePath;
|
||||
// incremental filtering
|
||||
private Option<InstantRange> instantRange = Option.empty();
|
||||
// auto scan default true
|
||||
private boolean autoScan = true;
|
||||
|
||||
public Builder withFileSystem(FileSystem fs) {
|
||||
this.fs = fs;
|
||||
@@ -215,6 +213,11 @@ public class HoodieMergedLogRecordScanner extends AbstractHoodieLogRecordScanner
|
||||
return this;
|
||||
}
|
||||
|
||||
public Builder withInstantRange(Option<InstantRange> instantRange) {
|
||||
this.instantRange = instantRange;
|
||||
return this;
|
||||
}
|
||||
|
||||
public Builder withMaxMemorySizeInBytes(Long maxMemorySizeInBytes) {
|
||||
this.maxMemorySizeInBytes = maxMemorySizeInBytes;
|
||||
return this;
|
||||
@@ -225,11 +228,16 @@ public class HoodieMergedLogRecordScanner extends AbstractHoodieLogRecordScanner
|
||||
return this;
|
||||
}
|
||||
|
||||
public Builder withAutoScan(boolean autoScan) {
|
||||
this.autoScan = autoScan;
|
||||
return this;
|
||||
}
|
||||
|
||||
@Override
|
||||
public HoodieMergedLogRecordScanner build() {
|
||||
return new HoodieMergedLogRecordScanner(fs, basePath, logFilePaths, readerSchema,
|
||||
latestInstantTime, maxMemorySizeInBytes, readBlocksLazily, reverseReader,
|
||||
bufferSize, spillableMapBasePath);
|
||||
bufferSize, spillableMapBasePath, instantRange, autoScan);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -21,6 +21,7 @@ package org.apache.hudi.common.table.log;
|
||||
import org.apache.hudi.common.model.HoodieKey;
|
||||
import org.apache.hudi.common.model.HoodieRecord;
|
||||
import org.apache.hudi.common.model.HoodieRecordPayload;
|
||||
import org.apache.hudi.common.util.Option;
|
||||
|
||||
import org.apache.avro.Schema;
|
||||
import org.apache.hadoop.fs.FileSystem;
|
||||
@@ -34,10 +35,9 @@ public class HoodieUnMergedLogRecordScanner extends AbstractHoodieLogRecordScann
|
||||
|
||||
private final LogRecordScannerCallback callback;
|
||||
|
||||
public HoodieUnMergedLogRecordScanner(FileSystem fs, String basePath, List<String> logFilePaths, Schema readerSchema,
|
||||
String latestInstantTime, boolean readBlocksLazily, boolean reverseReader, int bufferSize,
|
||||
LogRecordScannerCallback callback) {
|
||||
super(fs, basePath, logFilePaths, readerSchema, latestInstantTime, readBlocksLazily, reverseReader, bufferSize);
|
||||
private HoodieUnMergedLogRecordScanner(FileSystem fs, String basePath, List<String> logFilePaths, Schema readerSchema,
|
||||
String latestInstantTime, boolean readBlocksLazily, boolean reverseReader, int bufferSize, LogRecordScannerCallback callback) {
|
||||
super(fs, basePath, logFilePaths, readerSchema, latestInstantTime, readBlocksLazily, reverseReader, bufferSize, Option.empty());
|
||||
this.callback = callback;
|
||||
}
|
||||
|
||||
|
||||
@@ -0,0 +1,101 @@
|
||||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hudi.common.table.log;
|
||||
|
||||
import org.apache.hudi.common.table.timeline.HoodieTimeline;
|
||||
|
||||
import java.io.Serializable;
|
||||
import java.util.Objects;
|
||||
|
||||
/**
|
||||
* A instant commits range used for incremental reader filtering.
|
||||
*/
|
||||
public abstract class InstantRange implements Serializable {
|
||||
private static final long serialVersionUID = 1L;
|
||||
|
||||
protected final String startInstant;
|
||||
protected final String endInstant;
|
||||
|
||||
public InstantRange(String startInstant, String endInstant) {
|
||||
this.startInstant = Objects.requireNonNull(startInstant);
|
||||
this.endInstant = Objects.requireNonNull(endInstant);
|
||||
}
|
||||
|
||||
public static InstantRange getInstance(String startInstant, String endInstant, RangeType rangeType) {
|
||||
switch (rangeType) {
|
||||
case OPEN_CLOSE:
|
||||
return new OpenCloseRange(startInstant, endInstant);
|
||||
case CLOSE_CLOSE:
|
||||
return new CloseCloseRange(startInstant, endInstant);
|
||||
default:
|
||||
throw new AssertionError();
|
||||
}
|
||||
}
|
||||
|
||||
public String getStartInstant() {
|
||||
return startInstant;
|
||||
}
|
||||
|
||||
public String getEndInstant() {
|
||||
return endInstant;
|
||||
}
|
||||
|
||||
public abstract boolean isInRange(String instant);
|
||||
|
||||
// -------------------------------------------------------------------------
|
||||
// Inner Class
|
||||
// -------------------------------------------------------------------------
|
||||
|
||||
/**
|
||||
* Represents a range type.
|
||||
*/
|
||||
public enum RangeType {
|
||||
OPEN_CLOSE, CLOSE_CLOSE
|
||||
}
|
||||
|
||||
private static class OpenCloseRange extends InstantRange {
|
||||
|
||||
public OpenCloseRange(String startInstant, String endInstant) {
|
||||
super(startInstant, endInstant);
|
||||
}
|
||||
|
||||
@Override
|
||||
public boolean isInRange(String instant) {
|
||||
// No need to do comparison:
|
||||
// HoodieTimeline.compareTimestamps(instant, HoodieTimeline.LESSER_THAN_OR_EQUALS, endInstant)
|
||||
// because the logic is ensured by the log scanner
|
||||
return HoodieTimeline.compareTimestamps(instant, HoodieTimeline.GREATER_THAN, startInstant);
|
||||
}
|
||||
}
|
||||
|
||||
private static class CloseCloseRange extends InstantRange {
|
||||
|
||||
public CloseCloseRange(String startInstant, String endInstant) {
|
||||
super(startInstant, endInstant);
|
||||
}
|
||||
|
||||
@Override
|
||||
public boolean isInRange(String instant) {
|
||||
// No need to do comparison:
|
||||
// HoodieTimeline.compareTimestamps(instant, HoodieTimeline.LESSER_THAN_OR_EQUALS, endInstant)
|
||||
// because the logic is ensured by the log scanner
|
||||
return HoodieTimeline.compareTimestamps(instant, HoodieTimeline.GREATER_THAN_OR_EQUALS, startInstant);
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -214,8 +214,16 @@ public class HoodieBackedTableMetadata extends BaseTableMetadata {
|
||||
|
||||
// Load the schema
|
||||
Schema schema = HoodieAvroUtils.addMetadataFields(HoodieMetadataRecord.getClassSchema());
|
||||
logRecordScanner = new HoodieMetadataMergedLogRecordScanner(metaClient.getFs(), metadataBasePath, logFilePaths,
|
||||
schema, latestMetaInstantTimestamp, MAX_MEMORY_SIZE_IN_BYTES, BUFFER_SIZE, spillableMapDirectory, null);
|
||||
logRecordScanner = HoodieMetadataMergedLogRecordScanner.newBuilder()
|
||||
.withFileSystem(metaClient.getFs())
|
||||
.withBasePath(metadataBasePath)
|
||||
.withLogFilePaths(logFilePaths)
|
||||
.withReaderSchema(schema)
|
||||
.withLatestInstantTime(latestMetaInstantTimestamp)
|
||||
.withMaxMemorySizeInBytes(MAX_MEMORY_SIZE_IN_BYTES)
|
||||
.withBufferSize(BUFFER_SIZE)
|
||||
.withSpillableMapBasePath(spillableMapDirectory)
|
||||
.build();
|
||||
|
||||
logScannerOpenMs = timer.endTimer();
|
||||
LOG.info(String.format("Opened metadata log files from %s at instant (dataset instant=%s, metadata instant=%s) in %d ms",
|
||||
|
||||
@@ -39,12 +39,12 @@ public class HoodieMetadataMergedLogRecordScanner extends HoodieMergedLogRecordS
|
||||
// Set of all record keys that are to be read in memory
|
||||
private Set<String> mergeKeyFilter;
|
||||
|
||||
public HoodieMetadataMergedLogRecordScanner(FileSystem fs, String basePath, List<String> logFilePaths,
|
||||
private HoodieMetadataMergedLogRecordScanner(FileSystem fs, String basePath, List<String> logFilePaths,
|
||||
Schema readerSchema, String latestInstantTime, Long maxMemorySizeInBytes, int bufferSize,
|
||||
String spillableMapBasePath, Set<String> mergeKeyFilter) {
|
||||
super(fs, basePath, logFilePaths, readerSchema, latestInstantTime, maxMemorySizeInBytes, false, false, bufferSize,
|
||||
spillableMapBasePath, false);
|
||||
this.mergeKeyFilter = mergeKeyFilter != null ? mergeKeyFilter : Collections.emptySet();
|
||||
spillableMapBasePath, Option.empty(), false);
|
||||
this.mergeKeyFilter = mergeKeyFilter;
|
||||
|
||||
performScan();
|
||||
}
|
||||
@@ -63,6 +63,13 @@ public class HoodieMetadataMergedLogRecordScanner extends HoodieMergedLogRecordS
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns the builder for {@code HoodieMetadataMergedLogRecordScanner}.
|
||||
*/
|
||||
public static HoodieMetadataMergedLogRecordScanner.Builder newBuilder() {
|
||||
return new HoodieMetadataMergedLogRecordScanner.Builder();
|
||||
}
|
||||
|
||||
/**
|
||||
* Retrieve a record given its key.
|
||||
*
|
||||
@@ -72,4 +79,70 @@ public class HoodieMetadataMergedLogRecordScanner extends HoodieMergedLogRecordS
|
||||
public Option<HoodieRecord<HoodieMetadataPayload>> getRecordByKey(String key) {
|
||||
return Option.ofNullable((HoodieRecord) records.get(key));
|
||||
}
|
||||
|
||||
/**
|
||||
* Builder used to build {@code HoodieMetadataMergedLogRecordScanner}.
|
||||
*/
|
||||
public static class Builder extends HoodieMergedLogRecordScanner.Builder {
|
||||
private Set<String> mergeKeyFilter = Collections.emptySet();
|
||||
|
||||
public Builder withFileSystem(FileSystem fs) {
|
||||
this.fs = fs;
|
||||
return this;
|
||||
}
|
||||
|
||||
public Builder withBasePath(String basePath) {
|
||||
this.basePath = basePath;
|
||||
return this;
|
||||
}
|
||||
|
||||
public Builder withLogFilePaths(List<String> logFilePaths) {
|
||||
this.logFilePaths = logFilePaths;
|
||||
return this;
|
||||
}
|
||||
|
||||
public Builder withReaderSchema(Schema schema) {
|
||||
this.readerSchema = schema;
|
||||
return this;
|
||||
}
|
||||
|
||||
public Builder withLatestInstantTime(String latestInstantTime) {
|
||||
this.latestInstantTime = latestInstantTime;
|
||||
return this;
|
||||
}
|
||||
|
||||
public Builder withReadBlocksLazily(boolean readBlocksLazily) {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
|
||||
public Builder withReverseReader(boolean reverseReader) {
|
||||
throw new UnsupportedOperationException();
|
||||
}
|
||||
|
||||
public Builder withBufferSize(int bufferSize) {
|
||||
this.bufferSize = bufferSize;
|
||||
return this;
|
||||
}
|
||||
|
||||
public Builder withMaxMemorySizeInBytes(Long maxMemorySizeInBytes) {
|
||||
this.maxMemorySizeInBytes = maxMemorySizeInBytes;
|
||||
return this;
|
||||
}
|
||||
|
||||
public Builder withSpillableMapBasePath(String spillableMapBasePath) {
|
||||
this.spillableMapBasePath = spillableMapBasePath;
|
||||
return this;
|
||||
}
|
||||
|
||||
public Builder withMergeKeyFilter(Set<String> mergeKeyFilter) {
|
||||
this.mergeKeyFilter = mergeKeyFilter;
|
||||
return this;
|
||||
}
|
||||
|
||||
@Override
|
||||
public HoodieMetadataMergedLogRecordScanner build() {
|
||||
return new HoodieMetadataMergedLogRecordScanner(fs, basePath, logFilePaths, readerSchema,
|
||||
latestInstantTime, maxMemorySizeInBytes, bufferSize, spillableMapBasePath, mergeKeyFilter);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user