[HUDI-482] Fix missing @Override annotation on methods (#1156)
* [HUDI-482] Fix missing @Override annotation on methods
This commit is contained in:
@@ -30,6 +30,7 @@ import org.springframework.stereotype.Component;
|
|||||||
@Order(Ordered.HIGHEST_PRECEDENCE)
|
@Order(Ordered.HIGHEST_PRECEDENCE)
|
||||||
public class HoodieHistoryFileNameProvider extends DefaultHistoryFileNameProvider {
|
public class HoodieHistoryFileNameProvider extends DefaultHistoryFileNameProvider {
|
||||||
|
|
||||||
|
@Override
|
||||||
public String getHistoryFileName() {
|
public String getHistoryFileName() {
|
||||||
return "hoodie-cmd.log";
|
return "hoodie-cmd.log";
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -50,14 +50,17 @@ public class HoodieSplashScreen extends DefaultBannerProvider {
|
|||||||
+ "* *" + OsUtils.LINE_SEPARATOR
|
+ "* *" + OsUtils.LINE_SEPARATOR
|
||||||
+ "===================================================================" + OsUtils.LINE_SEPARATOR;
|
+ "===================================================================" + OsUtils.LINE_SEPARATOR;
|
||||||
|
|
||||||
|
@Override
|
||||||
public String getBanner() {
|
public String getBanner() {
|
||||||
return screen;
|
return screen;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public String getVersion() {
|
public String getVersion() {
|
||||||
return "1.0";
|
return "1.0";
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public String getWelcomeMessage() {
|
public String getWelcomeMessage() {
|
||||||
return "Welcome to Apache Hudi CLI. Please type help if you are looking for help. ";
|
return "Welcome to Apache Hudi CLI. Please type help if you are looking for help. ";
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -72,6 +72,7 @@ public abstract class AbstractHoodieClient implements Serializable, AutoCloseabl
|
|||||||
/**
|
/**
|
||||||
* Releases any resources used by the client.
|
* Releases any resources used by the client.
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
public void close() {
|
public void close() {
|
||||||
stopEmbeddedServerView(true);
|
stopEmbeddedServerView(true);
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -973,6 +973,7 @@ public class HoodieWriteClient<T extends HoodieRecordPayload> extends AbstractHo
|
|||||||
/**
|
/**
|
||||||
* Releases any resources used by the client.
|
* Releases any resources used by the client.
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
public void close() {
|
public void close() {
|
||||||
// Stop timeline-server if running
|
// Stop timeline-server if running
|
||||||
super.close();
|
super.close();
|
||||||
|
|||||||
@@ -48,6 +48,7 @@ public class SparkBoundedInMemoryExecutor<I, O, E> extends BoundedInMemoryExecut
|
|||||||
this.sparkThreadTaskContext = TaskContext.get();
|
this.sparkThreadTaskContext = TaskContext.get();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public void preExecute() {
|
public void preExecute() {
|
||||||
// Passing parent thread's TaskContext to newly launched thread for it to access original TaskContext properties.
|
// Passing parent thread's TaskContext to newly launched thread for it to access original TaskContext properties.
|
||||||
TaskContext$.MODULE$.setTaskContext(sparkThreadTaskContext);
|
TaskContext$.MODULE$.setTaskContext(sparkThreadTaskContext);
|
||||||
|
|||||||
@@ -88,6 +88,7 @@ public class BloomIndexFileInfo implements Serializable {
|
|||||||
return Objects.hashCode(fileId, minRecordKey, maxRecordKey);
|
return Objects.hashCode(fileId, minRecordKey, maxRecordKey);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public String toString() {
|
public String toString() {
|
||||||
final StringBuilder sb = new StringBuilder("BloomIndexFileInfo {");
|
final StringBuilder sb = new StringBuilder("BloomIndexFileInfo {");
|
||||||
sb.append(" fileId=").append(fileId);
|
sb.append(" fileId=").append(fileId);
|
||||||
|
|||||||
@@ -154,6 +154,7 @@ public class HBaseIndex<T extends HoodieRecordPayload> extends HoodieIndex<T> {
|
|||||||
*/
|
*/
|
||||||
private void addShutDownHook() {
|
private void addShutDownHook() {
|
||||||
Runtime.getRuntime().addShutdownHook(new Thread() {
|
Runtime.getRuntime().addShutdownHook(new Thread() {
|
||||||
|
@Override
|
||||||
public void run() {
|
public void run() {
|
||||||
try {
|
try {
|
||||||
hbaseConnection.close();
|
hbaseConnection.close();
|
||||||
@@ -167,6 +168,7 @@ public class HBaseIndex<T extends HoodieRecordPayload> extends HoodieIndex<T> {
|
|||||||
/**
|
/**
|
||||||
* Ensure that any resources used for indexing are released here.
|
* Ensure that any resources used for indexing are released here.
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
public void close() {
|
public void close() {
|
||||||
this.hBaseIndexQPSResourceAllocator.releaseQPSResources();
|
this.hBaseIndexQPSResourceAllocator.releaseQPSResources();
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -94,6 +94,7 @@ public class HoodieCreateHandle<T extends HoodieRecordPayload> extends HoodieWri
|
|||||||
/**
|
/**
|
||||||
* Perform the actual writing of the given record into the backing file.
|
* Perform the actual writing of the given record into the backing file.
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
public void write(HoodieRecord record, Option<IndexedRecord> avroRecord) {
|
public void write(HoodieRecord record, Option<IndexedRecord> avroRecord) {
|
||||||
Option recordMetadata = record.getData().getMetadata();
|
Option recordMetadata = record.getData().getMetadata();
|
||||||
try {
|
try {
|
||||||
|
|||||||
@@ -92,6 +92,7 @@ public class HoodieMergeHandle<T extends HoodieRecordPayload> extends HoodieWrit
|
|||||||
return HoodieAvroUtils.addMetadataFields(originalSchema);
|
return HoodieAvroUtils.addMetadataFields(originalSchema);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public Path makeNewPath(String partitionPath) {
|
public Path makeNewPath(String partitionPath) {
|
||||||
Path path = FSUtils.getPartitionPath(config.getBasePath(), partitionPath);
|
Path path = FSUtils.getPartitionPath(config.getBasePath(), partitionPath);
|
||||||
try {
|
try {
|
||||||
@@ -103,6 +104,7 @@ public class HoodieMergeHandle<T extends HoodieRecordPayload> extends HoodieWrit
|
|||||||
return new Path(path.toString(), FSUtils.makeDataFileName(instantTime, writeToken, fileId));
|
return new Path(path.toString(), FSUtils.makeDataFileName(instantTime, writeToken, fileId));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public Schema getWriterSchema() {
|
public Schema getWriterSchema() {
|
||||||
return writerSchema;
|
return writerSchema;
|
||||||
}
|
}
|
||||||
@@ -113,6 +115,7 @@ public class HoodieMergeHandle<T extends HoodieRecordPayload> extends HoodieWrit
|
|||||||
* - Whether it belongs to the same partitionPath as existing records - Whether the current file written bytes lt max
|
* - Whether it belongs to the same partitionPath as existing records - Whether the current file written bytes lt max
|
||||||
* file size
|
* file size
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
public boolean canWrite(HoodieRecord record) {
|
public boolean canWrite(HoodieRecord record) {
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
@@ -120,6 +123,7 @@ public class HoodieMergeHandle<T extends HoodieRecordPayload> extends HoodieWrit
|
|||||||
/**
|
/**
|
||||||
* Perform the actual writing of the given record into the backing file.
|
* Perform the actual writing of the given record into the backing file.
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
public void write(HoodieRecord record, Option<IndexedRecord> insertValue) {
|
public void write(HoodieRecord record, Option<IndexedRecord> insertValue) {
|
||||||
// NO_OP
|
// NO_OP
|
||||||
}
|
}
|
||||||
@@ -127,6 +131,7 @@ public class HoodieMergeHandle<T extends HoodieRecordPayload> extends HoodieWrit
|
|||||||
/**
|
/**
|
||||||
* Perform the actual writing of the given record into the backing file.
|
* Perform the actual writing of the given record into the backing file.
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
public void write(HoodieRecord record, Option<IndexedRecord> avroRecord, Option<Exception> exception) {
|
public void write(HoodieRecord record, Option<IndexedRecord> avroRecord, Option<Exception> exception) {
|
||||||
Option recordMetadata = record.getData().getMetadata();
|
Option recordMetadata = record.getData().getMetadata();
|
||||||
if (exception.isPresent() && exception.get() instanceof Throwable) {
|
if (exception.isPresent() && exception.get() instanceof Throwable) {
|
||||||
@@ -141,6 +146,7 @@ public class HoodieMergeHandle<T extends HoodieRecordPayload> extends HoodieWrit
|
|||||||
/**
|
/**
|
||||||
* Rewrite the GenericRecord with the Schema containing the Hoodie Metadata fields.
|
* Rewrite the GenericRecord with the Schema containing the Hoodie Metadata fields.
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
protected GenericRecord rewriteRecord(GenericRecord record) {
|
protected GenericRecord rewriteRecord(GenericRecord record) {
|
||||||
return HoodieAvroUtils.rewriteRecord(record, writerSchema);
|
return HoodieAvroUtils.rewriteRecord(record, writerSchema);
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -93,6 +93,7 @@ public class HoodieParquetWriter<T extends HoodieRecordPayload, R extends Indexe
|
|||||||
writeSupport.add(record.getRecordKey());
|
writeSupport.add(record.getRecordKey());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public boolean canWrite() {
|
public boolean canWrite() {
|
||||||
return fs.getBytesWritten(file) < maxFileSize;
|
return fs.getBytesWritten(file) < maxFileSize;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -279,6 +279,7 @@ public class HoodieCopyOnWriteTable<T extends HoodieRecordPayload> extends Hoodi
|
|||||||
* @param jsc JavaSparkContext
|
* @param jsc JavaSparkContext
|
||||||
* @return Cleaner Plan
|
* @return Cleaner Plan
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
public HoodieCleanerPlan scheduleClean(JavaSparkContext jsc) {
|
public HoodieCleanerPlan scheduleClean(JavaSparkContext jsc) {
|
||||||
try {
|
try {
|
||||||
HoodieCleanHelper cleaner = new HoodieCleanHelper(this, config);
|
HoodieCleanHelper cleaner = new HoodieCleanHelper(this, config);
|
||||||
|
|||||||
@@ -108,6 +108,7 @@ class HoodieLogFileReader implements HoodieLogFormat.Reader {
|
|||||||
*/
|
*/
|
||||||
private void addShutDownHook() {
|
private void addShutDownHook() {
|
||||||
Runtime.getRuntime().addShutdownHook(new Thread() {
|
Runtime.getRuntime().addShutdownHook(new Thread() {
|
||||||
|
@Override
|
||||||
public void run() {
|
public void run() {
|
||||||
try {
|
try {
|
||||||
close();
|
close();
|
||||||
|
|||||||
@@ -107,6 +107,7 @@ public class HoodieLogFormatWriter implements HoodieLogFormat.Writer {
|
|||||||
return fs;
|
return fs;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public HoodieLogFile getLogFile() {
|
public HoodieLogFile getLogFile() {
|
||||||
return logFile;
|
return logFile;
|
||||||
}
|
}
|
||||||
@@ -212,6 +213,7 @@ public class HoodieLogFormatWriter implements HoodieLogFormat.Writer {
|
|||||||
output.hsync();
|
output.hsync();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public long getCurrentSize() throws IOException {
|
public long getCurrentSize() throws IOException {
|
||||||
if (output == null) {
|
if (output == null) {
|
||||||
throw new IllegalStateException("Cannot get current size as the underlying stream has been closed already");
|
throw new IllegalStateException("Cannot get current size as the underlying stream has been closed already");
|
||||||
|
|||||||
@@ -145,6 +145,7 @@ public class HoodieActiveTimeline extends HoodieDefaultTimeline {
|
|||||||
* timeline * With Async compaction a requested/inflight compaction-instant is a valid baseInstant for a file-slice as
|
* timeline * With Async compaction a requested/inflight compaction-instant is a valid baseInstant for a file-slice as
|
||||||
* there could be delta-commits with that baseInstant.
|
* there could be delta-commits with that baseInstant.
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
public HoodieTimeline getCommitsAndCompactionTimeline() {
|
public HoodieTimeline getCommitsAndCompactionTimeline() {
|
||||||
return getTimelineOfActions(Sets.newHashSet(COMMIT_ACTION, DELTA_COMMIT_ACTION, COMPACTION_ACTION));
|
return getTimelineOfActions(Sets.newHashSet(COMMIT_ACTION, DELTA_COMMIT_ACTION, COMPACTION_ACTION));
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -182,6 +182,7 @@ public abstract class AbstractTableFileSystemView implements SyncableFileSystemV
|
|||||||
/**
|
/**
|
||||||
* Clears the partition Map and reset view states.
|
* Clears the partition Map and reset view states.
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
public final void reset() {
|
public final void reset() {
|
||||||
try {
|
try {
|
||||||
writeLock.lock();
|
writeLock.lock();
|
||||||
@@ -380,6 +381,7 @@ public abstract class AbstractTableFileSystemView implements SyncableFileSystemV
|
|||||||
/**
|
/**
|
||||||
* Get Latest data file for a partition and file-Id.
|
* Get Latest data file for a partition and file-Id.
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
public final Option<HoodieDataFile> getLatestDataFile(String partitionStr, String fileId) {
|
public final Option<HoodieDataFile> getLatestDataFile(String partitionStr, String fileId) {
|
||||||
try {
|
try {
|
||||||
readLock.lock();
|
readLock.lock();
|
||||||
@@ -434,6 +436,7 @@ public abstract class AbstractTableFileSystemView implements SyncableFileSystemV
|
|||||||
/**
|
/**
|
||||||
* Get Latest File Slice for a given fileId in a given partition.
|
* Get Latest File Slice for a given fileId in a given partition.
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
public final Option<FileSlice> getLatestFileSlice(String partitionStr, String fileId) {
|
public final Option<FileSlice> getLatestFileSlice(String partitionStr, String fileId) {
|
||||||
try {
|
try {
|
||||||
readLock.lock();
|
readLock.lock();
|
||||||
|
|||||||
@@ -205,6 +205,7 @@ public class HoodieTableFileSystemView extends IncrementalTimelineSyncFileSystem
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public void close() {
|
public void close() {
|
||||||
closed = true;
|
closed = true;
|
||||||
super.reset();
|
super.reset();
|
||||||
@@ -212,6 +213,7 @@ public class HoodieTableFileSystemView extends IncrementalTimelineSyncFileSystem
|
|||||||
fgIdToPendingCompaction = null;
|
fgIdToPendingCompaction = null;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public boolean isClosed() {
|
public boolean isClosed() {
|
||||||
return closed;
|
return closed;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -76,6 +76,7 @@ public class SpillableMapBasedFileSystemView extends HoodieTableFileSystemView {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
protected Map<HoodieFileGroupId, Pair<String, CompactionOperation>> createFileIdToPendingCompactionMap(
|
protected Map<HoodieFileGroupId, Pair<String, CompactionOperation>> createFileIdToPendingCompactionMap(
|
||||||
Map<HoodieFileGroupId, Pair<String, CompactionOperation>> fgIdToPendingCompaction) {
|
Map<HoodieFileGroupId, Pair<String, CompactionOperation>> fgIdToPendingCompaction) {
|
||||||
try {
|
try {
|
||||||
@@ -91,6 +92,7 @@ public class SpillableMapBasedFileSystemView extends HoodieTableFileSystemView {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public Stream<HoodieFileGroup> getAllFileGroups() {
|
public Stream<HoodieFileGroup> getAllFileGroups() {
|
||||||
return ((ExternalSpillableMap) partitionToFileGroupsMap).valueStream()
|
return ((ExternalSpillableMap) partitionToFileGroupsMap).valueStream()
|
||||||
.flatMap(fg -> ((List<HoodieFileGroup>) fg).stream());
|
.flatMap(fg -> ((List<HoodieFileGroup>) fg).stream());
|
||||||
|
|||||||
@@ -124,6 +124,7 @@ public class ObjectSizeCalculator {
|
|||||||
|
|
||||||
private final LoadingCache<Class<?>, ClassSizeInfo> classSizeInfos =
|
private final LoadingCache<Class<?>, ClassSizeInfo> classSizeInfos =
|
||||||
CacheBuilder.newBuilder().build(new CacheLoader<Class<?>, ClassSizeInfo>() {
|
CacheBuilder.newBuilder().build(new CacheLoader<Class<?>, ClassSizeInfo>() {
|
||||||
|
@Override
|
||||||
public ClassSizeInfo load(Class<?> clazz) {
|
public ClassSizeInfo load(Class<?> clazz) {
|
||||||
return new ClassSizeInfo(clazz);
|
return new ClassSizeInfo(clazz);
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -125,6 +125,7 @@ public final class DiskBasedMap<T extends Serializable, R extends Serializable>
|
|||||||
*/
|
*/
|
||||||
private void addShutDownHook() {
|
private void addShutDownHook() {
|
||||||
Runtime.getRuntime().addShutdownHook(new Thread() {
|
Runtime.getRuntime().addShutdownHook(new Thread() {
|
||||||
|
@Override
|
||||||
public void run() {
|
public void run() {
|
||||||
try {
|
try {
|
||||||
if (writeOnlyFileHandle != null) {
|
if (writeOnlyFileHandle != null) {
|
||||||
|
|||||||
@@ -115,6 +115,7 @@ public class LazyFileIterable<T, R> implements Iterable<R> {
|
|||||||
|
|
||||||
private void addShutdownHook() {
|
private void addShutdownHook() {
|
||||||
Runtime.getRuntime().addShutdownHook(new Thread() {
|
Runtime.getRuntime().addShutdownHook(new Thread() {
|
||||||
|
@Override
|
||||||
public void run() {
|
public void run() {
|
||||||
close();
|
close();
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -173,10 +173,12 @@ public class HoodieParquetInputFormat extends MapredParquetInputFormat implement
|
|||||||
return grouped;
|
return grouped;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public void setConf(Configuration conf) {
|
public void setConf(Configuration conf) {
|
||||||
this.conf = conf;
|
this.conf = conf;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public Configuration getConf() {
|
public Configuration getConf() {
|
||||||
return conf;
|
return conf;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -805,6 +805,7 @@ public class HoodieCombineHiveInputFormat<K extends WritableComparable, V extend
|
|||||||
|
|
||||||
public HoodieCombineFileInputFormatShim() {}
|
public HoodieCombineFileInputFormatShim() {}
|
||||||
|
|
||||||
|
@Override
|
||||||
public Path[] getInputPathsShim(JobConf conf) {
|
public Path[] getInputPathsShim(JobConf conf) {
|
||||||
try {
|
try {
|
||||||
return FileInputFormat.getInputPaths(conf);
|
return FileInputFormat.getInputPaths(conf);
|
||||||
@@ -813,6 +814,7 @@ public class HoodieCombineHiveInputFormat<K extends WritableComparable, V extend
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public void createPool(JobConf conf, PathFilter... filters) {
|
public void createPool(JobConf conf, PathFilter... filters) {
|
||||||
super.createPool(conf, filters);
|
super.createPool(conf, filters);
|
||||||
}
|
}
|
||||||
@@ -822,6 +824,7 @@ public class HoodieCombineHiveInputFormat<K extends WritableComparable, V extend
|
|||||||
throw new IOException("CombineFileInputFormat.getRecordReader not needed.");
|
throw new IOException("CombineFileInputFormat.getRecordReader not needed.");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
protected List<FileStatus> listStatus(JobContext job) throws IOException {
|
protected List<FileStatus> listStatus(JobContext job) throws IOException {
|
||||||
LOG.info("Listing status in HoodieCombineHiveInputFormat.HoodieCombineFileInputFormatShim");
|
LOG.info("Listing status in HoodieCombineHiveInputFormat.HoodieCombineFileInputFormatShim");
|
||||||
List<FileStatus> result;
|
List<FileStatus> result;
|
||||||
@@ -851,6 +854,7 @@ public class HoodieCombineHiveInputFormat<K extends WritableComparable, V extend
|
|||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public CombineFileSplit[] getSplits(JobConf job, int numSplits) throws IOException {
|
public CombineFileSplit[] getSplits(JobConf job, int numSplits) throws IOException {
|
||||||
long minSize = job.getLong(org.apache.hadoop.mapreduce.lib.input.FileInputFormat.SPLIT_MINSIZE, 0L);
|
long minSize = job.getLong(org.apache.hadoop.mapreduce.lib.input.FileInputFormat.SPLIT_MINSIZE, 0L);
|
||||||
if (job.getLong("mapreduce.input.fileinputformat.split.minsize.per.node", 0L) == 0L) {
|
if (job.getLong("mapreduce.input.fileinputformat.split.minsize.per.node", 0L) == 0L) {
|
||||||
@@ -879,10 +883,12 @@ public class HoodieCombineHiveInputFormat<K extends WritableComparable, V extend
|
|||||||
return (CombineFileSplit[]) inputSplitShims.toArray(new HadoopShimsSecure.InputSplitShim[inputSplitShims.size()]);
|
return (CombineFileSplit[]) inputSplitShims.toArray(new HadoopShimsSecure.InputSplitShim[inputSplitShims.size()]);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public HadoopShimsSecure.InputSplitShim getInputSplitShim() throws IOException {
|
public HadoopShimsSecure.InputSplitShim getInputSplitShim() throws IOException {
|
||||||
return new HadoopShimsSecure.InputSplitShim();
|
return new HadoopShimsSecure.InputSplitShim();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
public RecordReader getRecordReader(JobConf job, CombineFileSplit split, Reporter reporter,
|
public RecordReader getRecordReader(JobConf job, CombineFileSplit split, Reporter reporter,
|
||||||
Class<RecordReader<K, V>> rrClass) throws IOException {
|
Class<RecordReader<K, V>> rrClass) throws IOException {
|
||||||
return new HadoopShimsSecure.CombineFileRecordReader(job, split, reporter, rrClass);
|
return new HadoopShimsSecure.CombineFileRecordReader(job, split, reporter, rrClass);
|
||||||
|
|||||||
@@ -554,6 +554,7 @@ public class HoodieDeltaStreamer implements Serializable {
|
|||||||
/**
|
/**
|
||||||
* Start Compaction Service.
|
* Start Compaction Service.
|
||||||
*/
|
*/
|
||||||
|
@Override
|
||||||
protected Pair<CompletableFuture, ExecutorService> startService() {
|
protected Pair<CompletableFuture, ExecutorService> startService() {
|
||||||
ExecutorService executor = Executors.newFixedThreadPool(maxConcurrentCompaction);
|
ExecutorService executor = Executors.newFixedThreadPool(maxConcurrentCompaction);
|
||||||
List<CompletableFuture<Boolean>> compactionFutures =
|
List<CompletableFuture<Boolean>> compactionFutures =
|
||||||
|
|||||||
Reference in New Issue
Block a user