diff --git a/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieHistoryFileNameProvider.java b/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieHistoryFileNameProvider.java index af01f6675..95f983416 100644 --- a/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieHistoryFileNameProvider.java +++ b/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieHistoryFileNameProvider.java @@ -30,6 +30,7 @@ import org.springframework.stereotype.Component; @Order(Ordered.HIGHEST_PRECEDENCE) public class HoodieHistoryFileNameProvider extends DefaultHistoryFileNameProvider { + @Override public String getHistoryFileName() { return "hoodie-cmd.log"; } diff --git a/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieSplashScreen.java b/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieSplashScreen.java index b7f016b50..f2a458c19 100644 --- a/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieSplashScreen.java +++ b/hudi-cli/src/main/java/org/apache/hudi/cli/HoodieSplashScreen.java @@ -50,14 +50,17 @@ public class HoodieSplashScreen extends DefaultBannerProvider { + "* *" + OsUtils.LINE_SEPARATOR + "===================================================================" + OsUtils.LINE_SEPARATOR; + @Override public String getBanner() { return screen; } + @Override public String getVersion() { return "1.0"; } + @Override public String getWelcomeMessage() { return "Welcome to Apache Hudi CLI. Please type help if you are looking for help. "; } diff --git a/hudi-client/src/main/java/org/apache/hudi/AbstractHoodieClient.java b/hudi-client/src/main/java/org/apache/hudi/AbstractHoodieClient.java index 1d60840ae..dd108be02 100644 --- a/hudi-client/src/main/java/org/apache/hudi/AbstractHoodieClient.java +++ b/hudi-client/src/main/java/org/apache/hudi/AbstractHoodieClient.java @@ -72,6 +72,7 @@ public abstract class AbstractHoodieClient implements Serializable, AutoCloseabl /** * Releases any resources used by the client. */ + @Override public void close() { stopEmbeddedServerView(true); } diff --git a/hudi-client/src/main/java/org/apache/hudi/HoodieWriteClient.java b/hudi-client/src/main/java/org/apache/hudi/HoodieWriteClient.java index 0f5411648..9d0332d62 100644 --- a/hudi-client/src/main/java/org/apache/hudi/HoodieWriteClient.java +++ b/hudi-client/src/main/java/org/apache/hudi/HoodieWriteClient.java @@ -973,6 +973,7 @@ public class HoodieWriteClient extends AbstractHo /** * Releases any resources used by the client. */ + @Override public void close() { // Stop timeline-server if running super.close(); diff --git a/hudi-client/src/main/java/org/apache/hudi/func/SparkBoundedInMemoryExecutor.java b/hudi-client/src/main/java/org/apache/hudi/func/SparkBoundedInMemoryExecutor.java index 9741f7ce9..a576e3cc3 100644 --- a/hudi-client/src/main/java/org/apache/hudi/func/SparkBoundedInMemoryExecutor.java +++ b/hudi-client/src/main/java/org/apache/hudi/func/SparkBoundedInMemoryExecutor.java @@ -48,6 +48,7 @@ public class SparkBoundedInMemoryExecutor extends BoundedInMemoryExecut this.sparkThreadTaskContext = TaskContext.get(); } + @Override public void preExecute() { // Passing parent thread's TaskContext to newly launched thread for it to access original TaskContext properties. TaskContext$.MODULE$.setTaskContext(sparkThreadTaskContext); diff --git a/hudi-client/src/main/java/org/apache/hudi/index/bloom/BloomIndexFileInfo.java b/hudi-client/src/main/java/org/apache/hudi/index/bloom/BloomIndexFileInfo.java index 51e1b7652..655d99bb7 100644 --- a/hudi-client/src/main/java/org/apache/hudi/index/bloom/BloomIndexFileInfo.java +++ b/hudi-client/src/main/java/org/apache/hudi/index/bloom/BloomIndexFileInfo.java @@ -88,6 +88,7 @@ public class BloomIndexFileInfo implements Serializable { return Objects.hashCode(fileId, minRecordKey, maxRecordKey); } + @Override public String toString() { final StringBuilder sb = new StringBuilder("BloomIndexFileInfo {"); sb.append(" fileId=").append(fileId); diff --git a/hudi-client/src/main/java/org/apache/hudi/index/hbase/HBaseIndex.java b/hudi-client/src/main/java/org/apache/hudi/index/hbase/HBaseIndex.java index dfe9dab96..677408b52 100644 --- a/hudi-client/src/main/java/org/apache/hudi/index/hbase/HBaseIndex.java +++ b/hudi-client/src/main/java/org/apache/hudi/index/hbase/HBaseIndex.java @@ -154,6 +154,7 @@ public class HBaseIndex extends HoodieIndex { */ private void addShutDownHook() { Runtime.getRuntime().addShutdownHook(new Thread() { + @Override public void run() { try { hbaseConnection.close(); @@ -167,6 +168,7 @@ public class HBaseIndex extends HoodieIndex { /** * Ensure that any resources used for indexing are released here. */ + @Override public void close() { this.hBaseIndexQPSResourceAllocator.releaseQPSResources(); } diff --git a/hudi-client/src/main/java/org/apache/hudi/io/HoodieCreateHandle.java b/hudi-client/src/main/java/org/apache/hudi/io/HoodieCreateHandle.java index 6916bad78..095e0a0e3 100644 --- a/hudi-client/src/main/java/org/apache/hudi/io/HoodieCreateHandle.java +++ b/hudi-client/src/main/java/org/apache/hudi/io/HoodieCreateHandle.java @@ -94,6 +94,7 @@ public class HoodieCreateHandle extends HoodieWri /** * Perform the actual writing of the given record into the backing file. */ + @Override public void write(HoodieRecord record, Option avroRecord) { Option recordMetadata = record.getData().getMetadata(); try { diff --git a/hudi-client/src/main/java/org/apache/hudi/io/HoodieMergeHandle.java b/hudi-client/src/main/java/org/apache/hudi/io/HoodieMergeHandle.java index 45519acca..518b88334 100644 --- a/hudi-client/src/main/java/org/apache/hudi/io/HoodieMergeHandle.java +++ b/hudi-client/src/main/java/org/apache/hudi/io/HoodieMergeHandle.java @@ -92,6 +92,7 @@ public class HoodieMergeHandle extends HoodieWrit return HoodieAvroUtils.addMetadataFields(originalSchema); } + @Override public Path makeNewPath(String partitionPath) { Path path = FSUtils.getPartitionPath(config.getBasePath(), partitionPath); try { @@ -103,6 +104,7 @@ public class HoodieMergeHandle extends HoodieWrit return new Path(path.toString(), FSUtils.makeDataFileName(instantTime, writeToken, fileId)); } + @Override public Schema getWriterSchema() { return writerSchema; } @@ -113,6 +115,7 @@ public class HoodieMergeHandle extends HoodieWrit * - Whether it belongs to the same partitionPath as existing records - Whether the current file written bytes lt max * file size */ + @Override public boolean canWrite(HoodieRecord record) { return false; } @@ -120,6 +123,7 @@ public class HoodieMergeHandle extends HoodieWrit /** * Perform the actual writing of the given record into the backing file. */ + @Override public void write(HoodieRecord record, Option insertValue) { // NO_OP } @@ -127,6 +131,7 @@ public class HoodieMergeHandle extends HoodieWrit /** * Perform the actual writing of the given record into the backing file. */ + @Override public void write(HoodieRecord record, Option avroRecord, Option exception) { Option recordMetadata = record.getData().getMetadata(); if (exception.isPresent() && exception.get() instanceof Throwable) { @@ -141,6 +146,7 @@ public class HoodieMergeHandle extends HoodieWrit /** * Rewrite the GenericRecord with the Schema containing the Hoodie Metadata fields. */ + @Override protected GenericRecord rewriteRecord(GenericRecord record) { return HoodieAvroUtils.rewriteRecord(record, writerSchema); } diff --git a/hudi-client/src/main/java/org/apache/hudi/io/storage/HoodieParquetWriter.java b/hudi-client/src/main/java/org/apache/hudi/io/storage/HoodieParquetWriter.java index fa7b16528..5f1fb8fc5 100644 --- a/hudi-client/src/main/java/org/apache/hudi/io/storage/HoodieParquetWriter.java +++ b/hudi-client/src/main/java/org/apache/hudi/io/storage/HoodieParquetWriter.java @@ -93,6 +93,7 @@ public class HoodieParquetWriter extends Hoodi * @param jsc JavaSparkContext * @return Cleaner Plan */ + @Override public HoodieCleanerPlan scheduleClean(JavaSparkContext jsc) { try { HoodieCleanHelper cleaner = new HoodieCleanHelper(this, config); diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFileReader.java b/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFileReader.java index 985ac99c1..354f8090d 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFileReader.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFileReader.java @@ -108,6 +108,7 @@ class HoodieLogFileReader implements HoodieLogFormat.Reader { */ private void addShutDownHook() { Runtime.getRuntime().addShutdownHook(new Thread() { + @Override public void run() { try { close(); diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFormatWriter.java b/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFormatWriter.java index c6adbc3ca..b8d5f8933 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFormatWriter.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/log/HoodieLogFormatWriter.java @@ -107,6 +107,7 @@ public class HoodieLogFormatWriter implements HoodieLogFormat.Writer { return fs; } + @Override public HoodieLogFile getLogFile() { return logFile; } @@ -212,6 +213,7 @@ public class HoodieLogFormatWriter implements HoodieLogFormat.Writer { output.hsync(); } + @Override public long getCurrentSize() throws IOException { if (output == null) { throw new IllegalStateException("Cannot get current size as the underlying stream has been closed already"); diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieActiveTimeline.java b/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieActiveTimeline.java index 1798a5348..cdf37646d 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieActiveTimeline.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/timeline/HoodieActiveTimeline.java @@ -145,6 +145,7 @@ public class HoodieActiveTimeline extends HoodieDefaultTimeline { * timeline * With Async compaction a requested/inflight compaction-instant is a valid baseInstant for a file-slice as * there could be delta-commits with that baseInstant. */ + @Override public HoodieTimeline getCommitsAndCompactionTimeline() { return getTimelineOfActions(Sets.newHashSet(COMMIT_ACTION, DELTA_COMMIT_ACTION, COMPACTION_ACTION)); } diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java b/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java index 9beeea0c5..bd7462d83 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java @@ -182,6 +182,7 @@ public abstract class AbstractTableFileSystemView implements SyncableFileSystemV /** * Clears the partition Map and reset view states. */ + @Override public final void reset() { try { writeLock.lock(); @@ -380,6 +381,7 @@ public abstract class AbstractTableFileSystemView implements SyncableFileSystemV /** * Get Latest data file for a partition and file-Id. */ + @Override public final Option getLatestDataFile(String partitionStr, String fileId) { try { readLock.lock(); @@ -434,6 +436,7 @@ public abstract class AbstractTableFileSystemView implements SyncableFileSystemV /** * Get Latest File Slice for a given fileId in a given partition. */ + @Override public final Option getLatestFileSlice(String partitionStr, String fileId) { try { readLock.lock(); diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/view/HoodieTableFileSystemView.java b/hudi-common/src/main/java/org/apache/hudi/common/table/view/HoodieTableFileSystemView.java index 34dadb0fe..dd711244a 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/view/HoodieTableFileSystemView.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/view/HoodieTableFileSystemView.java @@ -205,6 +205,7 @@ public class HoodieTableFileSystemView extends IncrementalTimelineSyncFileSystem }); } + @Override public void close() { closed = true; super.reset(); @@ -212,6 +213,7 @@ public class HoodieTableFileSystemView extends IncrementalTimelineSyncFileSystem fgIdToPendingCompaction = null; } + @Override public boolean isClosed() { return closed; } diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/view/SpillableMapBasedFileSystemView.java b/hudi-common/src/main/java/org/apache/hudi/common/table/view/SpillableMapBasedFileSystemView.java index 928e060bb..3ada17e95 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/table/view/SpillableMapBasedFileSystemView.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/table/view/SpillableMapBasedFileSystemView.java @@ -76,6 +76,7 @@ public class SpillableMapBasedFileSystemView extends HoodieTableFileSystemView { } } + @Override protected Map> createFileIdToPendingCompactionMap( Map> fgIdToPendingCompaction) { try { @@ -91,6 +92,7 @@ public class SpillableMapBasedFileSystemView extends HoodieTableFileSystemView { } } + @Override public Stream getAllFileGroups() { return ((ExternalSpillableMap) partitionToFileGroupsMap).valueStream() .flatMap(fg -> ((List) fg).stream()); diff --git a/hudi-common/src/main/java/org/apache/hudi/common/util/ObjectSizeCalculator.java b/hudi-common/src/main/java/org/apache/hudi/common/util/ObjectSizeCalculator.java index dc50c5671..b11ac6c1d 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/util/ObjectSizeCalculator.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/util/ObjectSizeCalculator.java @@ -124,6 +124,7 @@ public class ObjectSizeCalculator { private final LoadingCache, ClassSizeInfo> classSizeInfos = CacheBuilder.newBuilder().build(new CacheLoader, ClassSizeInfo>() { + @Override public ClassSizeInfo load(Class clazz) { return new ClassSizeInfo(clazz); } diff --git a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java index e26f1f239..e764a170d 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/DiskBasedMap.java @@ -125,6 +125,7 @@ public final class DiskBasedMap */ private void addShutDownHook() { Runtime.getRuntime().addShutdownHook(new Thread() { + @Override public void run() { try { if (writeOnlyFileHandle != null) { diff --git a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/LazyFileIterable.java b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/LazyFileIterable.java index aaaf32e8e..4b8e71684 100644 --- a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/LazyFileIterable.java +++ b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/LazyFileIterable.java @@ -115,6 +115,7 @@ public class LazyFileIterable implements Iterable { private void addShutdownHook() { Runtime.getRuntime().addShutdownHook(new Thread() { + @Override public void run() { close(); } diff --git a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieParquetInputFormat.java b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieParquetInputFormat.java index 5d6e52c41..e8f7de0f8 100644 --- a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieParquetInputFormat.java +++ b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/HoodieParquetInputFormat.java @@ -173,10 +173,12 @@ public class HoodieParquetInputFormat extends MapredParquetInputFormat implement return grouped; } + @Override public void setConf(Configuration conf) { this.conf = conf; } + @Override public Configuration getConf() { return conf; } diff --git a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/hive/HoodieCombineHiveInputFormat.java b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/hive/HoodieCombineHiveInputFormat.java index 1bf0da1e2..cd1cea32e 100644 --- a/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/hive/HoodieCombineHiveInputFormat.java +++ b/hudi-hadoop-mr/src/main/java/org/apache/hudi/hadoop/hive/HoodieCombineHiveInputFormat.java @@ -805,6 +805,7 @@ public class HoodieCombineHiveInputFormat listStatus(JobContext job) throws IOException { LOG.info("Listing status in HoodieCombineHiveInputFormat.HoodieCombineFileInputFormatShim"); List result; @@ -851,6 +854,7 @@ public class HoodieCombineHiveInputFormat> rrClass) throws IOException { return new HadoopShimsSecure.CombineFileRecordReader(job, split, reporter, rrClass); diff --git a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java index a8eab1da0..f8ddadbe5 100644 --- a/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java +++ b/hudi-utilities/src/main/java/org/apache/hudi/utilities/deltastreamer/HoodieDeltaStreamer.java @@ -554,6 +554,7 @@ public class HoodieDeltaStreamer implements Serializable { /** * Start Compaction Service. */ + @Override protected Pair startService() { ExecutorService executor = Executors.newFixedThreadPool(maxConcurrentCompaction); List> compactionFutures =