1
0

[MINOR] Fix wrong javadoc and refactor some naming issues (#2156)

This commit is contained in:
vinoyang
2020-10-10 06:09:26 +08:00
committed by GitHub
parent 00271af64e
commit eafd7bf289
6 changed files with 17 additions and 17 deletions

View File

@@ -268,7 +268,7 @@ public abstract class AbstractHoodieWriteClient<T extends HoodieRecordPayload, I
* *
* @param preppedRecords Prepared HoodieRecords to upsert * @param preppedRecords Prepared HoodieRecords to upsert
* @param instantTime Instant time of the commit * @param instantTime Instant time of the commit
* @return JavaRDD[WriteStatus] - RDD of WriteStatus to inspect errors and counts * @return Collection of WriteStatus to inspect errors and counts
*/ */
public abstract O upsertPreppedRecords(I preppedRecords, final String instantTime); public abstract O upsertPreppedRecords(I preppedRecords, final String instantTime);
@@ -280,7 +280,7 @@ public abstract class AbstractHoodieWriteClient<T extends HoodieRecordPayload, I
* *
* @param records HoodieRecords to insert * @param records HoodieRecords to insert
* @param instantTime Instant time of the commit * @param instantTime Instant time of the commit
* @return JavaRDD[WriteStatus] - RDD of WriteStatus to inspect errors and counts * @return Collection of WriteStatus to inspect errors and counts
*/ */
public abstract O insert(I records, final String instantTime); public abstract O insert(I records, final String instantTime);
@@ -293,7 +293,7 @@ public abstract class AbstractHoodieWriteClient<T extends HoodieRecordPayload, I
* *
* @param preppedRecords HoodieRecords to insert * @param preppedRecords HoodieRecords to insert
* @param instantTime Instant time of the commit * @param instantTime Instant time of the commit
* @return JavaRDD[WriteStatus] - RDD of WriteStatus to inspect errors and counts * @return Collection of WriteStatus to inspect errors and counts
*/ */
public abstract O insertPreppedRecords(I preppedRecords, final String instantTime); public abstract O insertPreppedRecords(I preppedRecords, final String instantTime);
@@ -306,7 +306,7 @@ public abstract class AbstractHoodieWriteClient<T extends HoodieRecordPayload, I
* *
* @param records HoodieRecords to insert * @param records HoodieRecords to insert
* @param instantTime Instant time of the commit * @param instantTime Instant time of the commit
* @return JavaRDD[WriteStatus] - RDD of WriteStatus to inspect errors and counts * @return Collection of WriteStatus to inspect errors and counts
*/ */
public abstract O bulkInsert(I records, final String instantTime); public abstract O bulkInsert(I records, final String instantTime);
@@ -323,7 +323,7 @@ public abstract class AbstractHoodieWriteClient<T extends HoodieRecordPayload, I
* @param instantTime Instant time of the commit * @param instantTime Instant time of the commit
* @param userDefinedBulkInsertPartitioner If specified then it will be used to partition input records before they are inserted * @param userDefinedBulkInsertPartitioner If specified then it will be used to partition input records before they are inserted
* into hoodie. * into hoodie.
* @return JavaRDD[WriteStatus] - RDD of WriteStatus to inspect errors and counts * @return Collection of WriteStatus to inspect errors and counts
*/ */
public abstract O bulkInsert(I records, final String instantTime, public abstract O bulkInsert(I records, final String instantTime,
Option<BulkInsertPartitioner<I>> userDefinedBulkInsertPartitioner); Option<BulkInsertPartitioner<I>> userDefinedBulkInsertPartitioner);
@@ -343,7 +343,7 @@ public abstract class AbstractHoodieWriteClient<T extends HoodieRecordPayload, I
* @param instantTime Instant time of the commit * @param instantTime Instant time of the commit
* @param bulkInsertPartitioner If specified then it will be used to partition input records before they are inserted * @param bulkInsertPartitioner If specified then it will be used to partition input records before they are inserted
* into hoodie. * into hoodie.
* @return JavaRDD[WriteStatus] - RDD of WriteStatus to inspect errors and counts * @return Collection of WriteStatus to inspect errors and counts
*/ */
public abstract O bulkInsertPreppedRecords(I preppedRecords, final String instantTime, public abstract O bulkInsertPreppedRecords(I preppedRecords, final String instantTime,
Option<BulkInsertPartitioner<I>> bulkInsertPartitioner); Option<BulkInsertPartitioner<I>> bulkInsertPartitioner);
@@ -354,7 +354,7 @@ public abstract class AbstractHoodieWriteClient<T extends HoodieRecordPayload, I
* *
* @param keys {@link List} of {@link HoodieKey}s to be deleted * @param keys {@link List} of {@link HoodieKey}s to be deleted
* @param instantTime Commit time handle * @param instantTime Commit time handle
* @return JavaRDD[WriteStatus] - RDD of WriteStatus to inspect errors and counts * @return Collection of WriteStatus to inspect errors and counts
*/ */
public abstract O delete(K keys, final String instantTime); public abstract O delete(K keys, final String instantTime);
@@ -653,7 +653,7 @@ public abstract class AbstractHoodieWriteClient<T extends HoodieRecordPayload, I
* Performs Compaction for the workload stored in instant-time. * Performs Compaction for the workload stored in instant-time.
* *
* @param compactionInstantTime Compaction Instant Time * @param compactionInstantTime Compaction Instant Time
* @return RDD of WriteStatus to inspect errors and counts * @return Collection of WriteStatus to inspect errors and counts
*/ */
public O compact(String compactionInstantTime) { public O compact(String compactionInstantTime) {
return compact(compactionInstantTime, config.shouldAutoCommit()); return compact(compactionInstantTime, config.shouldAutoCommit());
@@ -663,7 +663,7 @@ public abstract class AbstractHoodieWriteClient<T extends HoodieRecordPayload, I
* Commit a compaction operation. Allow passing additional meta-data to be stored in commit instant file. * Commit a compaction operation. Allow passing additional meta-data to be stored in commit instant file.
* *
* @param compactionInstantTime Compaction Instant Time * @param compactionInstantTime Compaction Instant Time
* @param writeStatuses RDD of WriteStatus to inspect errors and counts * @param writeStatuses Collection of WriteStatus to inspect errors and counts
* @param extraMetadata Extra Metadata to be stored * @param extraMetadata Extra Metadata to be stored
*/ */
public abstract void commitCompaction(String compactionInstantTime, O writeStatuses, public abstract void commitCompaction(String compactionInstantTime, O writeStatuses,
@@ -710,7 +710,7 @@ public abstract class AbstractHoodieWriteClient<T extends HoodieRecordPayload, I
* Ensures compaction instant is in expected state and performs Compaction for the workload stored in instant-time. * Ensures compaction instant is in expected state and performs Compaction for the workload stored in instant-time.
* *
* @param compactionInstantTime Compaction Instant Time * @param compactionInstantTime Compaction Instant Time
* @return RDD of Write Status * @return Collection of Write Status
*/ */
protected abstract O compact(String compactionInstantTime, boolean shouldComplete); protected abstract O compact(String compactionInstantTime, boolean shouldComplete);

View File

@@ -45,7 +45,7 @@ public abstract class FullRecordBootstrapDataProvider<I> implements Serializable
} }
/** /**
* Generates a list of input partition and files and returns a RDD representing source. * Generates a list of input partition and files and returns a collection representing source.
* @param tableName Hudi Table Name * @param tableName Hudi Table Name
* @param sourceBasePath Source Base Path * @param sourceBasePath Source Base Path
* @param partitionPaths Partition Paths * @param partitionPaths Partition Paths

View File

@@ -62,7 +62,7 @@ public abstract class HoodieIndex<T extends HoodieRecordPayload, I, K, O> implem
* TODO(vc): We may need to propagate the record as well in a WriteStatus class * TODO(vc): We may need to propagate the record as well in a WriteStatus class
*/ */
@PublicAPIMethod(maturity = ApiMaturityLevel.STABLE) @PublicAPIMethod(maturity = ApiMaturityLevel.STABLE)
public abstract O updateLocation(O writeStatusRDD, HoodieEngineContext context, public abstract O updateLocation(O writeStatuses, HoodieEngineContext context,
HoodieTable<T, I, K, O> hoodieTable) throws HoodieIndexException; HoodieTable<T, I, K, O> hoodieTable) throws HoodieIndexException;
/** /**

View File

@@ -73,7 +73,7 @@ public class HoodieIndexUtils {
public static HoodieRecord getTaggedRecord(HoodieRecord inputRecord, Option<HoodieRecordLocation> location) { public static HoodieRecord getTaggedRecord(HoodieRecord inputRecord, Option<HoodieRecordLocation> location) {
HoodieRecord record = inputRecord; HoodieRecord record = inputRecord;
if (location.isPresent()) { if (location.isPresent()) {
// When you have a record in multiple files in the same partition, then rowKeyRecordPairRDD // When you have a record in multiple files in the same partition, then <row key, record> collection
// will have 2 entries with the same exact in memory copy of the HoodieRecord and the 2 // will have 2 entries with the same exact in memory copy of the HoodieRecord and the 2
// separate filenames that the record is found in. This will result in setting // separate filenames that the record is found in. This will result in setting
// currentLocation 2 times and it will fail the second time. So creating a new in memory // currentLocation 2 times and it will fail the second time. So creating a new in memory

View File

@@ -35,7 +35,7 @@ public interface BulkInsertPartitioner<I> {
I repartitionRecords(I records, int outputSparkPartitions); I repartitionRecords(I records, int outputSparkPartitions);
/** /**
* @return {@code true} if the records within a RDD partition are sorted; {@code false} otherwise. * @return {@code true} if the records within a partition are sorted; {@code false} otherwise.
*/ */
boolean arePartitionRecordsSorted(); boolean arePartitionRecordsSorted();
} }

View File

@@ -32,7 +32,7 @@ import java.time.Instant;
public abstract class AbstractWriteHelper<T extends HoodieRecordPayload, I, K, O, R> { public abstract class AbstractWriteHelper<T extends HoodieRecordPayload, I, K, O, R> {
public HoodieWriteMetadata<O> write(String instantTime, public HoodieWriteMetadata<O> write(String instantTime,
I inputRecordsRDD, I inputRecords,
HoodieEngineContext context, HoodieEngineContext context,
HoodieTable<T, I, K, O> table, HoodieTable<T, I, K, O> table,
boolean shouldCombine, boolean shouldCombine,
@@ -42,7 +42,7 @@ public abstract class AbstractWriteHelper<T extends HoodieRecordPayload, I, K, O
try { try {
// De-dupe/merge if needed // De-dupe/merge if needed
I dedupedRecords = I dedupedRecords =
combineOnCondition(shouldCombine, inputRecordsRDD, shuffleParallelism, table); combineOnCondition(shouldCombine, inputRecords, shuffleParallelism, table);
Instant lookupBegin = Instant.now(); Instant lookupBegin = Instant.now();
I taggedRecords = dedupedRecords; I taggedRecords = dedupedRecords;
@@ -79,7 +79,7 @@ public abstract class AbstractWriteHelper<T extends HoodieRecordPayload, I, K, O
* *
* @param records hoodieRecords to deduplicate * @param records hoodieRecords to deduplicate
* @param parallelism parallelism or partitions to be used while reducing/deduplicating * @param parallelism parallelism or partitions to be used while reducing/deduplicating
* @return RDD of HoodieRecord already be deduplicated * @return Collection of HoodieRecord already be deduplicated
*/ */
public I deduplicateRecords( public I deduplicateRecords(
I records, HoodieTable<T, I, K, O> table, int parallelism) { I records, HoodieTable<T, I, K, O> table, int parallelism) {