1
0

[HUDI-379] Refactor the codes based on new JavadocStyle code style rule (#1079)

This commit is contained in:
lamber-ken
2019-12-06 12:59:28 +08:00
committed by leesf
parent c06d89b648
commit 2745b7552f
137 changed files with 434 additions and 433 deletions

View File

@@ -28,17 +28,17 @@ import java.io.IOException;
import java.io.Serializable;
/**
* Base class for all AVRO record based payloads, that can be ordered based on a field
* Base class for all AVRO record based payloads, that can be ordered based on a field.
*/
public abstract class BaseAvroPayload implements Serializable {
/**
* Avro data extracted from the source converted to bytes
* Avro data extracted from the source converted to bytes.
*/
protected final byte[] recordBytes;
/**
* For purposes of preCombining
* For purposes of preCombining.
*/
protected final Comparable orderingVal;

View File

@@ -49,7 +49,7 @@ import java.util.Map;
import java.util.stream.Collectors;
/**
* Utilities used throughout the data source
* Utilities used throughout the data source.
*/
public class DataSourceUtils {
@@ -120,7 +120,7 @@ public class DataSourceUtils {
}
/**
* Create a partition value extractor class via reflection, passing in any configs needed
* Create a partition value extractor class via reflection, passing in any configs needed.
*/
public static PartitionValueExtractor createPartitionExtractor(String partitionExtractorClass) {
try {

View File

@@ -31,7 +31,7 @@ import java.util.List;
import java.util.stream.Collectors;
/**
* List of helpers to aid, construction of instanttime for read and write operations using datasource
* List of helpers to aid, construction of instanttime for read and write operations using datasource.
*/
public class HoodieDataSourceHelpers {
@@ -53,7 +53,7 @@ public class HoodieDataSourceHelpers {
}
/**
* Returns the last successful write operation's instant time
* Returns the last successful write operation's instant time.
*/
public static String latestCommit(FileSystem fs, String basePath) {
HoodieTimeline timeline = allCompletedCommitsCompactions(fs, basePath);

View File

@@ -26,7 +26,7 @@ import org.apache.avro.generic.GenericRecord;
import java.io.Serializable;
/**
* Abstract class to extend for plugging in extraction of {@link HoodieKey} from an Avro record
* Abstract class to extend for plugging in extraction of {@link HoodieKey} from an Avro record.
*/
public abstract class KeyGenerator implements Serializable {

View File

@@ -25,7 +25,7 @@ import org.apache.hudi.exception.HoodieKeyException;
import org.apache.avro.generic.GenericRecord;
/**
* Simple Key generator for unpartitioned Hive Tables
* Simple Key generator for unpartitioned Hive Tables.
*/
public class NonpartitionedKeyGenerator extends SimpleKeyGenerator {

View File

@@ -45,7 +45,7 @@ import java.util.ArrayList;
import java.util.List;
/**
* Sample program that writes & reads hoodie datasets via the Spark datasource
* Sample program that writes & reads hoodie datasets via the Spark datasource.
*/
public class HoodieJavaApp {
@@ -232,7 +232,7 @@ public class HoodieJavaApp {
}
/**
* Setup configs for syncing to hive
* Setup configs for syncing to hive.
*/
private DataFrameWriter<Row> updateHiveSyncConfig(DataFrameWriter<Row> writer) {
if (enableHiveSync) {

View File

@@ -45,7 +45,7 @@ import java.util.concurrent.Executors;
import java.util.concurrent.Future;
/**
* Sample program that writes & reads hoodie datasets via the Spark datasource streaming
* Sample program that writes & reads hoodie datasets via the Spark datasource streaming.
*/
public class HoodieJavaStreamingApp {
@@ -165,7 +165,7 @@ public class HoodieJavaStreamingApp {
}
/**
* Adding data to the streaming source and showing results over time
* Adding data to the streaming source and showing results over time.
*
* @param spark
* @param fs
@@ -215,7 +215,7 @@ public class HoodieJavaStreamingApp {
}
/**
* Hoodie spark streaming job
* Hoodie spark streaming job.
*
* @param streamingInput
* @throws Exception
@@ -236,7 +236,7 @@ public class HoodieJavaStreamingApp {
}
/**
* Setup configs for syncing to hive
* Setup configs for syncing to hive.
*
* @param writer
* @return