1
0

[HUDI-379] Refactor the codes based on new JavadocStyle code style rule (#1079)

This commit is contained in:
lamber-ken
2019-12-06 12:59:28 +08:00
committed by leesf
parent c06d89b648
commit 2745b7552f
137 changed files with 434 additions and 433 deletions

View File

@@ -21,7 +21,7 @@ package com.uber.hoodie.hadoop;
import org.apache.hudi.hadoop.HoodieParquetInputFormat;
/**
* Temporary class to allow seamless migration of com.uber.hoodie to org.apache.hudi
* Temporary class to allow seamless migration of com.uber.hoodie to org.apache.hudi .
*/
public class HoodieInputFormat extends HoodieParquetInputFormat {

View File

@@ -21,7 +21,7 @@ package com.uber.hoodie.hadoop.realtime;
import org.apache.hudi.hadoop.realtime.HoodieParquetRealtimeInputFormat;
/**
* Temporary class to allow seamless migration of com.uber.hoodie to org.apache.hudi
* Temporary class to allow seamless migration of com.uber.hoodie to org.apache.hudi .
*/
public class HoodieRealtimeInputFormat extends HoodieParquetRealtimeInputFormat {

View File

@@ -54,7 +54,7 @@ public class HoodieROTablePathFilter implements PathFilter, Serializable {
/**
* Its quite common, to have all files from a given partition path be passed into accept(), cache the check for hoodie
* metadata for known partition paths and the latest versions of files
* metadata for known partition paths and the latest versions of files.
*/
private HashMap<String, HashSet<Path>> hoodiePathCache;
@@ -72,7 +72,7 @@ public class HoodieROTablePathFilter implements PathFilter, Serializable {
}
/**
* Obtain the path, two levels from provided path
* Obtain the path, two levels from provided path.
*
* @return said path if available, null otherwise
*/

View File

@@ -29,7 +29,7 @@ import java.util.Iterator;
import java.util.NoSuchElementException;
/**
* Provides Iterator Interface to iterate value entries read from record reader
* Provides Iterator Interface to iterate value entries read from record reader.
*
* @param <K> Key Type
* @param <V> Value Type
@@ -42,7 +42,7 @@ public class RecordReaderValueIterator<K, V> implements Iterator<V> {
private V nextVal = null;
/**
* Construct RecordReaderValueIterator
* Construct RecordReaderValueIterator.
*
* @param reader reader
*/

View File

@@ -27,7 +27,7 @@ import java.lang.annotation.Target;
/**
* When annotated on a InputFormat, informs the query engines, that they should use the FileSplits provided by the input
* format to execute the queries
* format to execute the queries.
*/
@Inherited
@Documented

View File

@@ -486,7 +486,7 @@ public class HoodieCombineHiveInputFormat<K extends WritableComparable, V extend
}
/**
* Gets all the path indices that should not be combined
* Gets all the path indices that should not be combined.
*/
@VisibleForTesting
public Set<Integer> getNonCombinablePathIndices(JobConf job, Path[] paths, int numThreads)
@@ -602,8 +602,8 @@ public class HoodieCombineHiveInputFormat<K extends WritableComparable, V extend
}
/**
* MOD - Just added this for visibility
**/
* MOD - Just added this for visibility.
*/
Path[] getInputPaths(JobConf job) throws IOException {
Path[] dirs = FileInputFormat.getInputPaths(job);
if (dirs.length == 0) {
@@ -786,7 +786,7 @@ public class HoodieCombineHiveInputFormat<K extends WritableComparable, V extend
}
/**
* This is a marker interface that is used to identify the formats where combine split generation is not applicable
* This is a marker interface that is used to identify the formats where combine split generation is not applicable.
*/
public interface AvoidSplitCombination {
@@ -794,8 +794,8 @@ public class HoodieCombineHiveInputFormat<K extends WritableComparable, V extend
}
/**
* **MOD** this is the implementation of CombineFileInputFormat which is a copy of
* org.apache.hadoop.hive.shims.HadoopShimsSecure.CombineFileInputFormatShim with changes in listStatus
* **MOD** This is the implementation of CombineFileInputFormat which is a copy of
* org.apache.hadoop.hive.shims.HadoopShimsSecure.CombineFileInputFormatShim with changes in listStatus.
*/
public static class HoodieCombineFileInputFormatShim<K, V> extends CombineFileInputFormat<K, V>
implements org.apache.hadoop.hive.shims.HadoopShims.CombineFileInputFormatShim<K, V> {

View File

@@ -130,7 +130,7 @@ public abstract class AbstractRealtimeRecordReader {
}
/**
* Prints a JSON representation of the ArrayWritable for easier debuggability
* Prints a JSON representation of the ArrayWritable for easier debuggability.
*/
protected static String arrayWritableToString(ArrayWritable writable) {
if (writable == null) {
@@ -197,7 +197,7 @@ public abstract class AbstractRealtimeRecordReader {
}
/**
* Generate a reader schema off the provided writeSchema, to just project out the provided columns
* Generate a reader schema off the provided writeSchema, to just project out the provided columns.
*/
public static Schema generateProjectionSchema(Schema writeSchema, Map<String, Field> schemaFieldsMap,
List<String> fieldNames) {
@@ -234,7 +234,7 @@ public abstract class AbstractRealtimeRecordReader {
}
/**
* Convert the projected read from delta record into an array writable
* Convert the projected read from delta record into an array writable.
*/
public static Writable avroToArrayWritable(Object value, Schema schema) {

View File

@@ -60,7 +60,7 @@ import java.util.stream.Collectors;
import java.util.stream.Stream;
/**
* Input Format, that provides a real-time view of data in a Hoodie dataset
* Input Format, that provides a real-time view of data in a Hoodie dataset.
*/
@UseFileSplitsFromInputFormat
public class HoodieParquetRealtimeInputFormat extends HoodieParquetInputFormat implements Configurable {
@@ -160,7 +160,7 @@ public class HoodieParquetRealtimeInputFormat extends HoodieParquetInputFormat i
}
/**
* Add a field to the existing fields projected
* Add a field to the existing fields projected.
*/
private static Configuration addProjectionField(Configuration conf, String fieldName, int fieldIndex) {
String readColNames = conf.get(ColumnProjectionUtils.READ_COLUMN_NAMES_CONF_STR, "");

View File

@@ -21,7 +21,7 @@ package org.apache.hudi.hadoop.realtime;
import org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe;
/**
* Simply extends ParquetHiveSerDe
* Simply extends ParquetHiveSerDe.
*/
public class HoodieParquetSerde extends ParquetHiveSerDe {

View File

@@ -52,7 +52,7 @@ public class HoodieRealtimeRecordReader implements RecordReader<NullWritable, Ar
}
/**
* Construct record reader based on job configuration
* Construct record reader based on job configuration.
*
* @param split File Split
* @param jobConf Job Configuration

View File

@@ -60,7 +60,7 @@ class RealtimeUnmergedRecordReader extends AbstractRealtimeRecordReader
/**
* Construct a Unmerged record reader that parallely consumes both parquet and log records and buffers for upstream
* clients to consume
* clients to consume.
*
* @param split File split
* @param job Job Configuration