[HUDI-379] Refactor the codes based on new JavadocStyle code style rule (#1079)
This commit is contained in:
@@ -21,7 +21,7 @@ package com.uber.hoodie.hadoop;
|
||||
import org.apache.hudi.hadoop.HoodieParquetInputFormat;
|
||||
|
||||
/**
|
||||
* Temporary class to allow seamless migration of com.uber.hoodie to org.apache.hudi
|
||||
* Temporary class to allow seamless migration of com.uber.hoodie to org.apache.hudi .
|
||||
*/
|
||||
public class HoodieInputFormat extends HoodieParquetInputFormat {
|
||||
|
||||
|
||||
@@ -21,7 +21,7 @@ package com.uber.hoodie.hadoop.realtime;
|
||||
import org.apache.hudi.hadoop.realtime.HoodieParquetRealtimeInputFormat;
|
||||
|
||||
/**
|
||||
* Temporary class to allow seamless migration of com.uber.hoodie to org.apache.hudi
|
||||
* Temporary class to allow seamless migration of com.uber.hoodie to org.apache.hudi .
|
||||
*/
|
||||
public class HoodieRealtimeInputFormat extends HoodieParquetRealtimeInputFormat {
|
||||
|
||||
|
||||
@@ -54,7 +54,7 @@ public class HoodieROTablePathFilter implements PathFilter, Serializable {
|
||||
|
||||
/**
|
||||
* Its quite common, to have all files from a given partition path be passed into accept(), cache the check for hoodie
|
||||
* metadata for known partition paths and the latest versions of files
|
||||
* metadata for known partition paths and the latest versions of files.
|
||||
*/
|
||||
private HashMap<String, HashSet<Path>> hoodiePathCache;
|
||||
|
||||
@@ -72,7 +72,7 @@ public class HoodieROTablePathFilter implements PathFilter, Serializable {
|
||||
}
|
||||
|
||||
/**
|
||||
* Obtain the path, two levels from provided path
|
||||
* Obtain the path, two levels from provided path.
|
||||
*
|
||||
* @return said path if available, null otherwise
|
||||
*/
|
||||
|
||||
@@ -29,7 +29,7 @@ import java.util.Iterator;
|
||||
import java.util.NoSuchElementException;
|
||||
|
||||
/**
|
||||
* Provides Iterator Interface to iterate value entries read from record reader
|
||||
* Provides Iterator Interface to iterate value entries read from record reader.
|
||||
*
|
||||
* @param <K> Key Type
|
||||
* @param <V> Value Type
|
||||
@@ -42,7 +42,7 @@ public class RecordReaderValueIterator<K, V> implements Iterator<V> {
|
||||
private V nextVal = null;
|
||||
|
||||
/**
|
||||
* Construct RecordReaderValueIterator
|
||||
* Construct RecordReaderValueIterator.
|
||||
*
|
||||
* @param reader reader
|
||||
*/
|
||||
|
||||
@@ -27,7 +27,7 @@ import java.lang.annotation.Target;
|
||||
|
||||
/**
|
||||
* When annotated on a InputFormat, informs the query engines, that they should use the FileSplits provided by the input
|
||||
* format to execute the queries
|
||||
* format to execute the queries.
|
||||
*/
|
||||
@Inherited
|
||||
@Documented
|
||||
|
||||
@@ -486,7 +486,7 @@ public class HoodieCombineHiveInputFormat<K extends WritableComparable, V extend
|
||||
}
|
||||
|
||||
/**
|
||||
* Gets all the path indices that should not be combined
|
||||
* Gets all the path indices that should not be combined.
|
||||
*/
|
||||
@VisibleForTesting
|
||||
public Set<Integer> getNonCombinablePathIndices(JobConf job, Path[] paths, int numThreads)
|
||||
@@ -602,8 +602,8 @@ public class HoodieCombineHiveInputFormat<K extends WritableComparable, V extend
|
||||
}
|
||||
|
||||
/**
|
||||
* MOD - Just added this for visibility
|
||||
**/
|
||||
* MOD - Just added this for visibility.
|
||||
*/
|
||||
Path[] getInputPaths(JobConf job) throws IOException {
|
||||
Path[] dirs = FileInputFormat.getInputPaths(job);
|
||||
if (dirs.length == 0) {
|
||||
@@ -786,7 +786,7 @@ public class HoodieCombineHiveInputFormat<K extends WritableComparable, V extend
|
||||
}
|
||||
|
||||
/**
|
||||
* This is a marker interface that is used to identify the formats where combine split generation is not applicable
|
||||
* This is a marker interface that is used to identify the formats where combine split generation is not applicable.
|
||||
*/
|
||||
public interface AvoidSplitCombination {
|
||||
|
||||
@@ -794,8 +794,8 @@ public class HoodieCombineHiveInputFormat<K extends WritableComparable, V extend
|
||||
}
|
||||
|
||||
/**
|
||||
* **MOD** this is the implementation of CombineFileInputFormat which is a copy of
|
||||
* org.apache.hadoop.hive.shims.HadoopShimsSecure.CombineFileInputFormatShim with changes in listStatus
|
||||
* **MOD** This is the implementation of CombineFileInputFormat which is a copy of
|
||||
* org.apache.hadoop.hive.shims.HadoopShimsSecure.CombineFileInputFormatShim with changes in listStatus.
|
||||
*/
|
||||
public static class HoodieCombineFileInputFormatShim<K, V> extends CombineFileInputFormat<K, V>
|
||||
implements org.apache.hadoop.hive.shims.HadoopShims.CombineFileInputFormatShim<K, V> {
|
||||
|
||||
@@ -130,7 +130,7 @@ public abstract class AbstractRealtimeRecordReader {
|
||||
}
|
||||
|
||||
/**
|
||||
* Prints a JSON representation of the ArrayWritable for easier debuggability
|
||||
* Prints a JSON representation of the ArrayWritable for easier debuggability.
|
||||
*/
|
||||
protected static String arrayWritableToString(ArrayWritable writable) {
|
||||
if (writable == null) {
|
||||
@@ -197,7 +197,7 @@ public abstract class AbstractRealtimeRecordReader {
|
||||
}
|
||||
|
||||
/**
|
||||
* Generate a reader schema off the provided writeSchema, to just project out the provided columns
|
||||
* Generate a reader schema off the provided writeSchema, to just project out the provided columns.
|
||||
*/
|
||||
public static Schema generateProjectionSchema(Schema writeSchema, Map<String, Field> schemaFieldsMap,
|
||||
List<String> fieldNames) {
|
||||
@@ -234,7 +234,7 @@ public abstract class AbstractRealtimeRecordReader {
|
||||
}
|
||||
|
||||
/**
|
||||
* Convert the projected read from delta record into an array writable
|
||||
* Convert the projected read from delta record into an array writable.
|
||||
*/
|
||||
public static Writable avroToArrayWritable(Object value, Schema schema) {
|
||||
|
||||
|
||||
@@ -60,7 +60,7 @@ import java.util.stream.Collectors;
|
||||
import java.util.stream.Stream;
|
||||
|
||||
/**
|
||||
* Input Format, that provides a real-time view of data in a Hoodie dataset
|
||||
* Input Format, that provides a real-time view of data in a Hoodie dataset.
|
||||
*/
|
||||
@UseFileSplitsFromInputFormat
|
||||
public class HoodieParquetRealtimeInputFormat extends HoodieParquetInputFormat implements Configurable {
|
||||
@@ -160,7 +160,7 @@ public class HoodieParquetRealtimeInputFormat extends HoodieParquetInputFormat i
|
||||
}
|
||||
|
||||
/**
|
||||
* Add a field to the existing fields projected
|
||||
* Add a field to the existing fields projected.
|
||||
*/
|
||||
private static Configuration addProjectionField(Configuration conf, String fieldName, int fieldIndex) {
|
||||
String readColNames = conf.get(ColumnProjectionUtils.READ_COLUMN_NAMES_CONF_STR, "");
|
||||
|
||||
@@ -21,7 +21,7 @@ package org.apache.hudi.hadoop.realtime;
|
||||
import org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe;
|
||||
|
||||
/**
|
||||
* Simply extends ParquetHiveSerDe
|
||||
* Simply extends ParquetHiveSerDe.
|
||||
*/
|
||||
public class HoodieParquetSerde extends ParquetHiveSerDe {
|
||||
|
||||
|
||||
@@ -52,7 +52,7 @@ public class HoodieRealtimeRecordReader implements RecordReader<NullWritable, Ar
|
||||
}
|
||||
|
||||
/**
|
||||
* Construct record reader based on job configuration
|
||||
* Construct record reader based on job configuration.
|
||||
*
|
||||
* @param split File Split
|
||||
* @param jobConf Job Configuration
|
||||
|
||||
@@ -60,7 +60,7 @@ class RealtimeUnmergedRecordReader extends AbstractRealtimeRecordReader
|
||||
|
||||
/**
|
||||
* Construct a Unmerged record reader that parallely consumes both parquet and log records and buffers for upstream
|
||||
* clients to consume
|
||||
* clients to consume.
|
||||
*
|
||||
* @param split File split
|
||||
* @param job Job Configuration
|
||||
|
||||
Reference in New Issue
Block a user