[MINOR] Fix typos in Spark client related classes (#6204)
This commit is contained in:
@@ -136,7 +136,7 @@ public abstract class SingleSparkJobExecutionStrategy<T extends HoodieRecordPayl
|
|||||||
/**
|
/**
|
||||||
* Execute clustering to write inputRecords into new files as defined by rules in strategy parameters.
|
* Execute clustering to write inputRecords into new files as defined by rules in strategy parameters.
|
||||||
* The number of new file groups created is bounded by numOutputGroups.
|
* The number of new file groups created is bounded by numOutputGroups.
|
||||||
* Note that commit is not done as part of strategy. commit is callers responsibility.
|
* Note that commit is not done as part of strategy. Commit is callers responsibility.
|
||||||
*/
|
*/
|
||||||
public abstract Iterator<List<WriteStatus>> performClusteringWithRecordsIterator(final Iterator<HoodieRecord<T>> records, final int numOutputGroups,
|
public abstract Iterator<List<WriteStatus>> performClusteringWithRecordsIterator(final Iterator<HoodieRecord<T>> records, final int numOutputGroups,
|
||||||
final String instantTime,
|
final String instantTime,
|
||||||
|
|||||||
@@ -81,7 +81,7 @@ public class SparkInternalSchemaConverter {
|
|||||||
public static final String HOODIE_VALID_COMMITS_LIST = "hoodie.valid.commits.list";
|
public static final String HOODIE_VALID_COMMITS_LIST = "hoodie.valid.commits.list";
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Converts a spark schema to an hudi internal schema. Fields without IDs are kept and assigned fallback IDs.
|
* Convert a spark schema to an hudi internal schema. Fields without IDs are kept and assigned fallback IDs.
|
||||||
*
|
*
|
||||||
* @param sparkSchema a spark schema
|
* @param sparkSchema a spark schema
|
||||||
* @return a matching internal schema for the provided spark schema
|
* @return a matching internal schema for the provided spark schema
|
||||||
@@ -157,7 +157,7 @@ public class SparkInternalSchemaConverter {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Converts Spark schema to Hudi internal schema, and prune fields.
|
* Convert Spark schema to Hudi internal schema, and prune fields.
|
||||||
* Fields without IDs are kept and assigned fallback IDs.
|
* Fields without IDs are kept and assigned fallback IDs.
|
||||||
*
|
*
|
||||||
* @param sparkSchema a pruned spark schema
|
* @param sparkSchema a pruned spark schema
|
||||||
|
|||||||
@@ -50,7 +50,7 @@ import java.util.stream.Stream;
|
|||||||
import scala.collection.JavaConverters;
|
import scala.collection.JavaConverters;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Spark validator utils to verify and run any precommit validators configured.
|
* Spark validator utils to verify and run any pre-commit validators configured.
|
||||||
*/
|
*/
|
||||||
public class SparkValidatorUtils {
|
public class SparkValidatorUtils {
|
||||||
private static final Logger LOG = LogManager.getLogger(BaseSparkCommitActionExecutor.class);
|
private static final Logger LOG = LogManager.getLogger(BaseSparkCommitActionExecutor.class);
|
||||||
|
|||||||
@@ -308,7 +308,7 @@ public class HoodieAvroDataBlock extends HoodieDataBlock {
|
|||||||
ByteArrayOutputStream baos = new ByteArrayOutputStream();
|
ByteArrayOutputStream baos = new ByteArrayOutputStream();
|
||||||
DataOutputStream output = new DataOutputStream(baos);
|
DataOutputStream output = new DataOutputStream(baos);
|
||||||
|
|
||||||
// 2. Compress and Write schema out
|
// 1. Compress and Write schema out
|
||||||
byte[] schemaContent = compress(schema.toString());
|
byte[] schemaContent = compress(schema.toString());
|
||||||
output.writeInt(schemaContent.length);
|
output.writeInt(schemaContent.length);
|
||||||
output.write(schemaContent);
|
output.write(schemaContent);
|
||||||
@@ -318,10 +318,10 @@ public class HoodieAvroDataBlock extends HoodieDataBlock {
|
|||||||
recordItr.forEachRemaining(records::add);
|
recordItr.forEachRemaining(records::add);
|
||||||
}
|
}
|
||||||
|
|
||||||
// 3. Write total number of records
|
// 2. Write total number of records
|
||||||
output.writeInt(records.size());
|
output.writeInt(records.size());
|
||||||
|
|
||||||
// 4. Write the records
|
// 3. Write the records
|
||||||
Iterator<IndexedRecord> itr = records.iterator();
|
Iterator<IndexedRecord> itr = records.iterator();
|
||||||
while (itr.hasNext()) {
|
while (itr.hasNext()) {
|
||||||
IndexedRecord s = itr.next();
|
IndexedRecord s = itr.next();
|
||||||
|
|||||||
Reference in New Issue
Block a user