[MINOR] Fix typos in Spark client related classes (#4781)
This commit is contained in:
@@ -279,35 +279,35 @@ public class HoodieDeltaStreamerTestBase extends UtilitiesTestBase {
|
||||
HoodieTestDataGenerator dataGenerator = new HoodieTestDataGenerator();
|
||||
if (useCustomSchema) {
|
||||
Helpers.saveORCToDFS(Helpers.toGenericRecords(
|
||||
dataGenerator.generateInsertsAsPerSchema("000", numRecords, schemaStr),
|
||||
schema), new Path(path), HoodieTestDataGenerator.ORC_TRIP_SCHEMA);
|
||||
dataGenerator.generateInsertsAsPerSchema("000", numRecords, schemaStr),
|
||||
schema), new Path(path), HoodieTestDataGenerator.ORC_TRIP_SCHEMA);
|
||||
} else {
|
||||
Helpers.saveORCToDFS(Helpers.toGenericRecords(
|
||||
dataGenerator.generateInserts("000", numRecords)), new Path(path));
|
||||
dataGenerator.generateInserts("000", numRecords)), new Path(path));
|
||||
}
|
||||
}
|
||||
|
||||
static void addCommitToTimeline(HoodieTableMetaClient metaCient) throws IOException {
|
||||
addCommitToTimeline(metaCient, Collections.emptyMap());
|
||||
static void addCommitToTimeline(HoodieTableMetaClient metaClient) throws IOException {
|
||||
addCommitToTimeline(metaClient, Collections.emptyMap());
|
||||
}
|
||||
|
||||
static void addCommitToTimeline(HoodieTableMetaClient metaCient, Map<String, String> extraMetadata) throws IOException {
|
||||
addCommitToTimeline(metaCient, WriteOperationType.UPSERT, HoodieTimeline.COMMIT_ACTION, extraMetadata);
|
||||
static void addCommitToTimeline(HoodieTableMetaClient metaClient, Map<String, String> extraMetadata) throws IOException {
|
||||
addCommitToTimeline(metaClient, WriteOperationType.UPSERT, HoodieTimeline.COMMIT_ACTION, extraMetadata);
|
||||
}
|
||||
|
||||
static void addReplaceCommitToTimeline(HoodieTableMetaClient metaCient, Map<String, String> extraMetadata) throws IOException {
|
||||
addCommitToTimeline(metaCient, WriteOperationType.CLUSTER, HoodieTimeline.REPLACE_COMMIT_ACTION, extraMetadata);
|
||||
static void addReplaceCommitToTimeline(HoodieTableMetaClient metaClient, Map<String, String> extraMetadata) throws IOException {
|
||||
addCommitToTimeline(metaClient, WriteOperationType.CLUSTER, HoodieTimeline.REPLACE_COMMIT_ACTION, extraMetadata);
|
||||
}
|
||||
|
||||
static void addCommitToTimeline(HoodieTableMetaClient metaCient, WriteOperationType writeOperationType, String commitActiontype,
|
||||
static void addCommitToTimeline(HoodieTableMetaClient metaClient, WriteOperationType writeOperationType, String commitActiontype,
|
||||
Map<String, String> extraMetadata) throws IOException {
|
||||
HoodieCommitMetadata commitMetadata = new HoodieCommitMetadata();
|
||||
commitMetadata.setOperationType(writeOperationType);
|
||||
extraMetadata.forEach((k,v) -> commitMetadata.getExtraMetadata().put(k, v));
|
||||
extraMetadata.forEach((k, v) -> commitMetadata.getExtraMetadata().put(k, v));
|
||||
String commitTime = HoodieActiveTimeline.createNewInstantTime();
|
||||
metaCient.getActiveTimeline().createNewInstant(new HoodieInstant(HoodieInstant.State.REQUESTED, commitActiontype, commitTime));
|
||||
metaCient.getActiveTimeline().createNewInstant(new HoodieInstant(HoodieInstant.State.INFLIGHT, commitActiontype, commitTime));
|
||||
metaCient.getActiveTimeline().saveAsComplete(
|
||||
metaClient.getActiveTimeline().createNewInstant(new HoodieInstant(HoodieInstant.State.REQUESTED, commitActiontype, commitTime));
|
||||
metaClient.getActiveTimeline().createNewInstant(new HoodieInstant(HoodieInstant.State.INFLIGHT, commitActiontype, commitTime));
|
||||
metaClient.getActiveTimeline().saveAsComplete(
|
||||
new HoodieInstant(HoodieInstant.State.INFLIGHT, commitActiontype, commitTime),
|
||||
Option.of(commitMetadata.toJsonString().getBytes(StandardCharsets.UTF_8)));
|
||||
}
|
||||
|
||||
@@ -367,22 +367,22 @@ public class TestHoodieDeltaStreamerWithMultiWriter extends SparkClientFunctiona
|
||||
return true;
|
||||
};
|
||||
|
||||
AtomicBoolean continousFailed = new AtomicBoolean(false);
|
||||
AtomicBoolean continuousFailed = new AtomicBoolean(false);
|
||||
AtomicBoolean backfillFailed = new AtomicBoolean(false);
|
||||
try {
|
||||
Future regularIngestionJobFuture = service.submit(() -> {
|
||||
try {
|
||||
deltaStreamerTestRunner(ingestionJob, cfgIngestionJob, conditionForRegularIngestion, jobId);
|
||||
} catch (Throwable ex) {
|
||||
continousFailed.set(true);
|
||||
continuousFailed.set(true);
|
||||
LOG.error("Continuous job failed " + ex.getMessage());
|
||||
throw new RuntimeException(ex);
|
||||
}
|
||||
});
|
||||
Future backfillJobFuture = service.submit(() -> {
|
||||
try {
|
||||
// trigger backfill atleast after 1 requested entry is added to timline from continuous job. If not, there is a chance that backfill will complete even before
|
||||
// continous job starts.
|
||||
// trigger backfill atleast after 1 requested entry is added to timeline from continuous job. If not, there is a chance that backfill will complete even before
|
||||
// continuous job starts.
|
||||
awaitCondition(new GetCommitsAfterInstant(tableBasePath, lastSuccessfulCommit));
|
||||
backfillJob.sync();
|
||||
} catch (Throwable ex) {
|
||||
|
||||
Reference in New Issue
Block a user