[MINOR] Fix typos in Spark client related classes (#4781)
This commit is contained in:
@@ -74,7 +74,7 @@ public class TimelineServerPerf implements Serializable {
|
||||
public TimelineServerPerf(Config cfg) throws IOException {
|
||||
this.cfg = cfg;
|
||||
useExternalTimelineServer = (cfg.serverHost != null);
|
||||
TimelineService.Config timelineServiceConf = cfg.getTimelinServerConfig();
|
||||
TimelineService.Config timelineServiceConf = cfg.getTimelineServerConfig();
|
||||
this.timelineServer = new TimelineService(
|
||||
new HoodieLocalEngineContext(FSUtils.prepareHadoopConf(new Configuration())),
|
||||
new Configuration(), timelineServiceConf, FileSystem.get(new Configuration()),
|
||||
@@ -281,7 +281,7 @@ public class TimelineServerPerf implements Serializable {
|
||||
description = " Server Host (Set it for externally managed timeline service")
|
||||
public String serverHost = null;
|
||||
|
||||
@Parameter(names = {"--view-storage", "-st"}, description = "View Storage Type. Defaut - SPILLABLE_DISK")
|
||||
@Parameter(names = {"--view-storage", "-st"}, description = "View Storage Type. Default - SPILLABLE_DISK")
|
||||
public FileSystemViewStorageType viewStorageType = FileSystemViewStorageType.SPILLABLE_DISK;
|
||||
|
||||
@Parameter(names = {"--max-view-mem-per-table", "-mv"},
|
||||
@@ -310,7 +310,7 @@ public class TimelineServerPerf implements Serializable {
|
||||
@Parameter(names = {"--help", "-h"})
|
||||
public Boolean help = false;
|
||||
|
||||
public TimelineService.Config getTimelinServerConfig() {
|
||||
public TimelineService.Config getTimelineServerConfig() {
|
||||
TimelineService.Config c = new TimelineService.Config();
|
||||
c.viewStorageType = viewStorageType;
|
||||
c.baseStorePathForFileGroups = baseStorePathForFileGroups;
|
||||
|
||||
@@ -279,35 +279,35 @@ public class HoodieDeltaStreamerTestBase extends UtilitiesTestBase {
|
||||
HoodieTestDataGenerator dataGenerator = new HoodieTestDataGenerator();
|
||||
if (useCustomSchema) {
|
||||
Helpers.saveORCToDFS(Helpers.toGenericRecords(
|
||||
dataGenerator.generateInsertsAsPerSchema("000", numRecords, schemaStr),
|
||||
schema), new Path(path), HoodieTestDataGenerator.ORC_TRIP_SCHEMA);
|
||||
dataGenerator.generateInsertsAsPerSchema("000", numRecords, schemaStr),
|
||||
schema), new Path(path), HoodieTestDataGenerator.ORC_TRIP_SCHEMA);
|
||||
} else {
|
||||
Helpers.saveORCToDFS(Helpers.toGenericRecords(
|
||||
dataGenerator.generateInserts("000", numRecords)), new Path(path));
|
||||
dataGenerator.generateInserts("000", numRecords)), new Path(path));
|
||||
}
|
||||
}
|
||||
|
||||
static void addCommitToTimeline(HoodieTableMetaClient metaCient) throws IOException {
|
||||
addCommitToTimeline(metaCient, Collections.emptyMap());
|
||||
static void addCommitToTimeline(HoodieTableMetaClient metaClient) throws IOException {
|
||||
addCommitToTimeline(metaClient, Collections.emptyMap());
|
||||
}
|
||||
|
||||
static void addCommitToTimeline(HoodieTableMetaClient metaCient, Map<String, String> extraMetadata) throws IOException {
|
||||
addCommitToTimeline(metaCient, WriteOperationType.UPSERT, HoodieTimeline.COMMIT_ACTION, extraMetadata);
|
||||
static void addCommitToTimeline(HoodieTableMetaClient metaClient, Map<String, String> extraMetadata) throws IOException {
|
||||
addCommitToTimeline(metaClient, WriteOperationType.UPSERT, HoodieTimeline.COMMIT_ACTION, extraMetadata);
|
||||
}
|
||||
|
||||
static void addReplaceCommitToTimeline(HoodieTableMetaClient metaCient, Map<String, String> extraMetadata) throws IOException {
|
||||
addCommitToTimeline(metaCient, WriteOperationType.CLUSTER, HoodieTimeline.REPLACE_COMMIT_ACTION, extraMetadata);
|
||||
static void addReplaceCommitToTimeline(HoodieTableMetaClient metaClient, Map<String, String> extraMetadata) throws IOException {
|
||||
addCommitToTimeline(metaClient, WriteOperationType.CLUSTER, HoodieTimeline.REPLACE_COMMIT_ACTION, extraMetadata);
|
||||
}
|
||||
|
||||
static void addCommitToTimeline(HoodieTableMetaClient metaCient, WriteOperationType writeOperationType, String commitActiontype,
|
||||
static void addCommitToTimeline(HoodieTableMetaClient metaClient, WriteOperationType writeOperationType, String commitActiontype,
|
||||
Map<String, String> extraMetadata) throws IOException {
|
||||
HoodieCommitMetadata commitMetadata = new HoodieCommitMetadata();
|
||||
commitMetadata.setOperationType(writeOperationType);
|
||||
extraMetadata.forEach((k,v) -> commitMetadata.getExtraMetadata().put(k, v));
|
||||
extraMetadata.forEach((k, v) -> commitMetadata.getExtraMetadata().put(k, v));
|
||||
String commitTime = HoodieActiveTimeline.createNewInstantTime();
|
||||
metaCient.getActiveTimeline().createNewInstant(new HoodieInstant(HoodieInstant.State.REQUESTED, commitActiontype, commitTime));
|
||||
metaCient.getActiveTimeline().createNewInstant(new HoodieInstant(HoodieInstant.State.INFLIGHT, commitActiontype, commitTime));
|
||||
metaCient.getActiveTimeline().saveAsComplete(
|
||||
metaClient.getActiveTimeline().createNewInstant(new HoodieInstant(HoodieInstant.State.REQUESTED, commitActiontype, commitTime));
|
||||
metaClient.getActiveTimeline().createNewInstant(new HoodieInstant(HoodieInstant.State.INFLIGHT, commitActiontype, commitTime));
|
||||
metaClient.getActiveTimeline().saveAsComplete(
|
||||
new HoodieInstant(HoodieInstant.State.INFLIGHT, commitActiontype, commitTime),
|
||||
Option.of(commitMetadata.toJsonString().getBytes(StandardCharsets.UTF_8)));
|
||||
}
|
||||
|
||||
@@ -367,22 +367,22 @@ public class TestHoodieDeltaStreamerWithMultiWriter extends SparkClientFunctiona
|
||||
return true;
|
||||
};
|
||||
|
||||
AtomicBoolean continousFailed = new AtomicBoolean(false);
|
||||
AtomicBoolean continuousFailed = new AtomicBoolean(false);
|
||||
AtomicBoolean backfillFailed = new AtomicBoolean(false);
|
||||
try {
|
||||
Future regularIngestionJobFuture = service.submit(() -> {
|
||||
try {
|
||||
deltaStreamerTestRunner(ingestionJob, cfgIngestionJob, conditionForRegularIngestion, jobId);
|
||||
} catch (Throwable ex) {
|
||||
continousFailed.set(true);
|
||||
continuousFailed.set(true);
|
||||
LOG.error("Continuous job failed " + ex.getMessage());
|
||||
throw new RuntimeException(ex);
|
||||
}
|
||||
});
|
||||
Future backfillJobFuture = service.submit(() -> {
|
||||
try {
|
||||
// trigger backfill atleast after 1 requested entry is added to timline from continuous job. If not, there is a chance that backfill will complete even before
|
||||
// continous job starts.
|
||||
// trigger backfill atleast after 1 requested entry is added to timeline from continuous job. If not, there is a chance that backfill will complete even before
|
||||
// continuous job starts.
|
||||
awaitCondition(new GetCommitsAfterInstant(tableBasePath, lastSuccessfulCommit));
|
||||
backfillJob.sync();
|
||||
} catch (Throwable ex) {
|
||||
|
||||
Reference in New Issue
Block a user