[HUDI-65] commitTime rename to instantTime (#1431)
This commit is contained in:
@@ -196,20 +196,20 @@ public class DataSourceUtils {
|
||||
}
|
||||
|
||||
public static JavaRDD<WriteStatus> doWriteOperation(HoodieWriteClient client, JavaRDD<HoodieRecord> hoodieRecords,
|
||||
String commitTime, String operation) {
|
||||
String instantTime, String operation) {
|
||||
if (operation.equals(DataSourceWriteOptions.BULK_INSERT_OPERATION_OPT_VAL())) {
|
||||
return client.bulkInsert(hoodieRecords, commitTime);
|
||||
return client.bulkInsert(hoodieRecords, instantTime);
|
||||
} else if (operation.equals(DataSourceWriteOptions.INSERT_OPERATION_OPT_VAL())) {
|
||||
return client.insert(hoodieRecords, commitTime);
|
||||
return client.insert(hoodieRecords, instantTime);
|
||||
} else {
|
||||
// default is upsert
|
||||
return client.upsert(hoodieRecords, commitTime);
|
||||
return client.upsert(hoodieRecords, instantTime);
|
||||
}
|
||||
}
|
||||
|
||||
public static JavaRDD<WriteStatus> doDeleteOperation(HoodieWriteClient client, JavaRDD<HoodieKey> hoodieKeys,
|
||||
String commitTime) {
|
||||
return client.delete(hoodieKeys, commitTime);
|
||||
String instantTime) {
|
||||
return client.delete(hoodieKeys, instantTime);
|
||||
}
|
||||
|
||||
public static HoodieRecord createHoodieRecord(GenericRecord gr, Comparable orderingVal, HoodieKey hKey,
|
||||
|
||||
@@ -78,7 +78,7 @@ private[hudi] object HoodieSparkSqlWriter {
|
||||
|
||||
val jsc = new JavaSparkContext(sparkContext)
|
||||
val basePath = new Path(parameters("path"))
|
||||
val commitTime = HoodieActiveTimeline.createNewInstantTime()
|
||||
val instantTime = HoodieActiveTimeline.createNewInstantTime()
|
||||
val fs = basePath.getFileSystem(sparkContext.hadoopConfiguration)
|
||||
var exists = fs.exists(new Path(basePath, HoodieTableMetaClient.METAFOLDER_NAME))
|
||||
|
||||
@@ -140,8 +140,8 @@ private[hudi] object HoodieSparkSqlWriter {
|
||||
log.info("new batch has no new records, skipping...")
|
||||
(true, common.util.Option.empty())
|
||||
}
|
||||
client.startCommitWithTime(commitTime)
|
||||
val writeStatuses = DataSourceUtils.doWriteOperation(client, hoodieRecords, commitTime, operation)
|
||||
client.startCommitWithTime(instantTime)
|
||||
val writeStatuses = DataSourceUtils.doWriteOperation(client, hoodieRecords, instantTime, operation)
|
||||
(writeStatuses, client)
|
||||
} else {
|
||||
|
||||
@@ -172,14 +172,14 @@ private[hudi] object HoodieSparkSqlWriter {
|
||||
)
|
||||
|
||||
// Issue deletes
|
||||
client.startCommitWithTime(commitTime)
|
||||
val writeStatuses = DataSourceUtils.doDeleteOperation(client, hoodieKeysToDelete, commitTime)
|
||||
client.startCommitWithTime(instantTime)
|
||||
val writeStatuses = DataSourceUtils.doDeleteOperation(client, hoodieKeysToDelete, instantTime)
|
||||
(writeStatuses, client)
|
||||
}
|
||||
|
||||
// Check for errors and commit the write.
|
||||
val writeSuccessful = checkWriteStatus(writeStatuses, parameters, writeClient, commitTime, basePath, operation, jsc)
|
||||
(writeSuccessful, common.util.Option.ofNullable(commitTime))
|
||||
val writeSuccessful = checkWriteStatus(writeStatuses, parameters, writeClient, instantTime, basePath, operation, jsc)
|
||||
(writeSuccessful, common.util.Option.ofNullable(instantTime))
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -246,7 +246,7 @@ private[hudi] object HoodieSparkSqlWriter {
|
||||
private def checkWriteStatus(writeStatuses: JavaRDD[WriteStatus],
|
||||
parameters: Map[String, String],
|
||||
client: HoodieWriteClient[_],
|
||||
commitTime: String,
|
||||
instantTime: String,
|
||||
basePath: Path,
|
||||
operation: String,
|
||||
jsc: JavaSparkContext): Boolean = {
|
||||
@@ -256,17 +256,17 @@ private[hudi] object HoodieSparkSqlWriter {
|
||||
val metaMap = parameters.filter(kv =>
|
||||
kv._1.startsWith(parameters(COMMIT_METADATA_KEYPREFIX_OPT_KEY)))
|
||||
val commitSuccess = if (metaMap.isEmpty) {
|
||||
client.commit(commitTime, writeStatuses)
|
||||
client.commit(instantTime, writeStatuses)
|
||||
} else {
|
||||
client.commit(commitTime, writeStatuses,
|
||||
client.commit(instantTime, writeStatuses,
|
||||
common.util.Option.of(new util.HashMap[String, String](mapAsJavaMap(metaMap))))
|
||||
}
|
||||
|
||||
if (commitSuccess) {
|
||||
log.info("Commit " + commitTime + " successful!")
|
||||
log.info("Commit " + instantTime + " successful!")
|
||||
}
|
||||
else {
|
||||
log.info("Commit " + commitTime + " failed!")
|
||||
log.info("Commit " + instantTime + " failed!")
|
||||
}
|
||||
|
||||
val hiveSyncEnabled = parameters.get(HIVE_SYNC_ENABLED_OPT_KEY).exists(r => r.toBoolean)
|
||||
|
||||
Reference in New Issue
Block a user