[HUDI-2255] Refactor Datasource options (#3373)
Co-authored-by: Wenning Ding <wenningd@amazon.com>
This commit is contained in:
@@ -18,8 +18,8 @@
|
||||
|
||||
package org.apache.hudi.examples.spark
|
||||
|
||||
import org.apache.hudi.DataSourceReadOptions.{BEGIN_INSTANTTIME_OPT_KEY, END_INSTANTTIME_OPT_KEY, QUERY_TYPE_INCREMENTAL_OPT_VAL, QUERY_TYPE_OPT_KEY}
|
||||
import org.apache.hudi.DataSourceWriteOptions.{PARTITIONPATH_FIELD_OPT_KEY, PRECOMBINE_FIELD_OPT_KEY, RECORDKEY_FIELD_OPT_KEY}
|
||||
import org.apache.hudi.DataSourceReadOptions.{BEGIN_INSTANTTIME, END_INSTANTTIME, QUERY_TYPE_INCREMENTAL_OPT_VAL, QUERY_TYPE}
|
||||
import org.apache.hudi.DataSourceWriteOptions.{PARTITIONPATH_FIELD, PRECOMBINE_FIELD, RECORDKEY_FIELD}
|
||||
import org.apache.hudi.QuickstartUtils.getQuickstartWriteConfigs
|
||||
import org.apache.hudi.common.model.HoodieAvroPayload
|
||||
import org.apache.hudi.config.HoodieWriteConfig.TABLE_NAME
|
||||
@@ -76,9 +76,9 @@ object HoodieDataSourceExample {
|
||||
val df = spark.read.json(spark.sparkContext.parallelize(inserts, 1))
|
||||
df.write.format("org.apache.hudi").
|
||||
options(getQuickstartWriteConfigs).
|
||||
option(PRECOMBINE_FIELD_OPT_KEY.key, "ts").
|
||||
option(RECORDKEY_FIELD_OPT_KEY.key, "uuid").
|
||||
option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionpath").
|
||||
option(PRECOMBINE_FIELD.key, "ts").
|
||||
option(RECORDKEY_FIELD.key, "uuid").
|
||||
option(PARTITIONPATH_FIELD.key, "partitionpath").
|
||||
option(TABLE_NAME.key, tableName).
|
||||
mode(Overwrite).
|
||||
save(tablePath)
|
||||
@@ -121,9 +121,9 @@ object HoodieDataSourceExample {
|
||||
val df = spark.read.json(spark.sparkContext.parallelize(updates, 1))
|
||||
df.write.format("org.apache.hudi").
|
||||
options(getQuickstartWriteConfigs).
|
||||
option(PRECOMBINE_FIELD_OPT_KEY.key, "ts").
|
||||
option(RECORDKEY_FIELD_OPT_KEY.key, "uuid").
|
||||
option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionpath").
|
||||
option(PRECOMBINE_FIELD.key, "ts").
|
||||
option(RECORDKEY_FIELD.key, "uuid").
|
||||
option(PARTITIONPATH_FIELD.key, "partitionpath").
|
||||
option(TABLE_NAME.key, tableName).
|
||||
mode(Append).
|
||||
save(tablePath)
|
||||
@@ -143,8 +143,8 @@ object HoodieDataSourceExample {
|
||||
val incViewDF = spark.
|
||||
read.
|
||||
format("org.apache.hudi").
|
||||
option(QUERY_TYPE_OPT_KEY.key, QUERY_TYPE_INCREMENTAL_OPT_VAL).
|
||||
option(BEGIN_INSTANTTIME_OPT_KEY.key, beginTime).
|
||||
option(QUERY_TYPE.key, QUERY_TYPE_INCREMENTAL_OPT_VAL).
|
||||
option(BEGIN_INSTANTTIME.key, beginTime).
|
||||
load(tablePath)
|
||||
incViewDF.createOrReplaceTempView("hudi_incr_table")
|
||||
spark.sql("select `_hoodie_commit_time`, fare, begin_lon, begin_lat, ts from hudi_incr_table where fare > 20.0").show()
|
||||
@@ -163,9 +163,9 @@ object HoodieDataSourceExample {
|
||||
|
||||
//incrementally query data
|
||||
val incViewDF = spark.read.format("org.apache.hudi").
|
||||
option(QUERY_TYPE_OPT_KEY.key, QUERY_TYPE_INCREMENTAL_OPT_VAL).
|
||||
option(BEGIN_INSTANTTIME_OPT_KEY.key, beginTime).
|
||||
option(END_INSTANTTIME_OPT_KEY.key, endTime).
|
||||
option(QUERY_TYPE.key, QUERY_TYPE_INCREMENTAL_OPT_VAL).
|
||||
option(BEGIN_INSTANTTIME.key, beginTime).
|
||||
option(END_INSTANTTIME.key, endTime).
|
||||
load(tablePath)
|
||||
incViewDF.createOrReplaceTempView("hudi_incr_table")
|
||||
spark.sql("select `_hoodie_commit_time`, fare, begin_lon, begin_lat, ts from hudi_incr_table where fare > 20.0").show()
|
||||
|
||||
@@ -19,7 +19,7 @@
|
||||
|
||||
package org.apache.hudi.examples.spark
|
||||
|
||||
import org.apache.hudi.DataSourceWriteOptions.{PARTITIONPATH_FIELD_OPT_KEY, PRECOMBINE_FIELD_OPT_KEY, RECORDKEY_FIELD_OPT_KEY, TABLE_TYPE_OPT_KEY}
|
||||
import org.apache.hudi.DataSourceWriteOptions.{PARTITIONPATH_FIELD, PRECOMBINE_FIELD, RECORDKEY_FIELD, TABLE_TYPE}
|
||||
import org.apache.hudi.QuickstartUtils.getQuickstartWriteConfigs
|
||||
import org.apache.hudi.client.SparkRDDWriteClient
|
||||
import org.apache.hudi.client.common.HoodieSparkEngineContext
|
||||
@@ -86,11 +86,11 @@ object HoodieMorCompactionJob {
|
||||
val df = spark.read.json(spark.sparkContext.parallelize(inserts.asScala, 1))
|
||||
df.write.format("org.apache.hudi").
|
||||
options(getQuickstartWriteConfigs).
|
||||
option(PRECOMBINE_FIELD_OPT_KEY.key, "ts").
|
||||
option(RECORDKEY_FIELD_OPT_KEY.key, "uuid").
|
||||
option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionpath").
|
||||
option(PRECOMBINE_FIELD.key, "ts").
|
||||
option(RECORDKEY_FIELD.key, "uuid").
|
||||
option(PARTITIONPATH_FIELD.key, "partitionpath").
|
||||
option(TABLE_NAME.key, tableName).
|
||||
option(TABLE_TYPE_OPT_KEY.key, tableType).
|
||||
option(TABLE_TYPE.key, tableType).
|
||||
mode(Overwrite).
|
||||
save(tablePath)
|
||||
}
|
||||
@@ -102,11 +102,11 @@ object HoodieMorCompactionJob {
|
||||
val df = spark.read.json(spark.sparkContext.parallelize(updates.asScala, 1))
|
||||
df.write.format("org.apache.hudi").
|
||||
options(getQuickstartWriteConfigs).
|
||||
option(PRECOMBINE_FIELD_OPT_KEY.key, "ts").
|
||||
option(RECORDKEY_FIELD_OPT_KEY.key, "uuid").
|
||||
option(PARTITIONPATH_FIELD_OPT_KEY.key, "partitionpath").
|
||||
option(PRECOMBINE_FIELD.key, "ts").
|
||||
option(RECORDKEY_FIELD.key, "uuid").
|
||||
option(PARTITIONPATH_FIELD.key, "partitionpath").
|
||||
option(TABLE_NAME.key, tableName).
|
||||
option(TABLE_TYPE_OPT_KEY.key, tableType).
|
||||
option(TABLE_TYPE.key, tableType).
|
||||
mode(Append).
|
||||
save(tablePath)
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user