[MINOR] Fix typo,rename 'getUrlEncodePartitoning' to 'getUrlEncodePartitioning' (#4130)
This commit is contained in:
@@ -466,7 +466,7 @@ public class HoodieTableConfig extends HoodieConfig {
|
||||
return getString(HIVE_STYLE_PARTITIONING_ENABLE);
|
||||
}
|
||||
|
||||
public String getUrlEncodePartitoning() {
|
||||
public String getUrlEncodePartitioning() {
|
||||
return getString(URL_ENCODE_PARTITIONING);
|
||||
}
|
||||
|
||||
|
||||
@@ -72,7 +72,7 @@ case class DeleteHoodieTableCommand(deleteTable: DeleteFromTable) extends Runnab
|
||||
"path" -> path,
|
||||
TBL_NAME.key -> tableConfig.getTableName,
|
||||
HIVE_STYLE_PARTITIONING.key -> tableConfig.getHiveStylePartitioningEnable,
|
||||
URL_ENCODE_PARTITIONING.key -> tableConfig.getUrlEncodePartitoning,
|
||||
URL_ENCODE_PARTITIONING.key -> tableConfig.getUrlEncodePartitioning,
|
||||
KEYGENERATOR_CLASS_NAME.key -> classOf[SqlKeyGenerator].getCanonicalName,
|
||||
SqlKeyGenerator.ORIGIN_KEYGEN_CLASS_NAME -> tableConfig.getKeyGeneratorClassName,
|
||||
OPERATION.key -> DataSourceWriteOptions.DELETE_OPERATION_OPT_VAL,
|
||||
|
||||
@@ -212,7 +212,7 @@ object InsertIntoHoodieTableCommand extends Logging {
|
||||
val partitionFields = hoodieCatalogTable.partitionFields.mkString(",")
|
||||
|
||||
val hiveStylePartitioningEnable = Option(tableConfig.getHiveStylePartitioningEnable).getOrElse("true")
|
||||
val urlEncodePartitioning = Option(tableConfig.getUrlEncodePartitoning).getOrElse("false")
|
||||
val urlEncodePartitioning = Option(tableConfig.getUrlEncodePartitioning).getOrElse("false")
|
||||
val keyGeneratorClassName = Option(tableConfig.getKeyGeneratorClassName)
|
||||
.getOrElse(classOf[ComplexKeyGenerator].getCanonicalName)
|
||||
|
||||
|
||||
@@ -449,7 +449,7 @@ case class MergeIntoHoodieTableCommand(mergeInto: MergeIntoTable) extends Runnab
|
||||
PARTITIONPATH_FIELD.key -> tableConfig.getPartitionFieldProp,
|
||||
PAYLOAD_CLASS_NAME.key -> classOf[ExpressionPayload].getCanonicalName,
|
||||
HIVE_STYLE_PARTITIONING.key -> tableConfig.getHiveStylePartitioningEnable,
|
||||
URL_ENCODE_PARTITIONING.key -> tableConfig.getUrlEncodePartitoning,
|
||||
URL_ENCODE_PARTITIONING.key -> tableConfig.getUrlEncodePartitioning,
|
||||
KEYGENERATOR_CLASS_NAME.key -> classOf[SqlKeyGenerator].getCanonicalName,
|
||||
SqlKeyGenerator.ORIGIN_KEYGEN_CLASS_NAME -> tableConfig.getKeyGeneratorClassName,
|
||||
META_SYNC_ENABLED.key -> enableHive.toString,
|
||||
|
||||
@@ -98,7 +98,7 @@ case class UpdateHoodieTableCommand(updateTable: UpdateTable) extends RunnableCo
|
||||
PRECOMBINE_FIELD.key -> preCombineColumn,
|
||||
TBL_NAME.key -> hoodieCatalogTable.tableName,
|
||||
HIVE_STYLE_PARTITIONING.key -> tableConfig.getHiveStylePartitioningEnable,
|
||||
URL_ENCODE_PARTITIONING.key -> tableConfig.getUrlEncodePartitoning,
|
||||
URL_ENCODE_PARTITIONING.key -> tableConfig.getUrlEncodePartitioning,
|
||||
KEYGENERATOR_CLASS_NAME.key -> classOf[SqlKeyGenerator].getCanonicalName,
|
||||
SqlKeyGenerator.ORIGIN_KEYGEN_CLASS_NAME -> tableConfig.getKeyGeneratorClassName,
|
||||
OPERATION.key -> UPSERT_OPERATION_OPT_VAL,
|
||||
|
||||
@@ -853,7 +853,7 @@ class TestHoodieSparkSqlWriter {
|
||||
.setConf(spark.sparkContext.hadoopConfiguration)
|
||||
.setBasePath(tablePath1).build().getTableConfig
|
||||
assert(tableConfig1.getHiveStylePartitioningEnable == "true")
|
||||
assert(tableConfig1.getUrlEncodePartitoning == "false")
|
||||
assert(tableConfig1.getUrlEncodePartitioning == "false")
|
||||
assert(tableConfig1.getKeyGeneratorClassName == classOf[ComplexKeyGenerator].getName)
|
||||
df.write.format("hudi")
|
||||
.options(options)
|
||||
@@ -874,7 +874,7 @@ class TestHoodieSparkSqlWriter {
|
||||
.setConf(spark.sparkContext.hadoopConfiguration)
|
||||
.setBasePath(tablePath2).build().getTableConfig
|
||||
assert(tableConfig2.getHiveStylePartitioningEnable == "false")
|
||||
assert(tableConfig2.getUrlEncodePartitoning == "true")
|
||||
assert(tableConfig2.getUrlEncodePartitioning == "true")
|
||||
assert(tableConfig2.getKeyGeneratorClassName == classOf[SimpleKeyGenerator].getName)
|
||||
|
||||
val df2 = Seq((2, "a2", 20, 1000, "2021-10-16")).toDF("id", "name", "value", "ts", "dt")
|
||||
|
||||
Reference in New Issue
Block a user