[HUDI-89] Add configOption & refactor all configs based on that (#2833)
Co-authored-by: Wenning Ding <wenningd@amazon.com>
This commit is contained in:
@@ -173,7 +173,7 @@ public class FlinkOptions {
|
||||
// Write Options
|
||||
// ------------------------------------------------------------------------
|
||||
public static final ConfigOption<String> TABLE_NAME = ConfigOptions
|
||||
.key(HoodieWriteConfig.TABLE_NAME)
|
||||
.key(HoodieWriteConfig.TABLE_NAME.key())
|
||||
.stringType()
|
||||
.noDefaultValue()
|
||||
.withDescription("Table name to register to Hive metastore");
|
||||
@@ -240,7 +240,7 @@ public class FlinkOptions {
|
||||
+ "By default true (in favor of streaming progressing over data integrity)");
|
||||
|
||||
public static final ConfigOption<String> RECORD_KEY_FIELD = ConfigOptions
|
||||
.key(KeyGeneratorOptions.RECORDKEY_FIELD_OPT_KEY)
|
||||
.key(KeyGeneratorOptions.RECORDKEY_FIELD_OPT_KEY.key())
|
||||
.stringType()
|
||||
.defaultValue("uuid")
|
||||
.withDescription("Record key field. Value to be used as the `recordKey` component of `HoodieKey`.\n"
|
||||
@@ -248,7 +248,7 @@ public class FlinkOptions {
|
||||
+ "the dot notation eg: `a.b.c`");
|
||||
|
||||
public static final ConfigOption<String> PARTITION_PATH_FIELD = ConfigOptions
|
||||
.key(KeyGeneratorOptions.PARTITIONPATH_FIELD_OPT_KEY)
|
||||
.key(KeyGeneratorOptions.PARTITIONPATH_FIELD_OPT_KEY.key())
|
||||
.stringType()
|
||||
.defaultValue("")
|
||||
.withDescription("Partition path field. Value to be used at the `partitionPath` component of `HoodieKey`.\n"
|
||||
@@ -269,13 +269,13 @@ public class FlinkOptions {
|
||||
+ "By default false (the names of partition folders are only partition values)");
|
||||
|
||||
public static final ConfigOption<String> KEYGEN_CLASS = ConfigOptions
|
||||
.key(HoodieWriteConfig.KEYGENERATOR_CLASS_PROP)
|
||||
.key(HoodieWriteConfig.KEYGENERATOR_CLASS_PROP.key())
|
||||
.stringType()
|
||||
.defaultValue("")
|
||||
.withDescription("Key generator class, that implements will extract the key out of incoming record");
|
||||
|
||||
public static final ConfigOption<String> KEYGEN_TYPE = ConfigOptions
|
||||
.key(HoodieWriteConfig.KEYGENERATOR_TYPE_PROP)
|
||||
.key(HoodieWriteConfig.KEYGENERATOR_TYPE_PROP.key())
|
||||
.stringType()
|
||||
.defaultValue(KeyGeneratorType.SIMPLE.name())
|
||||
.withDescription("Key generator type, that implements will extract the key out of incoming record");
|
||||
|
||||
@@ -56,7 +56,7 @@ import java.util.List;
|
||||
import java.util.Locale;
|
||||
import java.util.Properties;
|
||||
|
||||
import static org.apache.hudi.common.table.HoodieTableConfig.DEFAULT_ARCHIVELOG_FOLDER;
|
||||
import static org.apache.hudi.common.table.HoodieTableConfig.HOODIE_ARCHIVELOG_FOLDER_PROP;
|
||||
|
||||
/**
|
||||
* Utilities for Flink stream read and write.
|
||||
@@ -209,7 +209,7 @@ public class StreamerUtil {
|
||||
.setTableType(conf.getString(FlinkOptions.TABLE_TYPE))
|
||||
.setTableName(conf.getString(FlinkOptions.TABLE_NAME))
|
||||
.setPayloadClassName(conf.getString(FlinkOptions.PAYLOAD_CLASS))
|
||||
.setArchiveLogFolder(DEFAULT_ARCHIVELOG_FOLDER)
|
||||
.setArchiveLogFolder(HOODIE_ARCHIVELOG_FOLDER_PROP.defaultValue())
|
||||
.setTimelineLayoutVersion(1)
|
||||
.initTable(hadoopConf, basePath);
|
||||
LOG.info("Table initialized under base path {}", basePath);
|
||||
|
||||
Reference in New Issue
Block a user