1
0

MINOR_CHECKSTYLE (#3616)

Fix checkstyle
This commit is contained in:
liujinhui
2021-09-07 18:19:39 +08:00
committed by GitHub
parent cf002b6918
commit eb5e7eec0a
11 changed files with 19 additions and 19 deletions

View File

@@ -77,9 +77,9 @@ public class ValidateAsyncOperations extends DagNode<Option<String>> {
} }
if (config.validateArchival() || config.validateClean()) { if (config.validateArchival() || config.validateClean()) {
Pattern ARCHIVE_FILE_PATTERN = final Pattern ARCHIVE_FILE_PATTERN =
Pattern.compile("\\.commits_\\.archive\\..*"); Pattern.compile("\\.commits_\\.archive\\..*");
Pattern CLEAN_FILE_PATTERN = final Pattern CLEAN_FILE_PATTERN =
Pattern.compile(".*\\.clean\\..*"); Pattern.compile(".*\\.clean\\..*");
String metadataPath = executionContext.getHoodieTestSuiteWriter().getCfg().targetBasePath + "/.hoodie"; String metadataPath = executionContext.getHoodieTestSuiteWriter().getCfg().targetBasePath + "/.hoodie";

View File

@@ -72,8 +72,8 @@ public class HoodieTestHiveBase extends ITTestBase {
} }
// Run Hoodie Java App // Run Hoodie Java App
String cmd = String.format("%s --hive-sync --table-path %s --hive-url %s --table-type %s --hive-table %s" + String cmd = String.format("%s --hive-sync --table-path %s --hive-url %s --table-type %s --hive-table %s"
" --commit-type %s --table-name %s", HOODIE_GENERATE_APP, hdfsUrl, HIVE_SERVER_JDBC_URL, + " --commit-type %s --table-name %s", HOODIE_GENERATE_APP, hdfsUrl, HIVE_SERVER_JDBC_URL,
tableType, hiveTableName, commitType, hoodieTableName); tableType, hiveTableName, commitType, hoodieTableName);
if (partitionType == PartitionType.MULTI_KEYS_PARTITIONED) { if (partitionType == PartitionType.MULTI_KEYS_PARTITIONED) {
cmd = cmd + " --use-multi-partition-keys"; cmd = cmd + " --use-multi-partition-keys";

View File

@@ -90,7 +90,6 @@ public class ITTestHoodieDemo extends ITTestBase {
+ " --hoodie-conf hoodie.datasource.hive_sync.database=default " + " --hoodie-conf hoodie.datasource.hive_sync.database=default "
+ " --hoodie-conf hoodie.datasource.hive_sync.table=%s"; + " --hoodie-conf hoodie.datasource.hive_sync.table=%s";
@AfterEach @AfterEach
public void clean() throws Exception { public void clean() throws Exception {
String hdfsCmd = "hdfs dfs -rm -R "; String hdfsCmd = "hdfs dfs -rm -R ";

View File

@@ -57,8 +57,8 @@ public class TestGenericRecordPayloadGenerator {
@Test @Test
public void testComplexPayload() throws IOException { public void testComplexPayload() throws IOException {
Schema schema = new Schema.Parser().parse(UtilitiesTestBase.Helpers Schema schema = new Schema.Parser().parse(UtilitiesTestBase.Helpers
.readFileFromAbsolutePath(System.getProperty("user.dir") + "/.." + .readFileFromAbsolutePath(System.getProperty("user.dir") + "/.."
COMPLEX_SOURCE_SCHEMA_DOCKER_DEMO_RELATIVE_PATH)); + COMPLEX_SOURCE_SCHEMA_DOCKER_DEMO_RELATIVE_PATH));
GenericRecordFullPayloadGenerator payloadGenerator = new GenericRecordFullPayloadGenerator(schema); GenericRecordFullPayloadGenerator payloadGenerator = new GenericRecordFullPayloadGenerator(schema);
GenericRecord record = payloadGenerator.getNewPayload(); GenericRecord record = payloadGenerator.getNewPayload();
// The generated payload should validate with the provided schema // The generated payload should validate with the provided schema
@@ -68,8 +68,8 @@ public class TestGenericRecordPayloadGenerator {
@Test @Test
public void testComplexPartialPayload() throws IOException { public void testComplexPartialPayload() throws IOException {
Schema schema = new Schema.Parser().parse(UtilitiesTestBase.Helpers Schema schema = new Schema.Parser().parse(UtilitiesTestBase.Helpers
.readFileFromAbsolutePath(System.getProperty("user.dir") + "/.." + .readFileFromAbsolutePath(System.getProperty("user.dir") + "/.."
COMPLEX_SOURCE_SCHEMA_DOCKER_DEMO_RELATIVE_PATH)); + COMPLEX_SOURCE_SCHEMA_DOCKER_DEMO_RELATIVE_PATH));
GenericRecordPartialPayloadGenerator payloadGenerator = new GenericRecordPartialPayloadGenerator(schema); GenericRecordPartialPayloadGenerator payloadGenerator = new GenericRecordPartialPayloadGenerator(schema);
IntStream.range(0, 10).forEach(a -> { IntStream.range(0, 10).forEach(a -> {
GenericRecord record = payloadGenerator.getNewPayload(); GenericRecord record = payloadGenerator.getNewPayload();
@@ -124,8 +124,8 @@ public class TestGenericRecordPayloadGenerator {
@Test @Test
public void testComplexPayloadWithLargeMinSize() throws Exception { public void testComplexPayloadWithLargeMinSize() throws Exception {
Schema schema = new Schema.Parser().parse(UtilitiesTestBase.Helpers Schema schema = new Schema.Parser().parse(UtilitiesTestBase.Helpers
.readFileFromAbsolutePath(System.getProperty("user.dir") + "/.." + .readFileFromAbsolutePath(System.getProperty("user.dir") + "/.."
COMPLEX_SOURCE_SCHEMA_DOCKER_DEMO_RELATIVE_PATH)); + COMPLEX_SOURCE_SCHEMA_DOCKER_DEMO_RELATIVE_PATH));
int minPayloadSize = 10000; int minPayloadSize = 10000;
GenericRecordFullPayloadGenerator payloadGenerator = new GenericRecordFullPayloadGenerator( GenericRecordFullPayloadGenerator payloadGenerator = new GenericRecordFullPayloadGenerator(
schema, minPayloadSize); schema, minPayloadSize);

View File

@@ -51,7 +51,8 @@ class TestSchemaRegistryProvider {
put("hoodie.deltastreamer.schemaprovider.registry.urlSuffix", "-value"); put("hoodie.deltastreamer.schemaprovider.registry.urlSuffix", "-value");
put("hoodie.deltastreamer.schemaprovider.registry.url", "http://foo:bar@localhost"); put("hoodie.deltastreamer.schemaprovider.registry.url", "http://foo:bar@localhost");
put("hoodie.deltastreamer.source.kafka.topic", "foo"); put("hoodie.deltastreamer.source.kafka.topic", "foo");
}}; }
};
} }
private Schema getExpectedSchema(String response) throws IOException { private Schema getExpectedSchema(String response) throws IOException {