@@ -77,9 +77,9 @@ public class ValidateAsyncOperations extends DagNode<Option<String>> {
|
||||
}
|
||||
|
||||
if (config.validateArchival() || config.validateClean()) {
|
||||
Pattern ARCHIVE_FILE_PATTERN =
|
||||
final Pattern ARCHIVE_FILE_PATTERN =
|
||||
Pattern.compile("\\.commits_\\.archive\\..*");
|
||||
Pattern CLEAN_FILE_PATTERN =
|
||||
final Pattern CLEAN_FILE_PATTERN =
|
||||
Pattern.compile(".*\\.clean\\..*");
|
||||
|
||||
String metadataPath = executionContext.getHoodieTestSuiteWriter().getCfg().targetBasePath + "/.hoodie";
|
||||
|
||||
@@ -72,8 +72,8 @@ public class HoodieTestHiveBase extends ITTestBase {
|
||||
}
|
||||
|
||||
// Run Hoodie Java App
|
||||
String cmd = String.format("%s --hive-sync --table-path %s --hive-url %s --table-type %s --hive-table %s" +
|
||||
" --commit-type %s --table-name %s", HOODIE_GENERATE_APP, hdfsUrl, HIVE_SERVER_JDBC_URL,
|
||||
String cmd = String.format("%s --hive-sync --table-path %s --hive-url %s --table-type %s --hive-table %s"
|
||||
+ " --commit-type %s --table-name %s", HOODIE_GENERATE_APP, hdfsUrl, HIVE_SERVER_JDBC_URL,
|
||||
tableType, hiveTableName, commitType, hoodieTableName);
|
||||
if (partitionType == PartitionType.MULTI_KEYS_PARTITIONED) {
|
||||
cmd = cmd + " --use-multi-partition-keys";
|
||||
|
||||
@@ -90,7 +90,6 @@ public class ITTestHoodieDemo extends ITTestBase {
|
||||
+ " --hoodie-conf hoodie.datasource.hive_sync.database=default "
|
||||
+ " --hoodie-conf hoodie.datasource.hive_sync.table=%s";
|
||||
|
||||
|
||||
@AfterEach
|
||||
public void clean() throws Exception {
|
||||
String hdfsCmd = "hdfs dfs -rm -R ";
|
||||
|
||||
@@ -57,8 +57,8 @@ public class TestGenericRecordPayloadGenerator {
|
||||
@Test
|
||||
public void testComplexPayload() throws IOException {
|
||||
Schema schema = new Schema.Parser().parse(UtilitiesTestBase.Helpers
|
||||
.readFileFromAbsolutePath(System.getProperty("user.dir") + "/.." +
|
||||
COMPLEX_SOURCE_SCHEMA_DOCKER_DEMO_RELATIVE_PATH));
|
||||
.readFileFromAbsolutePath(System.getProperty("user.dir") + "/.."
|
||||
+ COMPLEX_SOURCE_SCHEMA_DOCKER_DEMO_RELATIVE_PATH));
|
||||
GenericRecordFullPayloadGenerator payloadGenerator = new GenericRecordFullPayloadGenerator(schema);
|
||||
GenericRecord record = payloadGenerator.getNewPayload();
|
||||
// The generated payload should validate with the provided schema
|
||||
@@ -68,8 +68,8 @@ public class TestGenericRecordPayloadGenerator {
|
||||
@Test
|
||||
public void testComplexPartialPayload() throws IOException {
|
||||
Schema schema = new Schema.Parser().parse(UtilitiesTestBase.Helpers
|
||||
.readFileFromAbsolutePath(System.getProperty("user.dir") + "/.." +
|
||||
COMPLEX_SOURCE_SCHEMA_DOCKER_DEMO_RELATIVE_PATH));
|
||||
.readFileFromAbsolutePath(System.getProperty("user.dir") + "/.."
|
||||
+ COMPLEX_SOURCE_SCHEMA_DOCKER_DEMO_RELATIVE_PATH));
|
||||
GenericRecordPartialPayloadGenerator payloadGenerator = new GenericRecordPartialPayloadGenerator(schema);
|
||||
IntStream.range(0, 10).forEach(a -> {
|
||||
GenericRecord record = payloadGenerator.getNewPayload();
|
||||
@@ -124,8 +124,8 @@ public class TestGenericRecordPayloadGenerator {
|
||||
@Test
|
||||
public void testComplexPayloadWithLargeMinSize() throws Exception {
|
||||
Schema schema = new Schema.Parser().parse(UtilitiesTestBase.Helpers
|
||||
.readFileFromAbsolutePath(System.getProperty("user.dir") + "/.." +
|
||||
COMPLEX_SOURCE_SCHEMA_DOCKER_DEMO_RELATIVE_PATH));
|
||||
.readFileFromAbsolutePath(System.getProperty("user.dir") + "/.."
|
||||
+ COMPLEX_SOURCE_SCHEMA_DOCKER_DEMO_RELATIVE_PATH));
|
||||
int minPayloadSize = 10000;
|
||||
GenericRecordFullPayloadGenerator payloadGenerator = new GenericRecordFullPayloadGenerator(
|
||||
schema, minPayloadSize);
|
||||
|
||||
@@ -51,7 +51,8 @@ class TestSchemaRegistryProvider {
|
||||
put("hoodie.deltastreamer.schemaprovider.registry.urlSuffix", "-value");
|
||||
put("hoodie.deltastreamer.schemaprovider.registry.url", "http://foo:bar@localhost");
|
||||
put("hoodie.deltastreamer.source.kafka.topic", "foo");
|
||||
}};
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
private Schema getExpectedSchema(String response) throws IOException {
|
||||
|
||||
Reference in New Issue
Block a user