[MINOR] Fix typos in Spark client related classes (#4781)
This commit is contained in:
@@ -171,7 +171,7 @@ public class TestClientRollback extends HoodieClientTestBase {
|
||||
}
|
||||
|
||||
/**
|
||||
* Test Cases for effects of rollbacking completed/inflight commits.
|
||||
* Test Cases for effects of rolling back completed/inflight commits.
|
||||
*/
|
||||
@Test
|
||||
public void testRollbackCommit() throws Exception {
|
||||
|
||||
@@ -584,7 +584,7 @@ public class TestHoodieClientMultiWriter extends HoodieClientTestBase {
|
||||
private void createCommitWithInserts(HoodieWriteConfig cfg, SparkRDDWriteClient client,
|
||||
String prevCommitTime, String newCommitTime, int numRecords,
|
||||
boolean doCommit) throws Exception {
|
||||
// Finish first base commmit
|
||||
// Finish first base commit
|
||||
JavaRDD<WriteStatus> result = insertFirstBatch(cfg, client, newCommitTime, prevCommitTime, numRecords, SparkRDDWriteClient::bulkInsert,
|
||||
false, false, numRecords);
|
||||
if (doCommit) {
|
||||
|
||||
@@ -147,7 +147,7 @@ public class TestTableSchemaEvolution extends HoodieClientTestBase {
|
||||
+ TIP_NESTED_SCHEMA + EXTRA_FIELD_SCHEMA + EXTRA_FIELD_SCHEMA.replace("new_field", "new_new_field")
|
||||
+ TRIP_SCHEMA_SUFFIX;
|
||||
assertTrue(TableSchemaResolver.isSchemaCompatible(TRIP_EXAMPLE_SCHEMA, multipleAddedFieldSchema),
|
||||
"Multiple added fields with defauls are compatible");
|
||||
"Multiple added fields with defaults are compatible");
|
||||
|
||||
assertFalse(TableSchemaResolver.isSchemaCompatible(TRIP_EXAMPLE_SCHEMA,
|
||||
TRIP_SCHEMA_PREFIX + EXTRA_TYPE_SCHEMA + MAP_TYPE_SCHEMA
|
||||
@@ -205,7 +205,7 @@ public class TestTableSchemaEvolution extends HoodieClientTestBase {
|
||||
final List<HoodieRecord> failedRecords = generateInsertsWithSchema("004", numRecords, TRIP_EXAMPLE_SCHEMA_DEVOLVED);
|
||||
try {
|
||||
// We cannot use insertBatch directly here because we want to insert records
|
||||
// with a devolved schema and insertBatch inserts records using the TRIP_EXMPLE_SCHEMA.
|
||||
// with a devolved schema and insertBatch inserts records using the TRIP_EXAMPLE_SCHEMA.
|
||||
writeBatch(client, "005", "004", Option.empty(), "003", numRecords,
|
||||
(String s, Integer a) -> failedRecords, SparkRDDWriteClient::insert, false, 0, 0, 0, false);
|
||||
fail("Insert with devolved scheme should fail");
|
||||
@@ -233,7 +233,7 @@ public class TestTableSchemaEvolution extends HoodieClientTestBase {
|
||||
client = getHoodieWriteClient(hoodieEvolvedWriteConfig);
|
||||
|
||||
// We cannot use insertBatch directly here because we want to insert records
|
||||
// with a evolved schemaand insertBatch inserts records using the TRIP_EXMPLE_SCHEMA.
|
||||
// with an evolved schema and insertBatch inserts records using the TRIP_EXAMPLE_SCHEMA.
|
||||
final List<HoodieRecord> evolvedRecords = generateInsertsWithSchema("005", numRecords, TRIP_EXAMPLE_SCHEMA_EVOLVED);
|
||||
writeBatch(client, "005", "004", Option.empty(), initCommitTime, numRecords,
|
||||
(String s, Integer a) -> evolvedRecords, SparkRDDWriteClient::insert, false, 0, 0, 0, false);
|
||||
|
||||
@@ -80,7 +80,7 @@ public class TestKeyRangeLookupTree {
|
||||
* Tests for many duplicate entries in the tree.
|
||||
*/
|
||||
@Test
|
||||
public void testFileGroupLookUpManyDulicateEntries() {
|
||||
public void testFileGroupLookUpManyDuplicateEntries() {
|
||||
KeyRangeNode toInsert = new KeyRangeNode(Long.toString(1200), Long.toString(2000), UUID.randomUUID().toString());
|
||||
updateExpectedMatchesToTest(toInsert);
|
||||
keyRangeLookupTree.insert(toInsert);
|
||||
|
||||
@@ -191,7 +191,7 @@ public class TestSparkHoodieHBaseIndex extends SparkClientFunctionalTestHarness
|
||||
final String newCommitTime = "001";
|
||||
final int numRecords = 10;
|
||||
final String oldPartitionPath = "1970/01/01";
|
||||
final String emptyHoodieRecordPayloadClasssName = EmptyHoodieRecordPayload.class.getName();
|
||||
final String emptyHoodieRecordPayloadClassName = EmptyHoodieRecordPayload.class.getName();
|
||||
|
||||
List<HoodieRecord> newRecords = dataGen.generateInserts(newCommitTime, numRecords);
|
||||
List<HoodieRecord> oldRecords = new LinkedList();
|
||||
@@ -226,7 +226,7 @@ public class TestSparkHoodieHBaseIndex extends SparkClientFunctionalTestHarness
|
||||
assertEquals(numRecords * 2L, taggedRecords.stream().count());
|
||||
// Verify the number of deleted records
|
||||
assertEquals(numRecords, taggedRecords.stream().filter(record -> record.getKey().getPartitionPath().equals(oldPartitionPath)
|
||||
&& record.getData().getClass().getName().equals(emptyHoodieRecordPayloadClasssName)).count());
|
||||
&& record.getData().getClass().getName().equals(emptyHoodieRecordPayloadClassName)).count());
|
||||
// Verify the number of inserted records
|
||||
assertEquals(numRecords, taggedRecords.stream().filter(record -> !record.getKey().getPartitionPath().equals(oldPartitionPath)).count());
|
||||
|
||||
|
||||
@@ -18,7 +18,6 @@
|
||||
|
||||
package org.apache.hudi.io;
|
||||
|
||||
import org.apache.hadoop.fs.FileStatus;
|
||||
import org.apache.hudi.avro.model.HoodieRollbackMetadata;
|
||||
import org.apache.hudi.client.utils.MetadataConversionUtils;
|
||||
import org.apache.hudi.common.config.HoodieMetadataConfig;
|
||||
@@ -52,6 +51,7 @@ import org.apache.hudi.table.HoodieTimelineArchiveLog;
|
||||
import org.apache.hudi.testutils.HoodieClientTestHarness;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.fs.FileStatus;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.log4j.LogManager;
|
||||
import org.apache.log4j.Logger;
|
||||
@@ -655,7 +655,8 @@ public class TestHoodieTimelineArchiveLog extends HoodieClientTestHarness {
|
||||
public void testArchiveTableWithCleanCommits(boolean enableMetadata) throws Exception {
|
||||
HoodieWriteConfig writeConfig = initTestTableAndGetWriteConfig(enableMetadata, 2, 4, 2);
|
||||
|
||||
// min archival commits is 2 and max archival commits is 4(either clean commits has to be > 4 or commits has to be greater than 4.
|
||||
// min archival commits is 2 and max archival commits is 4
|
||||
// (either clean commits has to be > 4 or commits has to be greater than 4)
|
||||
// and so, after 5th commit, 3 commits will be archived.
|
||||
// 1,2,3,4,5,6 : after archival -> 1,5,6 (because, 2,3,4,5 and 6 are clean commits and are eligible for archival)
|
||||
// after 7th and 8th commit no-op wrt archival.
|
||||
|
||||
@@ -25,8 +25,8 @@ import org.apache.hudi.keygen.KeyGenerator;
|
||||
import org.apache.hudi.keygen.SimpleKeyGenerator;
|
||||
import org.apache.hudi.keygen.TestComplexKeyGenerator;
|
||||
import org.apache.hudi.keygen.constant.KeyGeneratorOptions;
|
||||
|
||||
import org.apache.hudi.keygen.constant.KeyGeneratorType;
|
||||
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
@@ -58,7 +58,7 @@ public class TestHoodieSparkKeyGeneratorFactory {
|
||||
// set both class name and keyGenerator type
|
||||
props.put(HoodieWriteConfig.KEYGENERATOR_TYPE.key(), KeyGeneratorType.CUSTOM.name());
|
||||
KeyGenerator keyGenerator3 = HoodieSparkKeyGeneratorFactory.createKeyGenerator(props);
|
||||
// KEYGENERATOR_TYPE_PROP was overitten by KEYGENERATOR_CLASS_PROP
|
||||
// KEYGENERATOR_TYPE_PROP was overwritten by KEYGENERATOR_CLASS_PROP
|
||||
Assertions.assertEquals(SimpleKeyGenerator.class.getName(), keyGenerator3.getClass().getName());
|
||||
|
||||
// set wrong class name
|
||||
|
||||
@@ -169,9 +169,9 @@ public class TestConsistencyGuard extends HoodieClientTestHarness {
|
||||
return getConsistencyGuardConfig(3, 10, 10);
|
||||
}
|
||||
|
||||
private ConsistencyGuardConfig getConsistencyGuardConfig(int maxChecks, int initalSleep, int maxSleep) {
|
||||
private ConsistencyGuardConfig getConsistencyGuardConfig(int maxChecks, int initialSleep, int maxSleep) {
|
||||
return ConsistencyGuardConfig.newBuilder().withConsistencyCheckEnabled(true)
|
||||
.withInitialConsistencyCheckIntervalMs(initalSleep).withMaxConsistencyCheckIntervalMs(maxSleep)
|
||||
.withInitialConsistencyCheckIntervalMs(initialSleep).withMaxConsistencyCheckIntervalMs(maxSleep)
|
||||
.withMaxConsistencyChecks(maxChecks).build();
|
||||
}
|
||||
}
|
||||
|
||||
@@ -18,8 +18,6 @@
|
||||
|
||||
package org.apache.hudi.table.action.compact;
|
||||
|
||||
import org.apache.hadoop.fs.FileStatus;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.hudi.client.HoodieReadClient;
|
||||
import org.apache.hudi.client.SparkRDDWriteClient;
|
||||
import org.apache.hudi.common.config.HoodieMetadataConfig;
|
||||
@@ -32,6 +30,9 @@ import org.apache.hudi.common.table.timeline.HoodieTimeline;
|
||||
import org.apache.hudi.config.HoodieWriteConfig;
|
||||
import org.apache.hudi.table.HoodieSparkTable;
|
||||
import org.apache.hudi.table.HoodieTable;
|
||||
|
||||
import org.apache.hadoop.fs.FileStatus;
|
||||
import org.apache.hadoop.fs.Path;
|
||||
import org.apache.spark.api.java.JavaRDD;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
@@ -204,8 +205,8 @@ public class TestAsyncCompaction extends CompactionTestBase {
|
||||
String compactionInstantTime = "006";
|
||||
int numRecs = 2000;
|
||||
|
||||
final List<HoodieRecord> initalRecords = dataGen.generateInserts(firstInstantTime, numRecs);
|
||||
final List<HoodieRecord> records = runNextDeltaCommits(client, readClient, Arrays.asList(firstInstantTime, secondInstantTime), initalRecords, cfg, true,
|
||||
final List<HoodieRecord> initialRecords = dataGen.generateInserts(firstInstantTime, numRecs);
|
||||
final List<HoodieRecord> records = runNextDeltaCommits(client, readClient, Arrays.asList(firstInstantTime, secondInstantTime), initialRecords, cfg, true,
|
||||
new ArrayList<>());
|
||||
|
||||
// Schedule compaction but do not run them
|
||||
|
||||
@@ -28,6 +28,7 @@ import org.apache.hudi.config.HoodieCompactionConfig;
|
||||
import org.apache.hudi.config.HoodieWriteConfig;
|
||||
import org.apache.hudi.table.HoodieSparkTable;
|
||||
import org.apache.hudi.table.marker.WriteMarkersFactory;
|
||||
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
import java.util.ArrayList;
|
||||
@@ -62,7 +63,7 @@ public class TestInlineCompaction extends CompactionTestBase {
|
||||
runNextDeltaCommits(writeClient, readClient, instants, records, cfg, true, new ArrayList<>());
|
||||
HoodieTableMetaClient metaClient = HoodieTableMetaClient.builder().setConf(hadoopConf).setBasePath(cfg.getBasePath()).build();
|
||||
|
||||
// Then: ensure no compaction is executedm since there are only 2 delta commits
|
||||
// Then: ensure no compaction is executed since there are only 2 delta commits
|
||||
assertEquals(2, metaClient.getActiveTimeline().getWriteTimeline().countInstants());
|
||||
}
|
||||
}
|
||||
@@ -152,7 +153,7 @@ public class TestInlineCompaction extends CompactionTestBase {
|
||||
runNextDeltaCommits(writeClient, readClient, instants, records, cfg, true, new ArrayList<>());
|
||||
HoodieTableMetaClient metaClient = HoodieTableMetaClient.builder().setConf(hadoopConf).setBasePath(cfg.getBasePath()).build();
|
||||
|
||||
// Then: ensure no compaction is executedm since there are only 3 delta commits
|
||||
// Then: ensure no compaction is executed since there are only 3 delta commits
|
||||
assertEquals(3, metaClient.getActiveTimeline().getWriteTimeline().countInstants());
|
||||
// 4th commit, that will trigger compaction
|
||||
metaClient = HoodieTableMetaClient.builder().setConf(hadoopConf).setBasePath(cfg.getBasePath()).build();
|
||||
|
||||
@@ -143,10 +143,10 @@ public class TestHoodieCompactionStrategy {
|
||||
"DayBasedCompactionStrategy should have resulted in fewer compactions");
|
||||
assertEquals(2, returned.size(), "DayBasedCompactionStrategy should have resulted in fewer compactions");
|
||||
|
||||
int comparision = strategy.getComparator().compare(returned.get(returned.size() - 1).getPartitionPath(),
|
||||
int comparison = strategy.getComparator().compare(returned.get(returned.size() - 1).getPartitionPath(),
|
||||
returned.get(0).getPartitionPath());
|
||||
// Either the partition paths are sorted in descending order or they are equal
|
||||
assertTrue(comparision >= 0, "DayBasedCompactionStrategy should sort partitions in descending order");
|
||||
assertTrue(comparison >= 0, "DayBasedCompactionStrategy should sort partitions in descending order");
|
||||
}
|
||||
|
||||
@Test
|
||||
@@ -192,10 +192,10 @@ public class TestHoodieCompactionStrategy {
|
||||
assertEquals(5, returned.size(),
|
||||
"BoundedPartitionAwareCompactionStrategy should have resulted in fewer compactions");
|
||||
|
||||
int comparision = strategy.getComparator().compare(returned.get(returned.size() - 1).getPartitionPath(),
|
||||
int comparison = strategy.getComparator().compare(returned.get(returned.size() - 1).getPartitionPath(),
|
||||
returned.get(0).getPartitionPath());
|
||||
// Either the partition paths are sorted in descending order or they are equal
|
||||
assertTrue(comparision >= 0, "BoundedPartitionAwareCompactionStrategy should sort partitions in descending order");
|
||||
assertTrue(comparison >= 0, "BoundedPartitionAwareCompactionStrategy should sort partitions in descending order");
|
||||
}
|
||||
|
||||
@Test
|
||||
|
||||
@@ -33,6 +33,7 @@ import org.apache.hudi.config.HoodieWriteConfig;
|
||||
import org.apache.hudi.table.HoodieTable;
|
||||
import org.apache.hudi.testutils.Assertions;
|
||||
import org.apache.hudi.testutils.HoodieClientTestBase;
|
||||
|
||||
import org.apache.spark.api.java.JavaRDD;
|
||||
|
||||
import java.io.IOException;
|
||||
@@ -78,18 +79,18 @@ public class HoodieClientRollbackTestBase extends HoodieClientTestBase {
|
||||
}
|
||||
|
||||
|
||||
//2. assert filegroup and get the first partition fileslice
|
||||
//2. assert file group and get the first partition file slice
|
||||
HoodieTable table = this.getHoodieTable(metaClient, cfg);
|
||||
SyncableFileSystemView fsView = getFileSystemViewWithUnCommittedSlices(table.getMetaClient());
|
||||
List<HoodieFileGroup> firstPartitionCommit2FileGroups = fsView.getAllFileGroups(DEFAULT_FIRST_PARTITION_PATH).collect(Collectors.toList());
|
||||
assertEquals(1, firstPartitionCommit2FileGroups.size());
|
||||
firstPartitionCommit2FileSlices.addAll(firstPartitionCommit2FileGroups.get(0).getAllFileSlices().collect(Collectors.toList()));
|
||||
//3. assert filegroup and get the second partition fileslice
|
||||
//3. assert file group and get the second partition file slice
|
||||
List<HoodieFileGroup> secondPartitionCommit2FileGroups = fsView.getAllFileGroups(DEFAULT_SECOND_PARTITION_PATH).collect(Collectors.toList());
|
||||
assertEquals(1, secondPartitionCommit2FileGroups.size());
|
||||
secondPartitionCommit2FileSlices.addAll(secondPartitionCommit2FileGroups.get(0).getAllFileSlices().collect(Collectors.toList()));
|
||||
|
||||
//4. assert fileslice
|
||||
//4. assert file slice
|
||||
HoodieTableType tableType = this.getTableType();
|
||||
if (tableType.equals(HoodieTableType.COPY_ON_WRITE)) {
|
||||
assertEquals(2, firstPartitionCommit2FileSlices.size());
|
||||
|
||||
@@ -112,7 +112,7 @@ public class TestMergeOnReadRollbackActionExecutor extends HoodieClientRollbackT
|
||||
assertTrue(meta.getSuccessDeleteFiles() == null || meta.getSuccessDeleteFiles().size() == 0);
|
||||
}
|
||||
|
||||
//4. assert filegroup after rollback, and compare to the rollbackstat
|
||||
//4. assert file group after rollback, and compare to the rollbackstat
|
||||
// assert the first partition data and log file size
|
||||
List<HoodieFileGroup> firstPartitionRollBack1FileGroups = table.getFileSystemView().getAllFileGroups(DEFAULT_FIRST_PARTITION_PATH).collect(Collectors.toList());
|
||||
assertEquals(1, firstPartitionRollBack1FileGroups.size());
|
||||
|
||||
@@ -103,7 +103,7 @@ public abstract class TestWriteMarkersBase extends HoodieCommonTestHarness {
|
||||
@ParameterizedTest
|
||||
@ValueSource(booleans = {true, false})
|
||||
public void testDataPathsWhenCreatingOrMerging(boolean isTablePartitioned) throws IOException {
|
||||
// add markfiles
|
||||
// add marker files
|
||||
createSomeMarkers(isTablePartitioned);
|
||||
// add invalid file
|
||||
createInvalidFile(isTablePartitioned ? "2020/06/01" : "", "invalid_file3");
|
||||
|
||||
@@ -207,7 +207,7 @@ public class HoodieClientTestUtils {
|
||||
}
|
||||
|
||||
/**
|
||||
* Reads the paths under the a hoodie table out as a DataFrame.
|
||||
* Reads the paths under the hoodie table out as a DataFrame.
|
||||
*/
|
||||
public static Dataset<Row> read(JavaSparkContext jsc, String basePath, SQLContext sqlContext, FileSystem fs,
|
||||
String... paths) {
|
||||
|
||||
Reference in New Issue
Block a user