1
0

Reduce logging in unit-test runs

This commit is contained in:
Balaji Varadarajan
2019-05-24 22:20:10 -07:00
committed by Balaji Varadarajan
parent f2d91a455e
commit d0d2fa0337
8 changed files with 25 additions and 16 deletions

View File

@@ -158,7 +158,7 @@ public class TestCompactionAdminClient extends TestHoodieClientBase {
Assert.assertTrue("Rename Files must be empty", renameFiles.isEmpty()); Assert.assertTrue("Rename Files must be empty", renameFiles.isEmpty());
} }
expRenameFiles.entrySet().stream().forEach(r -> { expRenameFiles.entrySet().stream().forEach(r -> {
System.out.println("Key :" + r.getKey() + " renamed to " + r.getValue() + " rolled back to " logger.info("Key :" + r.getKey() + " renamed to " + r.getValue() + " rolled back to "
+ renameFilesFromUndo.get(r.getKey())); + renameFilesFromUndo.get(r.getKey()));
}); });

View File

@@ -50,6 +50,8 @@ import org.apache.avro.Schema;
import org.apache.avro.generic.GenericRecord; import org.apache.avro.generic.GenericRecord;
import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.log4j.LogManager;
import org.apache.log4j.Logger;
import org.apache.parquet.avro.AvroSchemaConverter; import org.apache.parquet.avro.AvroSchemaConverter;
import org.apache.parquet.hadoop.ParquetWriter; import org.apache.parquet.hadoop.ParquetWriter;
import org.apache.parquet.hadoop.metadata.CompressionCodecName; import org.apache.parquet.hadoop.metadata.CompressionCodecName;
@@ -64,6 +66,7 @@ import org.apache.spark.sql.SQLContext;
*/ */
public class HoodieClientTestUtils { public class HoodieClientTestUtils {
private static final transient Logger log = LogManager.getLogger(HoodieClientTestUtils.class);
public static List<WriteStatus> collectStatuses(Iterator<List<WriteStatus>> statusListItr) { public static List<WriteStatus> collectStatuses(Iterator<List<WriteStatus>> statusListItr) {
List<WriteStatus> statuses = new ArrayList<>(); List<WriteStatus> statuses = new ArrayList<>();
@@ -137,7 +140,7 @@ public class HoodieClientTestUtils {
try { try {
HashMap<String, String> paths = getLatestFileIDsToFullPath(basePath, commitTimeline, HashMap<String, String> paths = getLatestFileIDsToFullPath(basePath, commitTimeline,
Arrays.asList(commitInstant)); Arrays.asList(commitInstant));
System.out.println("Path :" + paths.values()); log.info("Path :" + paths.values());
return sqlContext.read().parquet(paths.values().toArray(new String[paths.size()])) return sqlContext.read().parquet(paths.values().toArray(new String[paths.size()]))
.filter(String.format("%s ='%s'", HoodieRecord.COMMIT_TIME_METADATA_FIELD, commitTime)); .filter(String.format("%s ='%s'", HoodieRecord.COMMIT_TIME_METADATA_FIELD, commitTime));
} catch (Exception e) { } catch (Exception e) {

View File

@@ -80,7 +80,6 @@ public class TestUpdateMapFunction implements Serializable {
public void testSchemaEvolutionOnUpdate() throws Exception { public void testSchemaEvolutionOnUpdate() throws Exception {
// Create a bunch of records with a old version of schema // Create a bunch of records with a old version of schema
final HoodieWriteConfig config = makeHoodieClientConfig("/exampleSchema.txt"); final HoodieWriteConfig config = makeHoodieClientConfig("/exampleSchema.txt");
System.out.println("JSC =" + jsc);
final HoodieCopyOnWriteTable table = new HoodieCopyOnWriteTable(config, jsc); final HoodieCopyOnWriteTable table = new HoodieCopyOnWriteTable(config, jsc);
final List<WriteStatus> statuses = jsc.parallelize(Arrays.asList(1)).map(x -> { final List<WriteStatus> statuses = jsc.parallelize(Arrays.asList(1)).map(x -> {

View File

@@ -50,6 +50,8 @@ import java.util.UUID;
import org.apache.avro.generic.GenericRecord; import org.apache.avro.generic.GenericRecord;
import org.apache.commons.io.IOUtils; import org.apache.commons.io.IOUtils;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.log4j.LogManager;
import org.apache.log4j.Logger;
import org.apache.parquet.avro.AvroReadSupport; import org.apache.parquet.avro.AvroReadSupport;
import org.apache.parquet.hadoop.ParquetReader; import org.apache.parquet.hadoop.ParquetReader;
import org.apache.spark.TaskContext; import org.apache.spark.TaskContext;
@@ -64,6 +66,8 @@ import scala.Tuple2;
public class TestCopyOnWriteTable { public class TestCopyOnWriteTable {
protected static Logger log = LogManager.getLogger(TestCopyOnWriteTable.class);
private String basePath = null; private String basePath = null;
private transient JavaSparkContext jsc = null; private transient JavaSparkContext jsc = null;
@@ -378,7 +382,7 @@ public class TestCopyOnWriteTable {
int counts = 0; int counts = 0;
for (File file : new File(basePath + "/2016/01/31").listFiles()) { for (File file : new File(basePath + "/2016/01/31").listFiles()) {
if (file.getName().endsWith(".parquet") && FSUtils.getCommitTime(file.getName()).equals(commitTime)) { if (file.getName().endsWith(".parquet") && FSUtils.getCommitTime(file.getName()).equals(commitTime)) {
System.out.println(file.getName() + "-" + file.length()); log.info(file.getName() + "-" + file.length());
counts++; counts++;
} }
} }

View File

@@ -86,7 +86,7 @@ public class RocksDBDAO {
*/ */
private void init() throws HoodieException { private void init() throws HoodieException {
try { try {
log.warn("DELETING RocksDB persisted at " + rocksDBBasePath); log.info("DELETING RocksDB persisted at " + rocksDBBasePath);
FileUtils.deleteDirectory(new File(rocksDBBasePath)); FileUtils.deleteDirectory(new File(rocksDBBasePath));
managedHandlesMap = new ConcurrentHashMap<>(); managedHandlesMap = new ConcurrentHashMap<>();

View File

@@ -56,6 +56,8 @@ import java.util.stream.Collectors;
import java.util.stream.Stream; import java.util.stream.Stream;
import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.log4j.LogManager;
import org.apache.log4j.Logger;
import org.junit.Assert; import org.junit.Assert;
import org.junit.Before; import org.junit.Before;
import org.junit.Rule; import org.junit.Rule;
@@ -65,6 +67,8 @@ import org.junit.rules.TemporaryFolder;
@SuppressWarnings("ResultOfMethodCallIgnored") @SuppressWarnings("ResultOfMethodCallIgnored")
public class HoodieTableFileSystemViewTest { public class HoodieTableFileSystemViewTest {
private static final transient Logger log = LogManager.getLogger(HoodieTableFileSystemViewTest.class);
private static String TEST_WRITE_TOKEN = "1-0-1"; private static String TEST_WRITE_TOKEN = "1-0-1";
protected HoodieTableMetaClient metaClient; protected HoodieTableMetaClient metaClient;
@@ -502,7 +506,7 @@ public class HoodieTableFileSystemViewTest {
roView.getAllDataFiles(partitionPath); roView.getAllDataFiles(partitionPath);
fileSliceList = rtView.getLatestFileSlices(partitionPath).collect(Collectors.toList()); fileSliceList = rtView.getLatestFileSlices(partitionPath).collect(Collectors.toList());
System.out.println("FILESLICE LIST=" + fileSliceList); log.info("FILESLICE LIST=" + fileSliceList);
dataFiles = fileSliceList.stream().map(FileSlice::getDataFile) dataFiles = fileSliceList.stream().map(FileSlice::getDataFile)
.filter(Optional::isPresent).map(Optional::get).collect(Collectors.toList()); .filter(Optional::isPresent).map(Optional::get).collect(Collectors.toList());
assertEquals("Expect only one data-files in latest view as there is only one file-group", 1, dataFiles.size()); assertEquals("Expect only one data-files in latest view as there is only one file-group", 1, dataFiles.size());

View File

@@ -61,6 +61,8 @@ import java.util.UUID;
import java.util.stream.Collectors; import java.util.stream.Collectors;
import java.util.stream.IntStream; import java.util.stream.IntStream;
import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.Path;
import org.apache.log4j.LogManager;
import org.apache.log4j.Logger;
import org.junit.Assert; import org.junit.Assert;
import org.junit.Before; import org.junit.Before;
import org.junit.Rule; import org.junit.Rule;
@@ -69,6 +71,8 @@ import org.junit.rules.TemporaryFolder;
public class IncrementalFSViewSyncTest { public class IncrementalFSViewSyncTest {
private static final transient Logger log = LogManager.getLogger(IncrementalFSViewSyncTest.class);
private static String TEST_WRITE_TOKEN = "1-0-1"; private static String TEST_WRITE_TOKEN = "1-0-1";
protected HoodieTableMetaClient metaClient; protected HoodieTableMetaClient metaClient;
@@ -344,7 +348,7 @@ public class IncrementalFSViewSyncTest {
Assert.assertEquals(newCleanerInstants.size(), cleanedInstants.size()); Assert.assertEquals(newCleanerInstants.size(), cleanedInstants.size());
long initialFileSlices = partitions.stream().mapToLong(p -> view.getAllFileSlices(p).count()).findAny().getAsLong(); long initialFileSlices = partitions.stream().mapToLong(p -> view.getAllFileSlices(p).count()).findAny().getAsLong();
long exp = initialFileSlices; long exp = initialFileSlices;
System.out.println("Initial File Slices :" + exp); log.info("Initial File Slices :" + exp);
for (int idx = 0; idx < newCleanerInstants.size(); idx++) { for (int idx = 0; idx < newCleanerInstants.size(); idx++) {
String instant = cleanedInstants.get(idx); String instant = cleanedInstants.get(idx);
try { try {
@@ -361,8 +365,8 @@ public class IncrementalFSViewSyncTest {
Assert.assertEquals(State.COMPLETED, view.getLastInstant().get().getState()); Assert.assertEquals(State.COMPLETED, view.getLastInstant().get().getState());
Assert.assertEquals(HoodieTimeline.CLEAN_ACTION, view.getLastInstant().get().getAction()); Assert.assertEquals(HoodieTimeline.CLEAN_ACTION, view.getLastInstant().get().getAction());
partitions.forEach(p -> { partitions.forEach(p -> {
System.out.println("PARTTITION : " + p); log.info("PARTTITION : " + p);
System.out.println("\tFileSlices :" + view.getAllFileSlices(p).collect(Collectors.toList())); log.info("\tFileSlices :" + view.getAllFileSlices(p).collect(Collectors.toList()));
}); });
partitions.forEach(p -> Assert.assertEquals(fileIdsPerPartition.size(), view.getLatestFileSlices(p).count())); partitions.forEach(p -> Assert.assertEquals(fileIdsPerPartition.size(), view.getLatestFileSlices(p).count()));
@@ -404,7 +408,7 @@ public class IncrementalFSViewSyncTest {
initialFileSlices - ((idx + 1) * fileIdsPerPartition.size()); initialFileSlices - ((idx + 1) * fileIdsPerPartition.size());
view.sync(); view.sync();
Assert.assertTrue(view.getLastInstant().isPresent()); Assert.assertTrue(view.getLastInstant().isPresent());
System.out.println("Last Instant is :" + view.getLastInstant().get()); log.info("Last Instant is :" + view.getLastInstant().get());
if (isRestore) { if (isRestore) {
Assert.assertEquals(newRestoreInstants.get(idx), view.getLastInstant().get().getTimestamp()); Assert.assertEquals(newRestoreInstants.get(idx), view.getLastInstant().get().getTimestamp());
Assert.assertEquals(isRestore ? HoodieTimeline.RESTORE_ACTION : HoodieTimeline.ROLLBACK_ACTION, Assert.assertEquals(isRestore ? HoodieTimeline.RESTORE_ACTION : HoodieTimeline.ROLLBACK_ACTION,
@@ -645,7 +649,7 @@ public class IncrementalFSViewSyncTest {
int multiple = begin; int multiple = begin;
for (int idx = 0; idx < instants.size(); idx++) { for (int idx = 0; idx < instants.size(); idx++) {
String instant = instants.get(idx); String instant = instants.get(idx);
System.out.println("Adding instant=" + instant); log.info("Adding instant=" + instant);
HoodieInstant lastInstant = lastInstants.get(idx); HoodieInstant lastInstant = lastInstants.get(idx);
// Add a non-empty ingestion to COW table // Add a non-empty ingestion to COW table
List<String> filePaths = addInstant(metaClient, instant, deltaCommit, List<String> filePaths = addInstant(metaClient, instant, deltaCommit,

View File

@@ -29,17 +29,12 @@ import java.util.Objects;
import java.util.UUID; import java.util.UUID;
import java.util.stream.Collectors; import java.util.stream.Collectors;
import java.util.stream.IntStream; import java.util.stream.IntStream;
import org.apache.log4j.Level;
import org.junit.AfterClass; import org.junit.AfterClass;
import org.junit.Assert; import org.junit.Assert;
import org.junit.Test; import org.junit.Test;
public class TestRocksDBManager { public class TestRocksDBManager {
static {
RocksDBDAO.log.setLevel(Level.INFO);
}
private static RocksDBDAO dbManager; private static RocksDBDAO dbManager;
@AfterClass @AfterClass