From e2d13c63052a90ac109e26cad05b04c3678aa953 Mon Sep 17 00:00:00 2001 From: Nishith Agarwal Date: Mon, 14 Aug 2017 11:57:38 -0700 Subject: [PATCH] Fix build failing issues --- .../test/java/com/uber/hoodie/TestMergeOnReadTable.java | 8 ++++---- .../com/uber/hoodie/common/HoodieClientTestUtils.java | 2 +- .../uber/hoodie/utilities/TestHDFSParquetImporter.java | 2 +- 3 files changed, 6 insertions(+), 6 deletions(-) diff --git a/hoodie-client/src/test/java/com/uber/hoodie/TestMergeOnReadTable.java b/hoodie-client/src/test/java/com/uber/hoodie/TestMergeOnReadTable.java index d13dd3373..1c86bed08 100644 --- a/hoodie-client/src/test/java/com/uber/hoodie/TestMergeOnReadTable.java +++ b/hoodie-client/src/test/java/com/uber/hoodie/TestMergeOnReadTable.java @@ -90,15 +90,15 @@ public class TestMergeOnReadTable { dfsCluster.shutdown();; } FSUtils.setFs(null); - // TEMPFIX(vc): Fix failing build - //FileSystem.closeAll(); + // Need to closeAll to clear FileSystem.Cache, required because DFS and LocalFS used in the same JVM + FileSystem.closeAll(); HoodieTestUtils.resetFS(); } @BeforeClass public static void setUpDFS() throws IOException { - // TEMPFIX(vc): Fix failing build - //FileSystem.closeAll(); + // Need to closeAll to clear FileSystem.Cache, required because DFS and LocalFS used in the same JVM + FileSystem.closeAll(); if (hdfsTestService == null) { hdfsTestService = new HdfsTestService(); dfsCluster = hdfsTestService.start(true); diff --git a/hoodie-client/src/test/java/com/uber/hoodie/common/HoodieClientTestUtils.java b/hoodie-client/src/test/java/com/uber/hoodie/common/HoodieClientTestUtils.java index 5b0158565..1aac5785f 100644 --- a/hoodie-client/src/test/java/com/uber/hoodie/common/HoodieClientTestUtils.java +++ b/hoodie-client/src/test/java/com/uber/hoodie/common/HoodieClientTestUtils.java @@ -87,7 +87,7 @@ public class HoodieClientTestUtils { SparkConf sparkConf = new SparkConf() .setAppName(appName) .set("spark.serializer", "org.apache.spark.serializer.KryoSerializer") - .setMaster("local[4]"); + .setMaster("local[1]"); return HoodieReadClient.addHoodieSupport(sparkConf); } } diff --git a/hoodie-utilities/src/test/java/com/uber/hoodie/utilities/TestHDFSParquetImporter.java b/hoodie-utilities/src/test/java/com/uber/hoodie/utilities/TestHDFSParquetImporter.java index 36021c2fa..38feb184c 100644 --- a/hoodie-utilities/src/test/java/com/uber/hoodie/utilities/TestHDFSParquetImporter.java +++ b/hoodie-utilities/src/test/java/com/uber/hoodie/utilities/TestHDFSParquetImporter.java @@ -284,7 +284,7 @@ public class TestHDFSParquetImporter implements Serializable { private JavaSparkContext getJavaSparkContext() { // Initialize a local spark env - SparkConf sparkConf = new SparkConf().setAppName("TestConversionCommand").setMaster("local[4]"); + SparkConf sparkConf = new SparkConf().setAppName("TestConversionCommand").setMaster("local[1]"); sparkConf = HoodieWriteClient.registerClasses(sparkConf); return new JavaSparkContext(HoodieReadClient.addHoodieSupport(sparkConf)); }