[HUDI-1175] Commenting out testsuite tests from Integration tests until we investigate the CI flakiness (#1945)
This commit is contained in:
committed by
GitHub
parent
b2e703d442
commit
9c24151929
21
docker/demo/compaction-bootstrap.commands
Normal file
21
docker/demo/compaction-bootstrap.commands
Normal file
@@ -0,0 +1,21 @@
|
|||||||
|
|
||||||
|
# Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
# or more contributor license agreements. See the NOTICE file
|
||||||
|
# distributed with this work for additional information
|
||||||
|
# regarding copyright ownership. The ASF licenses this file
|
||||||
|
# to you under the Apache License, Version 2.0 (the
|
||||||
|
# "License"); you may not use this file except in compliance
|
||||||
|
# with the License. You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
|
||||||
|
connect --path /user/hive/warehouse/stock_ticks_mor_bs
|
||||||
|
compactions show all
|
||||||
|
compaction schedule --hoodieConfigs hoodie.compact.inline.max.delta.commits=1
|
||||||
|
compaction run --parallelism 2 --sparkMemory 1G --schemaFilePath /var/demo/config/schema.avsc --retry 1
|
||||||
@@ -19,7 +19,3 @@ connect --path /user/hive/warehouse/stock_ticks_mor
|
|||||||
compactions show all
|
compactions show all
|
||||||
compaction schedule --hoodieConfigs hoodie.compact.inline.max.delta.commits=1
|
compaction schedule --hoodieConfigs hoodie.compact.inline.max.delta.commits=1
|
||||||
compaction run --parallelism 2 --sparkMemory 1G --schemaFilePath /var/demo/config/schema.avsc --retry 1
|
compaction run --parallelism 2 --sparkMemory 1G --schemaFilePath /var/demo/config/schema.avsc --retry 1
|
||||||
connect --path /user/hive/warehouse/stock_ticks_mor_bs
|
|
||||||
compactions show all
|
|
||||||
compaction schedule --hoodieConfigs hoodie.compact.inline.max.delta.commits=1
|
|
||||||
compaction run --parallelism 2 --sparkMemory 1G --schemaFilePath /var/demo/config/schema.avsc --retry 1
|
|
||||||
|
|||||||
@@ -115,11 +115,9 @@ public abstract class ITTestBase {
|
|||||||
}
|
}
|
||||||
|
|
||||||
static String getPrestoConsoleCommand(String commandFile) {
|
static String getPrestoConsoleCommand(String commandFile) {
|
||||||
StringBuilder builder = new StringBuilder().append("presto --server " + PRESTO_COORDINATOR_URL)
|
return new StringBuilder().append("presto --server " + PRESTO_COORDINATOR_URL)
|
||||||
.append(" --catalog hive --schema default")
|
.append(" --catalog hive --schema default")
|
||||||
.append(" -f " + commandFile);
|
.append(" -f " + commandFile).toString();
|
||||||
System.out.println("Presto comamnd " + builder.toString());
|
|
||||||
return builder.toString();
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@BeforeEach
|
@BeforeEach
|
||||||
@@ -166,14 +164,14 @@ public abstract class ITTestBase {
|
|||||||
|
|
||||||
boolean completed =
|
boolean completed =
|
||||||
dockerClient.execStartCmd(createCmdResponse.getId()).withDetach(false).withTty(false).exec(callback)
|
dockerClient.execStartCmd(createCmdResponse.getId()).withDetach(false).withTty(false).exec(callback)
|
||||||
.awaitCompletion(900, SECONDS);
|
.awaitCompletion(540, SECONDS);
|
||||||
if (!completed) {
|
if (!completed) {
|
||||||
callback.getStderr().flush();
|
callback.getStderr().flush();
|
||||||
callback.getStdout().flush();
|
callback.getStdout().flush();
|
||||||
LOG.error("\n\n ###### Timed Out Command : " + Arrays.asList(command));
|
LOG.error("\n\n ###### Timed Out Command : " + Arrays.asList(command));
|
||||||
LOG.error("\n\n ###### Stderr of timed-out command #######\n" + callback.getStderr().toString());
|
LOG.error("\n\n ###### Stderr of timed-out command #######\n" + callback.getStderr().toString());
|
||||||
LOG.error("\n\n ###### stdout of timed-out command #######\n" + callback.getStderr().toString());
|
LOG.error("\n\n ###### stdout of timed-out command #######\n" + callback.getStdout().toString());
|
||||||
throw new TimeoutException("Command " + command + " has been running for more than 15 minutes. "
|
throw new TimeoutException("Command " + command + " has been running for more than 9 minutes. "
|
||||||
+ "Killing and failing !!");
|
+ "Killing and failing !!");
|
||||||
}
|
}
|
||||||
int exitCode = dockerClient.inspectExecCmd(createCmdResponse.getId()).exec().getExitCode();
|
int exitCode = dockerClient.inspectExecCmd(createCmdResponse.getId()).exec().getExitCode();
|
||||||
|
|||||||
@@ -65,6 +65,7 @@ public class ITTestHoodieDemo extends ITTestBase {
|
|||||||
private static final String MIN_COMMIT_TIME_MOR_SCRIPT = HOODIE_WS_ROOT + "/docker/demo/get_min_commit_time_mor.sh";
|
private static final String MIN_COMMIT_TIME_MOR_SCRIPT = HOODIE_WS_ROOT + "/docker/demo/get_min_commit_time_mor.sh";
|
||||||
private static final String HUDI_CLI_TOOL = HOODIE_WS_ROOT + "/hudi-cli/hudi-cli.sh";
|
private static final String HUDI_CLI_TOOL = HOODIE_WS_ROOT + "/hudi-cli/hudi-cli.sh";
|
||||||
private static final String COMPACTION_COMMANDS = HOODIE_WS_ROOT + "/docker/demo/compaction.commands";
|
private static final String COMPACTION_COMMANDS = HOODIE_WS_ROOT + "/docker/demo/compaction.commands";
|
||||||
|
private static final String COMPACTION_BOOTSTRAP_COMMANDS = HOODIE_WS_ROOT + "/docker/demo/compaction-bootstrap.commands";
|
||||||
private static final String SPARKSQL_BS_PREP_COMMANDS = HOODIE_WS_ROOT + "/docker/demo/sparksql-bootstrap-prep-source.commands";
|
private static final String SPARKSQL_BS_PREP_COMMANDS = HOODIE_WS_ROOT + "/docker/demo/sparksql-bootstrap-prep-source.commands";
|
||||||
private static final String SPARKSQL_BATCH1_COMMANDS = HOODIE_WS_ROOT + "/docker/demo/sparksql-batch1.commands";
|
private static final String SPARKSQL_BATCH1_COMMANDS = HOODIE_WS_ROOT + "/docker/demo/sparksql-batch1.commands";
|
||||||
private static final String SPARKSQL_BATCH2_COMMANDS = HOODIE_WS_ROOT + "/docker/demo/sparksql-batch2.commands";
|
private static final String SPARKSQL_BATCH2_COMMANDS = HOODIE_WS_ROOT + "/docker/demo/sparksql-batch2.commands";
|
||||||
@@ -407,5 +408,6 @@ public class ITTestHoodieDemo extends ITTestBase {
|
|||||||
|
|
||||||
private void scheduleAndRunCompaction() throws Exception {
|
private void scheduleAndRunCompaction() throws Exception {
|
||||||
executeCommandStringInDocker(ADHOC_1_CONTAINER, HUDI_CLI_TOOL + " --cmdfile " + COMPACTION_COMMANDS, true);
|
executeCommandStringInDocker(ADHOC_1_CONTAINER, HUDI_CLI_TOOL + " --cmdfile " + COMPACTION_COMMANDS, true);
|
||||||
|
executeCommandStringInDocker(ADHOC_1_CONTAINER, HUDI_CLI_TOOL + " --cmdfile " + COMPACTION_BOOTSTRAP_COMMANDS, true);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
1
pom.xml
1
pom.xml
@@ -1121,7 +1121,6 @@
|
|||||||
<configuration combine.self="override">
|
<configuration combine.self="override">
|
||||||
<skip>${skipITs}</skip>
|
<skip>${skipITs}</skip>
|
||||||
<includes>
|
<includes>
|
||||||
<include>**/testsuite/**/Test*.java</include>
|
|
||||||
<include>**/IT*.java</include>
|
<include>**/IT*.java</include>
|
||||||
</includes>
|
</includes>
|
||||||
</configuration>
|
</configuration>
|
||||||
|
|||||||
Reference in New Issue
Block a user