diff --git a/hudi-cli/src/main/java/org/apache/hudi/cli/Table.java b/hudi-cli/src/main/java/org/apache/hudi/cli/Table.java index 2efad37de..bebc7fc8d 100644 --- a/hudi-cli/src/main/java/org/apache/hudi/cli/Table.java +++ b/hudi-cli/src/main/java/org/apache/hudi/cli/Table.java @@ -85,7 +85,7 @@ public class Table implements Iterable> { /** * Add all rows. * - * @param rows Rows to be aded + * @param rows Rows to be added * @return */ public Table addAll(List> rows) { diff --git a/hudi-client/src/main/java/org/apache/hudi/HoodieWriteClient.java b/hudi-client/src/main/java/org/apache/hudi/HoodieWriteClient.java index 81ab68b08..0f5411648 100644 --- a/hudi-client/src/main/java/org/apache/hudi/HoodieWriteClient.java +++ b/hudi-client/src/main/java/org/apache/hudi/HoodieWriteClient.java @@ -92,7 +92,7 @@ import scala.Tuple2; * Hoodie Write Client helps you build datasets on HDFS [insert()] and then perform efficient mutations on an HDFS * dataset [upsert()] *

- * Note that, at any given time, there can only be one Spark job performing these operatons on a Hoodie dataset. + * Note that, at any given time, there can only be one Spark job performing these operations on a Hoodie dataset. */ public class HoodieWriteClient extends AbstractHoodieClient {