diff --git a/hudi-spark/src/main/scala/org/apache/hudi/HoodieMergeOnReadRDD.scala b/hudi-spark/src/main/scala/org/apache/hudi/HoodieMergeOnReadRDD.scala index f272084b5..32d6a0945 100644 --- a/hudi-spark/src/main/scala/org/apache/hudi/HoodieMergeOnReadRDD.scala +++ b/hudi-spark/src/main/scala/org/apache/hudi/HoodieMergeOnReadRDD.scala @@ -31,7 +31,7 @@ import org.apache.spark.{Partition, SerializableWritable, SparkContext, TaskCont import org.apache.spark.rdd.RDD import org.apache.spark.sql.avro.{AvroDeserializer, AvroSerializer} import org.apache.spark.sql.catalyst.InternalRow -import org.apache.spark.sql.catalyst.expressions.{SpecificInternalRow, UnsafeProjection, UnsafeRow} +import org.apache.spark.sql.catalyst.expressions.{SpecificInternalRow, UnsafeProjection} import org.apache.spark.sql.execution.datasources.PartitionedFile import org.apache.spark.sql.vectorized.ColumnarBatch diff --git a/hudi-spark/src/main/scala/org/apache/hudi/HoodieSparkSqlWriter.scala b/hudi-spark/src/main/scala/org/apache/hudi/HoodieSparkSqlWriter.scala index b9731df8b..335a7cdb7 100644 --- a/hudi-spark/src/main/scala/org/apache/hudi/HoodieSparkSqlWriter.scala +++ b/hudi-spark/src/main/scala/org/apache/hudi/HoodieSparkSqlWriter.scala @@ -27,8 +27,8 @@ import org.apache.hadoop.fs.{FileSystem, Path} import org.apache.hadoop.hive.conf.HiveConf import org.apache.hudi.DataSourceWriteOptions._ import org.apache.hudi.avro.HoodieAvroUtils -import org.apache.hudi.client.{SparkRDDWriteClient, HoodieWriteResult} -import org.apache.hudi.client.{SparkRDDWriteClient, WriteStatus} +import org.apache.hudi.client.HoodieWriteResult +import org.apache.hudi.client.SparkRDDWriteClient import org.apache.hudi.common.config.TypedProperties import org.apache.hudi.common.model.{HoodieRecordPayload, HoodieTableType, WriteOperationType} import org.apache.hudi.common.table.{HoodieTableConfig, HoodieTableMetaClient} diff --git a/hudi-spark/src/main/scala/org/apache/hudi/HoodieStreamingSink.scala b/hudi-spark/src/main/scala/org/apache/hudi/HoodieStreamingSink.scala index 9f57fb53e..846212d7e 100644 --- a/hudi-spark/src/main/scala/org/apache/hudi/HoodieStreamingSink.scala +++ b/hudi-spark/src/main/scala/org/apache/hudi/HoodieStreamingSink.scala @@ -17,7 +17,7 @@ package org.apache.hudi import java.lang -import java.util.function.{Function, Supplier} +import java.util.function.Function import org.apache.hudi.async.{AsyncCompactService, SparkStreamingAsyncCompactService} import org.apache.hudi.client.SparkRDDWriteClient @@ -31,7 +31,7 @@ import org.apache.hudi.exception.HoodieCorruptedDataException import org.apache.log4j.LogManager import org.apache.spark.api.java.JavaSparkContext import org.apache.spark.sql.execution.streaming.Sink -import org.apache.spark.sql.streaming.{OutputMode, StreamingQueryListener} +import org.apache.spark.sql.streaming.OutputMode import org.apache.spark.sql.{DataFrame, SQLContext, SaveMode} import scala.util.{Failure, Success, Try} diff --git a/hudi-spark/src/main/scala/org/apache/hudi/IncrementalRelation.scala b/hudi-spark/src/main/scala/org/apache/hudi/IncrementalRelation.scala index ff68ef077..f8345780b 100644 --- a/hudi-spark/src/main/scala/org/apache/hudi/IncrementalRelation.scala +++ b/hudi-spark/src/main/scala/org/apache/hudi/IncrementalRelation.scala @@ -17,16 +17,9 @@ package org.apache.hudi -import com.google.common.collect.Lists -import org.apache.avro.Schema -import org.apache.hadoop.fs.GlobPattern -import org.apache.hadoop.fs.Path -import org.apache.hudi.avro.HoodieAvroUtils -import org.apache.hudi.common.bootstrap.index.BootstrapIndex import org.apache.hudi.common.model.{HoodieCommitMetadata, HoodieRecord, HoodieTableType} import org.apache.hudi.common.table.{HoodieTableMetaClient, TableSchemaResolver} import org.apache.hudi.common.table.timeline.HoodieTimeline -import org.apache.hudi.common.util.ParquetUtils import org.apache.hudi.config.HoodieWriteConfig import org.apache.hudi.exception.HoodieException import org.apache.hadoop.fs.GlobPattern @@ -36,7 +29,7 @@ import org.apache.log4j.LogManager import org.apache.spark.api.java.JavaSparkContext import org.apache.spark.rdd.RDD import org.apache.spark.sql.sources.{BaseRelation, TableScan} -import org.apache.spark.sql.types.{StringType, StructField, StructType} +import org.apache.spark.sql.types.StructType import org.apache.spark.sql.{DataFrame, Row, SQLContext} import scala.collection.JavaConversions._