From ca427240c026bc6ba6e19b9b8b0a069a6330147a Mon Sep 17 00:00:00 2001 From: Yann Byron Date: Fri, 3 Dec 2021 16:37:13 +0800 Subject: [PATCH] [MINOR] use catalog schema if can not find table schema (#4182) --- .../apache/spark/sql/catalyst/catalog/HoodieCatalogTable.scala | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/catalyst/catalog/HoodieCatalogTable.scala b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/catalyst/catalog/HoodieCatalogTable.scala index 78081220e..fdf1b062e 100644 --- a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/catalyst/catalog/HoodieCatalogTable.scala +++ b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/catalyst/catalog/HoodieCatalogTable.scala @@ -117,7 +117,7 @@ class HoodieCatalogTable(val spark: SparkSession, val table: CatalogTable) exten * Make StructField nullable. */ lazy val tableSchema: StructType = { - val originSchema = getTableSqlSchema(metaClient, includeMetadataFields = true).get + val originSchema = getTableSqlSchema(metaClient, includeMetadataFields = true).getOrElse(table.schema) StructType(originSchema.map(_.copy(nullable = true))) }