[MINOR] Remove redundant plus operator (#1097)
This commit is contained in:
@@ -135,7 +135,7 @@ public class CompactionOperation implements Serializable {
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
return "CompactionOperation{" + "baseInstantTime='" + baseInstantTime + '\'' + ", dataFileCommitTime="
|
||||
return "CompactionOperation{baseInstantTime='" + baseInstantTime + '\'' + ", dataFileCommitTime="
|
||||
+ dataFileCommitTime + ", deltaFileNames=" + deltaFileNames + ", dataFileName=" + dataFileName + ", id='" + id
|
||||
+ '\'' + ", metrics=" + metrics + '}';
|
||||
}
|
||||
|
||||
@@ -341,7 +341,7 @@ public class HoodieCommitMetadata implements Serializable {
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
return "HoodieCommitMetadata{" + "partitionToWriteStats=" + partitionToWriteStats + ", compacted=" + compacted
|
||||
return "HoodieCommitMetadata{partitionToWriteStats=" + partitionToWriteStats + ", compacted=" + compacted
|
||||
+ ", extraMetadataMap=" + extraMetadataMap + '}';
|
||||
}
|
||||
}
|
||||
|
||||
@@ -98,6 +98,6 @@ public class HoodieDataFile implements Serializable {
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
return "HoodieDataFile{" + "fullPath=" + fullPath + ", fileLen=" + fileLen + '}';
|
||||
return "HoodieDataFile{fullPath=" + fullPath + ", fileLen=" + fileLen + '}';
|
||||
}
|
||||
}
|
||||
|
||||
@@ -62,6 +62,6 @@ public class HoodieFileGroupId implements Serializable {
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
return "HoodieFileGroupId{" + "partitionPath='" + partitionPath + '\'' + ", fileId='" + fileId + '\'' + '}';
|
||||
return "HoodieFileGroupId{partitionPath='" + partitionPath + '\'' + ", fileId='" + fileId + '\'' + '}';
|
||||
}
|
||||
}
|
||||
|
||||
@@ -179,6 +179,6 @@ public class HoodieLogFile implements Serializable {
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
return "HoodieLogFile{" + "pathStr='" + pathStr + '\'' + ", fileLen=" + fileLen + '}';
|
||||
return "HoodieLogFile{pathStr='" + pathStr + '\'' + ", fileLen=" + fileLen + '}';
|
||||
}
|
||||
}
|
||||
|
||||
@@ -101,7 +101,7 @@ public class HoodiePartitionMetadata {
|
||||
fs.rename(tmpMetaPath, metaPath);
|
||||
}
|
||||
} catch (IOException ioe) {
|
||||
LOG.warn("Error trying to save partition metadata (this is okay, as long as " + "atleast 1 of these succced), "
|
||||
LOG.warn("Error trying to save partition metadata (this is okay, as long as atleast 1 of these succced), "
|
||||
+ partitionPath, ioe);
|
||||
} finally {
|
||||
if (!metafileExists) {
|
||||
|
||||
@@ -320,7 +320,7 @@ public class HoodieWriteStat implements Serializable {
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
return "HoodieWriteStat{" + "fileId='" + fileId + '\'' + ", path='" + path + '\'' + ", prevCommit='" + prevCommit
|
||||
return "HoodieWriteStat{fileId='" + fileId + '\'' + ", path='" + path + '\'' + ", prevCommit='" + prevCommit
|
||||
+ '\'' + ", numWrites=" + numWrites + ", numDeletes=" + numDeletes + ", numUpdateWrites=" + numUpdateWrites
|
||||
+ ", totalWriteBytes=" + totalWriteBytes + ", totalWriteErrors=" + totalWriteErrors + ", tempPath='" + tempPath
|
||||
+ '\'' + ", partitionPath='" + partitionPath + '\'' + ", totalLogRecords=" + totalLogRecords
|
||||
|
||||
@@ -140,6 +140,6 @@ public final class Option<T> implements Serializable {
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
return "Option{" + "val=" + val + '}';
|
||||
return "Option{val=" + val + '}';
|
||||
}
|
||||
}
|
||||
|
||||
@@ -129,7 +129,7 @@ public class ParquetUtils {
|
||||
footerVals.put(footerName, metadata.get(footerName));
|
||||
} else if (required) {
|
||||
throw new MetadataNotFoundException(
|
||||
"Could not find index in Parquet footer. " + "Looked for key " + footerName + " in " + parquetFilePath);
|
||||
"Could not find index in Parquet footer. Looked for key " + footerName + " in " + parquetFilePath);
|
||||
}
|
||||
}
|
||||
return footerVals;
|
||||
|
||||
@@ -65,7 +65,7 @@ public class SpillableMapUtils {
|
||||
long crcOfReadValue = generateChecksum(value);
|
||||
if (crc != crcOfReadValue) {
|
||||
throw new HoodieCorruptedDataException(
|
||||
"checksum of payload written to external disk does not match, " + "data may be corrupted");
|
||||
"checksum of payload written to external disk does not match, data may be corrupted");
|
||||
}
|
||||
return new FileEntry(crc, keySize, valueSize, key, value, timestamp);
|
||||
}
|
||||
|
||||
@@ -64,7 +64,7 @@ public class TimelineDiffHelper {
|
||||
if (!lostPendingCompactions.isEmpty()) {
|
||||
// If a compaction is unscheduled, fall back to complete refresh of fs view since some log files could have been
|
||||
// moved. Its unsafe to incrementally sync in that case.
|
||||
LOG.warn("Some pending compactions are no longer in new timeline (unscheduled ?)." + "They are :"
|
||||
LOG.warn("Some pending compactions are no longer in new timeline (unscheduled ?). They are :"
|
||||
+ lostPendingCompactions);
|
||||
return TimelineDiffResult.UNSAFE_SYNC_RESULT;
|
||||
}
|
||||
@@ -132,7 +132,7 @@ public class TimelineDiffHelper {
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
return "TimelineDiffResult{" + "newlySeenInstants=" + newlySeenInstants + ", finishedCompactionInstants="
|
||||
return "TimelineDiffResult{newlySeenInstants=" + newlySeenInstants + ", finishedCompactionInstants="
|
||||
+ finishedCompactionInstants + ", canSyncIncrementally=" + canSyncIncrementally + '}';
|
||||
}
|
||||
}
|
||||
|
||||
@@ -103,7 +103,7 @@ public class TestHoodieLogFormatAppendFailure {
|
||||
HoodieAvroDataBlock dataBlock = new HoodieAvroDataBlock(records, header);
|
||||
|
||||
Writer writer = HoodieLogFormat.newWriterBuilder().onParentPath(testPath)
|
||||
.withFileExtension(HoodieArchivedLogFile.ARCHIVE_EXTENSION).withFileId("commits" + ".archive")
|
||||
.withFileExtension(HoodieArchivedLogFile.ARCHIVE_EXTENSION).withFileId("commits.archive")
|
||||
.overBaseCommit("").withFs(fs).build();
|
||||
|
||||
writer = writer.appendBlock(dataBlock);
|
||||
@@ -134,7 +134,7 @@ public class TestHoodieLogFormatAppendFailure {
|
||||
// Opening a new Writer right now will throw IOException. The code should handle this, rollover the logfile and
|
||||
// return a new writer with a bumped up logVersion
|
||||
writer = HoodieLogFormat.newWriterBuilder().onParentPath(testPath)
|
||||
.withFileExtension(HoodieArchivedLogFile.ARCHIVE_EXTENSION).withFileId("commits" + ".archive")
|
||||
.withFileExtension(HoodieArchivedLogFile.ARCHIVE_EXTENSION).withFileId("commits.archive")
|
||||
.overBaseCommit("").withFs(fs).build();
|
||||
// The log version should be different for this new writer
|
||||
Assert.assertFalse(writer.getLogFile().getLogVersion() == logFileVersion);
|
||||
|
||||
@@ -30,8 +30,8 @@ import java.util.Map;
|
||||
*/
|
||||
public class TestHoodieAvroUtils {
|
||||
|
||||
private static String EXAMPLE_SCHEMA = "{\"type\": \"record\"," + "\"name\": \"testrec\"," + "\"fields\": [ "
|
||||
+ "{\"name\": \"timestamp\",\"type\": \"double\"}," + "{\"name\": \"_row_key\", \"type\": \"string\"},"
|
||||
private static String EXAMPLE_SCHEMA = "{\"type\": \"record\",\"name\": \"testrec\",\"fields\": [ "
|
||||
+ "{\"name\": \"timestamp\",\"type\": \"double\"},{\"name\": \"_row_key\", \"type\": \"string\"},"
|
||||
+ "{\"name\": \"non_pii_col\", \"type\": \"string\"},"
|
||||
+ "{\"name\": \"pii_col\", \"type\": \"string\", \"column_category\": \"user_profile\"}]}";
|
||||
|
||||
|
||||
Reference in New Issue
Block a user