1
0

[HUDI-553] Building/Running Hudi on higher java versions (#1369)

This commit is contained in:
lamber-ken
2020-03-07 03:27:40 -06:00
committed by GitHub
parent ee5b32f5d4
commit 170ee88457
7 changed files with 110 additions and 9 deletions

View File

@@ -44,6 +44,11 @@
<groupId>net.alchim31.maven</groupId>
<artifactId>scala-maven-plugin</artifactId>
<version>${scala-maven-plugin.version}</version>
<configuration>
<args>
<arg>-nobootcp</arg>
</args>
</configuration>
</plugin>
</plugins>
</pluginManagement>

View File

@@ -18,12 +18,11 @@
package org.apache.hudi.common.bloom.filter;
import org.apache.hudi.common.util.Base64CodecUtil;
import org.apache.hudi.exception.HoodieIndexException;
import org.apache.hadoop.util.bloom.Key;
import javax.xml.bind.DatatypeConverter;
import java.io.ByteArrayInputStream;
import java.io.ByteArrayOutputStream;
import java.io.DataInputStream;
@@ -66,7 +65,7 @@ public class HoodieDynamicBoundedBloomFilter implements BloomFilter {
*/
HoodieDynamicBoundedBloomFilter(String serString, BloomFilterTypeCode typeCode) {
// ignoring the type code for now, since we have just one version
byte[] bytes = DatatypeConverter.parseBase64Binary(serString);
byte[] bytes = Base64CodecUtil.decode(serString);
DataInputStream dis = new DataInputStream(new ByteArrayInputStream(bytes));
try {
internalDynamicBloomFilter = new InternalDynamicBloomFilter();
@@ -95,7 +94,7 @@ public class HoodieDynamicBoundedBloomFilter implements BloomFilter {
internalDynamicBloomFilter.write(dos);
byte[] bytes = baos.toByteArray();
dos.close();
return DatatypeConverter.printBase64Binary(bytes);
return Base64CodecUtil.encode(bytes);
} catch (IOException e) {
throw new HoodieIndexException("Could not serialize BloomFilter instance", e);
}

View File

@@ -18,12 +18,11 @@
package org.apache.hudi.common.bloom.filter;
import org.apache.hudi.common.util.Base64CodecUtil;
import org.apache.hudi.exception.HoodieIndexException;
import org.apache.hadoop.util.bloom.Key;
import javax.xml.bind.DatatypeConverter;
import java.io.ByteArrayInputStream;
import java.io.ByteArrayOutputStream;
import java.io.DataInput;
@@ -66,7 +65,7 @@ public class SimpleBloomFilter implements BloomFilter {
*/
public SimpleBloomFilter(String serString) {
this.filter = new org.apache.hadoop.util.bloom.BloomFilter();
byte[] bytes = DatatypeConverter.parseBase64Binary(serString);
byte[] bytes = Base64CodecUtil.decode(serString);
DataInputStream dis = new DataInputStream(new ByteArrayInputStream(bytes));
try {
this.filter.readFields(dis);
@@ -103,7 +102,7 @@ public class SimpleBloomFilter implements BloomFilter {
filter.write(dos);
byte[] bytes = baos.toByteArray();
dos.close();
return DatatypeConverter.printBase64Binary(bytes);
return Base64CodecUtil.encode(bytes);
} catch (IOException e) {
throw new HoodieIndexException("Could not serialize BloomFilter instance", e);
}

View File

@@ -0,0 +1,46 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hudi.common.util;
import java.nio.charset.StandardCharsets;
import java.util.Base64;
public final class Base64CodecUtil {
/**
* Decodes data from the input string into using the encoding scheme.
*
* @param serString
* @return A newly-allocated byte array containing the decoded bytes.
*/
public static byte[] decode(String serString) {
return Base64.getDecoder().decode(serString.getBytes(StandardCharsets.UTF_8));
}
/**
* Encodes all bytes from the specified byte array into String using StandardCharsets.UTF_8.
*
* @param data byte[] source data
* @return base64 encoded data
*/
public static String encode(byte[] data) {
return new String(Base64.getEncoder().encode(data), StandardCharsets.UTF_8);
}
}

View File

@@ -0,0 +1,47 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hudi.common.util;
import org.junit.Assert;
import org.junit.Test;
import java.nio.charset.StandardCharsets;
import java.util.UUID;
public class TestBase64CodecUtil {
@Test
public void testCodec() {
int times = 100;
UUID uuid = UUID.randomUUID();
for (int i = 0; i < times; i++) {
byte[] originalData = uuid.toString().getBytes(StandardCharsets.UTF_8);
String encodeData = Base64CodecUtil.encode(originalData);
byte[] decodeData = Base64CodecUtil.decode(encodeData);
Assert.assertArrayEquals(originalData, decodeData);
}
}
}

View File

@@ -42,6 +42,11 @@
<groupId>net.alchim31.maven</groupId>
<artifactId>scala-maven-plugin</artifactId>
<version>${scala-maven-plugin.version}</version>
<configuration>
<args>
<arg>-nobootcp</arg>
</args>
</configuration>
</plugin>
<plugin>
<groupId>org.apache.maven.plugins</groupId>

View File

@@ -94,7 +94,7 @@
<metrics.version>4.1.1</metrics.version>
<spark.version>2.4.4</spark.version>
<avro.version>1.8.2</avro.version>
<scala.version>2.11.8</scala.version>
<scala.version>2.11.12</scala.version>
<scala.binary.version>2.11</scala.binary.version>
<apache-rat-plugin.version>0.12</apache-rat-plugin.version>
<scala-maven-plugin.version>3.3.1</scala-maven-plugin.version>