[HUDI-1996] Adding functionality to allow the providing of basic auth creds for confluent cloud schema registry (#3097)
* adding support for basic auth with confluent cloud schema registry
This commit is contained in:
@@ -77,7 +77,7 @@ public class TestKeyRangeLookupTree {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Tests for many duplicte entries in the tree.
|
* Tests for many duplicate entries in the tree.
|
||||||
*/
|
*/
|
||||||
@Test
|
@Test
|
||||||
public void testFileGroupLookUpManyDulicateEntries() {
|
public void testFileGroupLookUpManyDulicateEntries() {
|
||||||
|
|||||||
@@ -261,22 +261,22 @@
|
|||||||
<dependency>
|
<dependency>
|
||||||
<groupId>io.confluent</groupId>
|
<groupId>io.confluent</groupId>
|
||||||
<artifactId>kafka-avro-serializer</artifactId>
|
<artifactId>kafka-avro-serializer</artifactId>
|
||||||
<version>3.0.0</version>
|
<version>${confluent.version}</version>
|
||||||
</dependency>
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>io.confluent</groupId>
|
<groupId>io.confluent</groupId>
|
||||||
<artifactId>common-config</artifactId>
|
<artifactId>common-config</artifactId>
|
||||||
<version>3.0.0</version>
|
<version>${confluent.version}</version>
|
||||||
</dependency>
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>io.confluent</groupId>
|
<groupId>io.confluent</groupId>
|
||||||
<artifactId>common-utils</artifactId>
|
<artifactId>common-utils</artifactId>
|
||||||
<version>3.0.0</version>
|
<version>${confluent.version}</version>
|
||||||
</dependency>
|
</dependency>
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>io.confluent</groupId>
|
<groupId>io.confluent</groupId>
|
||||||
<artifactId>kafka-schema-registry-client</artifactId>
|
<artifactId>kafka-schema-registry-client</artifactId>
|
||||||
<version>3.0.0</version>
|
<version>${confluent.version}</version>
|
||||||
</dependency>
|
</dependency>
|
||||||
|
|
||||||
<dependency>
|
<dependency>
|
||||||
|
|||||||
@@ -28,8 +28,14 @@ import org.apache.avro.Schema;
|
|||||||
import org.apache.spark.api.java.JavaSparkContext;
|
import org.apache.spark.api.java.JavaSparkContext;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
|
import java.io.InputStream;
|
||||||
|
import java.net.HttpURLConnection;
|
||||||
import java.net.URL;
|
import java.net.URL;
|
||||||
|
import java.nio.charset.StandardCharsets;
|
||||||
|
import java.util.Base64;
|
||||||
import java.util.Collections;
|
import java.util.Collections;
|
||||||
|
import java.util.regex.Matcher;
|
||||||
|
import java.util.regex.Pattern;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Obtains latest schema from the Confluent/Kafka schema-registry.
|
* Obtains latest schema from the Confluent/Kafka schema-registry.
|
||||||
@@ -48,19 +54,49 @@ public class SchemaRegistryProvider extends SchemaProvider {
|
|||||||
"hoodie.deltastreamer.schemaprovider.registry.targetUrl";
|
"hoodie.deltastreamer.schemaprovider.registry.targetUrl";
|
||||||
}
|
}
|
||||||
|
|
||||||
private static String fetchSchemaFromRegistry(String registryUrl) throws IOException {
|
/**
|
||||||
URL registry = new URL(registryUrl);
|
* The method takes the provided url {@code registryUrl} and gets the schema from the schema registry using that url.
|
||||||
|
* If the caller provides userInfo credentials in the url (e.g "https://foo:bar@schemaregistry.org") then the credentials
|
||||||
|
* are extracted the url using the Matcher and the extracted credentials are set on the request as an Authorization
|
||||||
|
* header.
|
||||||
|
* @param registryUrl
|
||||||
|
* @return the Schema in String form.
|
||||||
|
* @throws IOException
|
||||||
|
*/
|
||||||
|
public String fetchSchemaFromRegistry(String registryUrl) throws IOException {
|
||||||
|
URL registry;
|
||||||
|
HttpURLConnection connection;
|
||||||
|
Matcher matcher = Pattern.compile("://(.*?)@").matcher(registryUrl);
|
||||||
|
if (matcher.find()) {
|
||||||
|
String creds = matcher.group(1);
|
||||||
|
String urlWithoutCreds = registryUrl.replace(creds + "@", "");
|
||||||
|
registry = new URL(urlWithoutCreds);
|
||||||
|
connection = (HttpURLConnection) registry.openConnection();
|
||||||
|
setAuthorizationHeader(matcher.group(1), connection);
|
||||||
|
} else {
|
||||||
|
registry = new URL(registryUrl);
|
||||||
|
connection = (HttpURLConnection) registry.openConnection();
|
||||||
|
}
|
||||||
ObjectMapper mapper = new ObjectMapper();
|
ObjectMapper mapper = new ObjectMapper();
|
||||||
JsonNode node = mapper.readTree(registry.openStream());
|
JsonNode node = mapper.readTree(getStream(connection));
|
||||||
return node.get("schema").asText();
|
return node.get("schema").asText();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
protected void setAuthorizationHeader(String creds, HttpURLConnection connection) {
|
||||||
|
String encodedAuth = Base64.getEncoder().encodeToString(creds.getBytes(StandardCharsets.UTF_8));
|
||||||
|
connection.setRequestProperty("Authorization", "Basic " + encodedAuth);
|
||||||
|
}
|
||||||
|
|
||||||
|
protected InputStream getStream(HttpURLConnection connection) throws IOException {
|
||||||
|
return connection.getInputStream();
|
||||||
|
}
|
||||||
|
|
||||||
public SchemaRegistryProvider(TypedProperties props, JavaSparkContext jssc) {
|
public SchemaRegistryProvider(TypedProperties props, JavaSparkContext jssc) {
|
||||||
super(props, jssc);
|
super(props, jssc);
|
||||||
DataSourceUtils.checkRequiredProperties(props, Collections.singletonList(Config.SRC_SCHEMA_REGISTRY_URL_PROP));
|
DataSourceUtils.checkRequiredProperties(props, Collections.singletonList(Config.SRC_SCHEMA_REGISTRY_URL_PROP));
|
||||||
}
|
}
|
||||||
|
|
||||||
private static Schema getSchema(String registryUrl) throws IOException {
|
private Schema getSchema(String registryUrl) throws IOException {
|
||||||
return new Schema.Parser().parse(fetchSchemaFromRegistry(registryUrl));
|
return new Schema.Parser().parse(fetchSchemaFromRegistry(registryUrl));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -94,7 +94,7 @@ public class TestKafkaAvroSchemaDeserializer extends UtilitiesTestBase {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Tests {@link KafkaAvroSchemaDeserializer#deserialize(boolean, String, Boolean, byte[], Schema)}.
|
* Tests {@link KafkaAvroSchemaDeserializer#deserialize(Boolean, String, Boolean, byte[], Schema)}.
|
||||||
*/
|
*/
|
||||||
@Test
|
@Test
|
||||||
public void testKafkaAvroSchemaDeserializer() {
|
public void testKafkaAvroSchemaDeserializer() {
|
||||||
|
|||||||
@@ -0,0 +1,111 @@
|
|||||||
|
/*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one
|
||||||
|
* or more contributor license agreements. See the NOTICE file
|
||||||
|
* distributed with this work for additional information
|
||||||
|
* regarding copyright ownership. The ASF licenses this file
|
||||||
|
* to you under the Apache License, Version 2.0 (the
|
||||||
|
* "License"); you may not use this file except in compliance
|
||||||
|
* with the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package org.apache.hudi.utilities.schema;
|
||||||
|
|
||||||
|
import com.fasterxml.jackson.databind.JsonNode;
|
||||||
|
import com.fasterxml.jackson.databind.ObjectMapper;
|
||||||
|
import org.apache.avro.Schema;
|
||||||
|
import org.apache.hudi.common.config.TypedProperties;
|
||||||
|
import org.junit.jupiter.api.Test;
|
||||||
|
import org.mockito.Mockito;
|
||||||
|
|
||||||
|
import java.io.ByteArrayInputStream;
|
||||||
|
import java.io.IOException;
|
||||||
|
import java.io.InputStream;
|
||||||
|
import java.net.HttpURLConnection;
|
||||||
|
import java.nio.charset.StandardCharsets;
|
||||||
|
|
||||||
|
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||||
|
import static org.junit.jupiter.api.Assertions.assertNotNull;
|
||||||
|
import static org.mockito.ArgumentMatchers.eq;
|
||||||
|
import static org.mockito.Mockito.times;
|
||||||
|
import static org.mockito.Mockito.verify;
|
||||||
|
|
||||||
|
class TestSchemaRegistryProvider {
|
||||||
|
|
||||||
|
private final String basicAuth = "foo:bar";
|
||||||
|
|
||||||
|
private final String json = "{\"schema\":\"{\\\"type\\\": \\\"record\\\", \\\"namespace\\\": \\\"example\\\", "
|
||||||
|
+ "\\\"name\\\": \\\"FullName\\\",\\\"fields\\\": [{ \\\"name\\\": \\\"first\\\", \\\"type\\\": "
|
||||||
|
+ "\\\"string\\\" }]}\"}";
|
||||||
|
|
||||||
|
private TypedProperties getProps() {
|
||||||
|
return new TypedProperties() {{
|
||||||
|
put("hoodie.deltastreamer.schemaprovider.registry.baseUrl", "http://" + basicAuth + "@localhost");
|
||||||
|
put("hoodie.deltastreamer.schemaprovider.registry.urlSuffix", "-value");
|
||||||
|
put("hoodie.deltastreamer.schemaprovider.registry.url", "http://foo:bar@localhost");
|
||||||
|
put("hoodie.deltastreamer.source.kafka.topic", "foo");
|
||||||
|
}};
|
||||||
|
}
|
||||||
|
|
||||||
|
private Schema getExpectedSchema(String response) throws IOException {
|
||||||
|
ObjectMapper mapper = new ObjectMapper();
|
||||||
|
JsonNode node = mapper.readTree(new ByteArrayInputStream(response.getBytes(StandardCharsets.UTF_8)));
|
||||||
|
return (new Schema.Parser()).parse(node.get("schema").asText());
|
||||||
|
}
|
||||||
|
|
||||||
|
private SchemaRegistryProvider getUnderTest(TypedProperties props) throws IOException {
|
||||||
|
InputStream is = new ByteArrayInputStream(json.getBytes(StandardCharsets.UTF_8));
|
||||||
|
SchemaRegistryProvider spyUnderTest = Mockito.spy(new SchemaRegistryProvider(props, null));
|
||||||
|
Mockito.doReturn(is).when(spyUnderTest).getStream(Mockito.any());
|
||||||
|
return spyUnderTest;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testGetSourceSchemaShouldRequestSchemaWithCreds() throws IOException {
|
||||||
|
SchemaRegistryProvider spyUnderTest = getUnderTest(getProps());
|
||||||
|
Schema actual = spyUnderTest.getSourceSchema();
|
||||||
|
assertNotNull(actual);
|
||||||
|
assertEquals(actual, getExpectedSchema(json));
|
||||||
|
verify(spyUnderTest, times(1)).setAuthorizationHeader(eq(basicAuth),
|
||||||
|
Mockito.any(HttpURLConnection.class));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testGetTargetSchemaShouldRequestSchemaWithCreds() throws IOException {
|
||||||
|
SchemaRegistryProvider spyUnderTest = getUnderTest(getProps());
|
||||||
|
Schema actual = spyUnderTest.getTargetSchema();
|
||||||
|
assertNotNull(actual);
|
||||||
|
assertEquals(actual, getExpectedSchema(json));
|
||||||
|
verify(spyUnderTest, times(1)).setAuthorizationHeader(eq(basicAuth),
|
||||||
|
Mockito.any(HttpURLConnection.class));
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testGetSourceSchemaShouldRequestSchemaWithoutCreds() throws IOException {
|
||||||
|
TypedProperties props = getProps();
|
||||||
|
props.put("hoodie.deltastreamer.schemaprovider.registry.url", "http://localhost");
|
||||||
|
SchemaRegistryProvider spyUnderTest = getUnderTest(props);
|
||||||
|
Schema actual = spyUnderTest.getSourceSchema();
|
||||||
|
assertNotNull(actual);
|
||||||
|
assertEquals(actual, getExpectedSchema(json));
|
||||||
|
verify(spyUnderTest, times(0)).setAuthorizationHeader(Mockito.any(), Mockito.any());
|
||||||
|
}
|
||||||
|
|
||||||
|
@Test
|
||||||
|
public void testGetTargetSchemaShouldRequestSchemaWithoutCreds() throws IOException {
|
||||||
|
TypedProperties props = getProps();
|
||||||
|
props.put("hoodie.deltastreamer.schemaprovider.registry.url", "http://localhost");
|
||||||
|
SchemaRegistryProvider spyUnderTest = getUnderTest(props);
|
||||||
|
Schema actual = spyUnderTest.getTargetSchema();
|
||||||
|
assertNotNull(actual);
|
||||||
|
assertEquals(actual, getExpectedSchema(json));
|
||||||
|
verify(spyUnderTest, times(0)).setAuthorizationHeader(Mockito.any(), Mockito.any());
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -615,25 +615,25 @@
|
|||||||
<dependency>
|
<dependency>
|
||||||
<groupId>io.confluent</groupId>
|
<groupId>io.confluent</groupId>
|
||||||
<artifactId>kafka-avro-serializer</artifactId>
|
<artifactId>kafka-avro-serializer</artifactId>
|
||||||
<version>3.0.0</version>
|
<version>${confluent.version}</version>
|
||||||
</dependency>
|
</dependency>
|
||||||
|
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>io.confluent</groupId>
|
<groupId>io.confluent</groupId>
|
||||||
<artifactId>common-config</artifactId>
|
<artifactId>common-config</artifactId>
|
||||||
<version>3.0.0</version>
|
<version>${confluent.version}</version>
|
||||||
</dependency>
|
</dependency>
|
||||||
|
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>io.confluent</groupId>
|
<groupId>io.confluent</groupId>
|
||||||
<artifactId>common-utils</artifactId>
|
<artifactId>common-utils</artifactId>
|
||||||
<version>3.0.0</version>
|
<version>${confluent.version}</version>
|
||||||
</dependency>
|
</dependency>
|
||||||
|
|
||||||
<dependency>
|
<dependency>
|
||||||
<groupId>io.confluent</groupId>
|
<groupId>io.confluent</groupId>
|
||||||
<artifactId>kafka-schema-registry-client</artifactId>
|
<artifactId>kafka-schema-registry-client</artifactId>
|
||||||
<version>3.0.0</version>
|
<version>${confluent.version}</version>
|
||||||
</dependency>
|
</dependency>
|
||||||
|
|
||||||
<dependency>
|
<dependency>
|
||||||
|
|||||||
@@ -93,6 +93,7 @@
|
|||||||
<include>com.twitter:bijection-avro_${scala.binary.version}</include>
|
<include>com.twitter:bijection-avro_${scala.binary.version}</include>
|
||||||
<include>com.twitter:bijection-core_${scala.binary.version}</include>
|
<include>com.twitter:bijection-core_${scala.binary.version}</include>
|
||||||
<include>io.confluent:kafka-avro-serializer</include>
|
<include>io.confluent:kafka-avro-serializer</include>
|
||||||
|
<include>io.confluent:kafka-schema-serializer</include>
|
||||||
<include>io.confluent:common-config</include>
|
<include>io.confluent:common-config</include>
|
||||||
<include>io.confluent:common-utils</include>
|
<include>io.confluent:common-utils</include>
|
||||||
<include>io.confluent:kafka-schema-registry-client</include>
|
<include>io.confluent:kafka-schema-registry-client</include>
|
||||||
|
|||||||
1
pom.xml
1
pom.xml
@@ -89,6 +89,7 @@
|
|||||||
<fasterxml.jackson.dataformat.yaml.version>2.7.4</fasterxml.jackson.dataformat.yaml.version>
|
<fasterxml.jackson.dataformat.yaml.version>2.7.4</fasterxml.jackson.dataformat.yaml.version>
|
||||||
<fasterxml.spark3.version>2.10.0</fasterxml.spark3.version>
|
<fasterxml.spark3.version>2.10.0</fasterxml.spark3.version>
|
||||||
<kafka.version>2.0.0</kafka.version>
|
<kafka.version>2.0.0</kafka.version>
|
||||||
|
<confluent.version>5.3.4</confluent.version>
|
||||||
<glassfish.version>2.17</glassfish.version>
|
<glassfish.version>2.17</glassfish.version>
|
||||||
<parquet.version>1.10.1</parquet.version>
|
<parquet.version>1.10.1</parquet.version>
|
||||||
<junit.jupiter.version>5.7.0-M1</junit.jupiter.version>
|
<junit.jupiter.version>5.7.0-M1</junit.jupiter.version>
|
||||||
|
|||||||
Reference in New Issue
Block a user