[HUDI-3569] Introduce ChainedJsonKafkaSourePostProcessor to support setting multi processors at once (#4969)
This commit is contained in:
@@ -43,6 +43,7 @@ import org.apache.hudi.exception.HoodieIOException;
|
||||
import org.apache.hudi.index.HoodieIndex;
|
||||
import org.apache.hudi.utilities.checkpointing.InitialCheckPointProvider;
|
||||
import org.apache.hudi.utilities.deltastreamer.HoodieDeltaStreamerMetrics;
|
||||
import org.apache.hudi.utilities.exception.HoodieSourcePostProcessException;
|
||||
import org.apache.hudi.utilities.exception.HoodieSchemaPostProcessException;
|
||||
import org.apache.hudi.utilities.schema.ChainedSchemaPostProcessor;
|
||||
import org.apache.hudi.utilities.schema.DelegatingSchemaProvider;
|
||||
@@ -53,6 +54,7 @@ import org.apache.hudi.utilities.schema.SchemaProvider;
|
||||
import org.apache.hudi.utilities.schema.SchemaProviderWithPostProcessor;
|
||||
import org.apache.hudi.utilities.schema.SparkAvroPostProcessor;
|
||||
import org.apache.hudi.utilities.sources.Source;
|
||||
import org.apache.hudi.utilities.sources.processor.ChainedJsonKafkaSourcePostProcessor;
|
||||
import org.apache.hudi.utilities.sources.processor.JsonKafkaSourcePostProcessor;
|
||||
import org.apache.hudi.utilities.transform.ChainedTransformer;
|
||||
import org.apache.hudi.utilities.transform.Transformer;
|
||||
@@ -125,12 +127,19 @@ public class UtilHelpers {
|
||||
}
|
||||
}
|
||||
|
||||
public static JsonKafkaSourcePostProcessor createJsonKafkaSourcePostProcessor(String postProcessorClassName, TypedProperties props) throws IOException {
|
||||
public static JsonKafkaSourcePostProcessor createJsonKafkaSourcePostProcessor(String postProcessorClassNames, TypedProperties props) throws IOException {
|
||||
if (StringUtils.isNullOrEmpty(postProcessorClassNames)) {
|
||||
return null;
|
||||
}
|
||||
|
||||
try {
|
||||
return StringUtils.isNullOrEmpty(postProcessorClassName) ? null
|
||||
: (JsonKafkaSourcePostProcessor) ReflectionUtils.loadClass(postProcessorClassName, props);
|
||||
List<JsonKafkaSourcePostProcessor> processors = new ArrayList<>();
|
||||
for (String className : (postProcessorClassNames.split(","))) {
|
||||
processors.add((JsonKafkaSourcePostProcessor) ReflectionUtils.loadClass(className, props));
|
||||
}
|
||||
return new ChainedJsonKafkaSourcePostProcessor(processors, props);
|
||||
} catch (Throwable e) {
|
||||
throw new IOException("Could not load json kafka source post processor class " + postProcessorClassName, e);
|
||||
throw new HoodieSourcePostProcessException("Could not load postProcessorClassNames class(es) " + postProcessorClassNames, e);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -0,0 +1,47 @@
|
||||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.hudi.utilities.sources.processor;
|
||||
|
||||
import org.apache.hudi.common.config.TypedProperties;
|
||||
|
||||
import org.apache.spark.api.java.JavaRDD;
|
||||
|
||||
import java.util.List;
|
||||
|
||||
/**
|
||||
* A {@link JsonKafkaSourcePostProcessor} to chain other {@link JsonKafkaSourcePostProcessor}s and apply sequentially.
|
||||
*/
|
||||
public class ChainedJsonKafkaSourcePostProcessor extends JsonKafkaSourcePostProcessor {
|
||||
|
||||
private final List<JsonKafkaSourcePostProcessor> processors;
|
||||
|
||||
public ChainedJsonKafkaSourcePostProcessor(List<JsonKafkaSourcePostProcessor> processors, TypedProperties props) {
|
||||
super(props);
|
||||
this.processors = processors;
|
||||
}
|
||||
|
||||
@Override
|
||||
public JavaRDD<String> process(JavaRDD<String> inputJsonRecords) {
|
||||
JavaRDD<String> targetRDD = inputJsonRecords;
|
||||
for (JsonKafkaSourcePostProcessor processor : processors) {
|
||||
targetRDD = processor.process(targetRDD);
|
||||
}
|
||||
return targetRDD;
|
||||
}
|
||||
}
|
||||
@@ -30,6 +30,8 @@ import org.apache.spark.api.java.JavaRDD;
|
||||
import org.junit.jupiter.api.Assertions;
|
||||
import org.junit.jupiter.api.Test;
|
||||
|
||||
import java.util.Objects;
|
||||
|
||||
import static org.apache.hudi.utilities.sources.helpers.KafkaOffsetGen.Config.JSON_KAFKA_PROCESSOR_CLASS_OPT;
|
||||
import static org.apache.hudi.utilities.testutils.UtilitiesTestBase.Helpers.jsonifyRecords;
|
||||
import static org.junit.jupiter.api.Assertions.assertEquals;
|
||||
@@ -96,6 +98,28 @@ public class TestJsonKafkaSourcePostProcessor extends TestJsonKafkaSource {
|
||||
() -> kafkaSource.fetchNewDataInAvroFormat(Option.empty(), 900));
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testChainedJsonKafkaSourcePostProcessor() {
|
||||
// topic setup.
|
||||
final String topic = TEST_TOPIC_PREFIX + "testChainedJsonKafkaSourcePostProcessor";
|
||||
testUtils.createTopic(topic, 2);
|
||||
|
||||
HoodieTestDataGenerator dataGenerator = new HoodieTestDataGenerator();
|
||||
TypedProperties props = createPropsForJsonSource(topic, null, "earliest");
|
||||
|
||||
// processor class name setup
|
||||
props.setProperty(JSON_KAFKA_PROCESSOR_CLASS_OPT.key(), SampleJsonKafkaSourcePostProcessor.class.getName()
|
||||
+ "," + DummyJsonKafkaSourcePostProcessor.class.getName());
|
||||
|
||||
Source jsonSource = new JsonKafkaSource(props, jsc(), spark(), schemaProvider, metrics);
|
||||
SourceFormatAdapter kafkaSource = new SourceFormatAdapter(jsonSource);
|
||||
|
||||
testUtils.sendMessages(topic, jsonifyRecords(dataGenerator.generateInserts("000", 1000)));
|
||||
InputBatch<JavaRDD<GenericRecord>> fetch1 = kafkaSource.fetchNewDataInAvroFormat(Option.empty(), 900);
|
||||
|
||||
assertEquals(0, fetch1.getBatch().get().count());
|
||||
}
|
||||
|
||||
/**
|
||||
* JsonKafkaSourcePostProcessor that return a sub RDD of the incoming data which get the data from incoming data using
|
||||
* {org.apache.spark.api.java.JavaRDD#sample(boolean, double, long)} method.
|
||||
@@ -112,4 +136,16 @@ public class TestJsonKafkaSourcePostProcessor extends TestJsonKafkaSource {
|
||||
}
|
||||
}
|
||||
|
||||
public static class DummyJsonKafkaSourcePostProcessor extends JsonKafkaSourcePostProcessor {
|
||||
public DummyJsonKafkaSourcePostProcessor(TypedProperties props) {
|
||||
super(props);
|
||||
}
|
||||
|
||||
@Override
|
||||
public JavaRDD<String> process(JavaRDD<String> inputJsonRecords) {
|
||||
// return empty RDD
|
||||
return inputJsonRecords.map(x -> "").filter(x -> !Objects.equals(x, ""));
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user