Class SparkInputRecordProcessor
java.lang.Object
com.linkedin.venice.hadoop.task.datawriter.AbstractDataWriterTask
com.linkedin.venice.hadoop.task.datawriter.AbstractInputRecordProcessor<ByteBuffer,ByteBuffer>
com.linkedin.venice.spark.datawriter.recordprocessor.SparkInputRecordProcessor
- All Implemented Interfaces:
Closeable,AutoCloseable
An implementation of
AbstractInputRecordProcessor for Spark that processes input records from the dataframe
and emits an Iterator of Row with DEFAULT_SCHEMA as the schema.-
Field Summary
Fields inherited from class com.linkedin.venice.hadoop.task.datawriter.AbstractInputRecordProcessor
EMPTY_BYTES, veniceRecordReaderFields inherited from class com.linkedin.venice.hadoop.task.datawriter.AbstractDataWriterTask
TASK_ID_NOT_SET -
Constructor Summary
ConstructorsConstructorDescriptionSparkInputRecordProcessor(Properties jobProperties, DataWriterAccumulators accumulators) -
Method Summary
Modifier and TypeMethodDescriptionprotected AbstractVeniceRecordReader<ByteBuffer,ByteBuffer> getRecordReader(VeniceProperties props) A method for child classes to setupAbstractInputRecordProcessor.veniceRecordReader.Iterator<org.apache.spark.sql.Row>processRecord(org.apache.spark.sql.Row record) Methods inherited from class com.linkedin.venice.hadoop.task.datawriter.AbstractInputRecordProcessor
close, configureTask, process, processRecord, readDictionaryFromKafkaMethods inherited from class com.linkedin.venice.hadoop.task.datawriter.AbstractDataWriterTask
configure, getEngineTaskConfigProvider, getPartitionCount, getTaskId, isChunkingEnabled, isRmdChunkingEnabled, setChunkingEnabled
-
Constructor Details
-
SparkInputRecordProcessor
-
-
Method Details
-
processRecord
-
getRecordReader
Description copied from class:AbstractInputRecordProcessorA method for child classes to setupAbstractInputRecordProcessor.veniceRecordReader.- Specified by:
getRecordReaderin classAbstractInputRecordProcessor<ByteBuffer,ByteBuffer>
-