IN - The type of input record to the SparkSink.@Beta public abstract class SparkSink<IN> extends BatchConfigurable<SparkPluginContext> implements Serializable
run(co.cask.cdap.etl.api.batch.SparkExecutionPluginContext, org.apache.spark.api.java.JavaRDD<IN>) method is called inside the Batch Run while BatchConfigurable.prepareRun(T) and
BatchConfigurable.onRunFinish(boolean, T) methods are called on the client side, which launches the Batch run, before the
Batch run starts and after it finishes respectively.| Modifier and Type | Field and Description |
|---|---|
static String |
PLUGIN_TYPE |
| Constructor and Description |
|---|
SparkSink() |
| Modifier and Type | Method and Description |
|---|---|
abstract void |
run(SparkExecutionPluginContext context,
org.apache.spark.api.java.JavaRDD<IN> input)
User Spark job which will be executed and is responsible for persisting any data.
|
configurePipeline, onRunFinish, prepareRunpublic static final String PLUGIN_TYPE
public abstract void run(SparkExecutionPluginContext context, org.apache.spark.api.java.JavaRDD<IN> input) throws Exception
context - SparkExecutionPluginContext for this jobinput - the input from previous stages of the Batch run.ExceptionCopyright © 2017 Cask Data, Inc. Licensed under the Apache License, Version 2.0.