public class SparkCompat
extends java.lang.Object
Modifier and Type | Method and Description |
---|---|
static org.apache.spark.scheduler.SparkListenerApplicationStart |
buildSparkListenerApplicationStart(org.apache.spark.api.java.JavaSparkContext jsc,
SparkPipelineOptions options,
long time,
PipelineResult result) |
static <K,InputT,AccumT,OutputT> |
extractOutput(org.apache.spark.api.java.JavaPairRDD<K,<any>> accumulatePerKey,
<any> sparkCombineFn)
Extracts the output for a given collection of WindowedAccumulators.
|
static <T> org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>> |
joinStreams(org.apache.spark.streaming.api.java.JavaStreamingContext streamingContext,
java.util.List<org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>>> dStreams)
Union of dStreams in the given StreamingContext.
|
public static <T> org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>> joinStreams(org.apache.spark.streaming.api.java.JavaStreamingContext streamingContext, java.util.List<org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>>> dStreams)
This is required because the API to join (union) DStreams is different among Spark versions. See https://issues.apache.org/jira/browse/SPARK-25737
public static <K,InputT,AccumT,OutputT> org.apache.spark.api.java.JavaPairRDD<K,org.apache.beam.sdk.util.WindowedValue<OutputT>> extractOutput(org.apache.spark.api.java.JavaPairRDD<K,<any>> accumulatePerKey, <any> sparkCombineFn)
This is required because the API of JavaPairRDD.flatMapValues is different among Spark versions. See https://issues.apache.org/jira/browse/SPARK-19287
public static org.apache.spark.scheduler.SparkListenerApplicationStart buildSparkListenerApplicationStart(org.apache.spark.api.java.JavaSparkContext jsc, SparkPipelineOptions options, long time, PipelineResult result)