public class SparkCompat
extends java.lang.Object
| Modifier and Type | Method and Description | 
|---|---|
| static org.apache.spark.scheduler.SparkListenerApplicationStart | buildSparkListenerApplicationStart(org.apache.spark.api.java.JavaSparkContext jsc,
                                  SparkPipelineOptions options,
                                  long time,
                                  PipelineResult result) | 
| static <K,InputT,AccumT,OutputT> | extractOutput(org.apache.spark.api.java.JavaPairRDD<K,<any>> accumulatePerKey,
             <any> sparkCombineFn)Extracts the output for a given collection of WindowedAccumulators. | 
| static <T> org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>> | joinStreams(org.apache.spark.streaming.api.java.JavaStreamingContext streamingContext,
           java.util.List<org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>>> dStreams)Union of dStreams in the given StreamingContext. | 
public static <T> org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>> joinStreams(org.apache.spark.streaming.api.java.JavaStreamingContext streamingContext,
                                                                                                                         java.util.List<org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>>> dStreams)
This is required because the API to join (union) DStreams is different among Spark versions. See https://issues.apache.org/jira/browse/SPARK-25737
public static <K,InputT,AccumT,OutputT> org.apache.spark.api.java.JavaPairRDD<K,org.apache.beam.sdk.util.WindowedValue<OutputT>> extractOutput(org.apache.spark.api.java.JavaPairRDD<K,<any>> accumulatePerKey,
                                                                                                                                               <any> sparkCombineFn)
This is required because the API of JavaPairRDD.flatMapValues is different among Spark versions. See https://issues.apache.org/jira/browse/SPARK-19287
public static org.apache.spark.scheduler.SparkListenerApplicationStart buildSparkListenerApplicationStart(org.apache.spark.api.java.JavaSparkContext jsc,
                                                                                                          SparkPipelineOptions options,
                                                                                                          long time,
                                                                                                          PipelineResult result)