@Internal public class SparkCompat extends java.lang.Object
| Constructor and Description | 
|---|
SparkCompat()  | 
| Modifier and Type | Method and Description | 
|---|---|
static <K,InputT,AccumT,OutputT> | 
extractOutput(org.apache.spark.api.java.JavaPairRDD<K,<any>> accumulatePerKey,
             <any> sparkCombineFn)
Extracts the output for a given collection of WindowedAccumulators. 
 | 
static <T> org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>> | 
joinStreams(org.apache.spark.streaming.api.java.JavaStreamingContext streamingContext,
           java.util.List<org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>>> dStreams)
Union of dStreams in the given StreamingContext. 
 | 
public static <T> org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>> joinStreams(org.apache.spark.streaming.api.java.JavaStreamingContext streamingContext,
                                                                                                                         java.util.List<org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>>> dStreams)
This is required because the API to join (union) DStreams is different among Spark versions. See https://issues.apache.org/jira/browse/SPARK-25737
public static <K,InputT,AccumT,OutputT> org.apache.spark.api.java.JavaPairRDD<K,org.apache.beam.sdk.util.WindowedValue<OutputT>> extractOutput(org.apache.spark.api.java.JavaPairRDD<K,<any>> accumulatePerKey,
                                                                                                                                               <any> sparkCombineFn)
This is required because the API of JavaPairRDD.flatMapValues is different among Spark versions. See https://issues.apache.org/jira/browse/SPARK-19287