public class SparkCompat
extends java.lang.Object
Constructor and Description |
---|
SparkCompat() |
Modifier and Type | Method and Description |
---|---|
static <K,InputT,AccumT,OutputT> |
extractOutput(org.apache.spark.api.java.JavaPairRDD<K,<any>> accumulatePerKey,
<any> sparkCombineFn)
Extracts the output for a given collection of WindowedAccumulators.
|
static <T> org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>> |
joinStreams(org.apache.spark.streaming.api.java.JavaStreamingContext streamingContext,
java.util.List<org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>>> dStreams)
Union of dStreams in the given StreamingContext.
|
public static <T> org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>> joinStreams(org.apache.spark.streaming.api.java.JavaStreamingContext streamingContext, java.util.List<org.apache.spark.streaming.api.java.JavaDStream<org.apache.beam.sdk.util.WindowedValue<T>>> dStreams)
This is required because the API to join (union) DStreams is different among Spark versions. See https://issues.apache.org/jira/browse/SPARK-25737
public static <K,InputT,AccumT,OutputT> org.apache.spark.api.java.JavaPairRDD<K,org.apache.beam.sdk.util.WindowedValue<OutputT>> extractOutput(org.apache.spark.api.java.JavaPairRDD<K,<any>> accumulatePerKey, <any> sparkCombineFn)
This is required because the API of JavaPairRDD.flatMapValues is different among Spark versions. See https://issues.apache.org/jira/browse/SPARK-19287