- absolute(String, String...) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData.Path
-
Construct a path from an absolute component path hierarchy.
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlArithmeticExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlCaseExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlCastExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlExpression
-
assertion to make sure the input and output are supported in this expression.
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlInputRefExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlPrimitive
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlReinterpretExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlUdfExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlWindowEndExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlWindowExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlWindowStartExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlCompareExpression
-
Compare operation must have 2 operands.
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlIsNotNullExpression
-
only one operand is required.
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlIsNullExpression
-
only one operand is required.
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date.BeamSqlCurrentDateExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date.BeamSqlCurrentTimeExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date.BeamSqlCurrentTimestampExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date.BeamSqlDateCeilExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date.BeamSqlDateFloorExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date.BeamSqlExtractExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.logical.BeamSqlLogicalExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.logical.BeamSqlNotExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlMathBinaryExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlMathUnaryExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlPiExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlRandExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlRandIntegerExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlConcatExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlOverlayExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlPositionExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlStringUnaryExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlSubstringExpression
-
- accept() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlTrimExpression
-
- AccumulatingCombineFn() - Constructor for class org.apache.beam.sdk.transforms.Combine.AccumulatingCombineFn
-
- accumulatingFiredPanes() - Method in class org.apache.beam.sdk.transforms.windowing.Window
-
Returns a new Window
PTransform
that uses the registered WindowFn and
Triggering behavior, and that accumulates elements in a pane after they are triggered.
- ACCUMULATOR_NAME - Static variable in class org.apache.beam.runners.flink.metrics.FlinkMetricContainer
-
- AccumulatorCheckpointingSparkListener() - Constructor for class org.apache.beam.runners.spark.aggregators.AggregatorsAccumulator.AccumulatorCheckpointingSparkListener
-
- AccumulatorCheckpointingSparkListener() - Constructor for class org.apache.beam.runners.spark.metrics.MetricsAccumulator.AccumulatorCheckpointingSparkListener
-
- add(MetricsContainerStepMap) - Method in class org.apache.beam.runners.flink.metrics.MetricsAccumulator
-
- add(int, GlobalWatermarkHolder.SparkWatermarks) - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
-
- add(KV<byte[], byte[]>) - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter
-
- add(InputT) - Method in interface org.apache.beam.sdk.state.GroupingState
-
Add a value to the buffer.
- add(Iterable<TimestampedValue<T>>) - Static method in class org.apache.beam.sdk.testing.TestStream.ElementEvent
-
For internal use only: no backwards compatibility guarantees.
- add(Long) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn.LargestUnique
-
Adds a value to the heap, returning whether the value is (large enough
to be) in the heap.
- add(DisplayData.ItemSpec<?>) - Method in interface org.apache.beam.sdk.transforms.display.DisplayData.Builder
-
Register the given display item.
- addAccum(AccumT) - Method in interface org.apache.beam.sdk.state.CombiningState
-
Add an accumulator to this state cell.
- addAccumulator(NamedAggregators, NamedAggregators) - Method in class org.apache.beam.runners.spark.aggregators.AggAccumParam
-
- addAll(Map<Integer, Queue<GlobalWatermarkHolder.SparkWatermarks>>) - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
-
- addBoolean(Map<String, Object>, String, boolean) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- addCollectionToSingletonOutput(PCollection<?>, PCollectionView<?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
-
Adds an output to this CollectionToSingleton
Dataflow step, consuming the specified
input PValue
and producing the specified output PValue
.
- addDictionary(Map<String, Object>, String, Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- addDouble(Map<String, Object>, String, Double) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- addElements(T, T...) - Method in class org.apache.beam.sdk.testing.TestStream.Builder
-
Adds the specified elements to the source with timestamp equal to the current watermark.
- addElements(TimestampedValue<T>, TimestampedValue<T>...) - Method in class org.apache.beam.sdk.testing.TestStream.Builder
-
Adds the specified elements to the source with the provided timestamps.
- addEncodingInput(Coder<?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
-
Sets the encoding for this Dataflow step.
- addIfAbsent(T) - Method in interface org.apache.beam.sdk.state.SetState
-
Ensures a value is a member of the set, returning true
if it was added and false
otherwise.
- addIfNotDefault(DisplayData.ItemSpec<T>, T) - Method in interface org.apache.beam.sdk.transforms.display.DisplayData.Builder
-
Register the given display item if the value is different than the specified default.
- addIfNotNull(DisplayData.ItemSpec<?>) - Method in interface org.apache.beam.sdk.transforms.display.DisplayData.Builder
-
Register the given display item if the value is not null.
- addInPlace(NamedAggregators, NamedAggregators) - Method in class org.apache.beam.runners.spark.aggregators.AggAccumParam
-
- addInput(String, Boolean) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
-
Adds an input with the given name and value to this Dataflow step.
- addInput(String, String) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
-
Adds an input with the given name and value to this Dataflow step.
- addInput(String, Long) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
-
Adds an input with the given name and value to this Dataflow step.
- addInput(String, PInput) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
-
Adds an input with the given name to this Dataflow step, coming from the specified input
PValue.
- addInput(String, Map<String, Object>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
-
Adds an input that is a dictionary of strings to objects.
- addInput(String, List<? extends Map<String, Object>>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
-
Adds an input that is a list of objects.
- addInput(BeamAggregationTransforms.AggregationAccumulator, BeamRecord) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamAggregationTransforms.AggregationAdaptor
-
- addInput(ApproximateUnique.ApproximateUniqueCombineFn.LargestUnique, T) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn
-
- addInput(InputT) - Method in interface org.apache.beam.sdk.transforms.Combine.AccumulatingCombineFn.Accumulator
-
Adds the given input value to this accumulator, modifying
this accumulator.
- addInput(AccumT, InputT) - Method in class org.apache.beam.sdk.transforms.Combine.AccumulatingCombineFn
-
- addInput(double[], Double) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
-
- addInput(Combine.Holder<V>, V) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
-
- addInput(int[], Integer) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
-
- addInput(long[], Long) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
-
- addInput(AccumT, InputT) - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
-
Adds the given input value to the given accumulator, returning the
new accumulator value.
- addInput(List<V>, V) - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
-
- addInput(Object[], DataT) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
-
- addInput(Object[], DataT, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
-
- addInput(AccumT, InputT, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
-
Adds the given input value to the given accumulator, returning the
new accumulator value.
- addInput(Top.BoundedHeap<KV<Integer, T>, SerializableComparator<KV<Integer, T>>>, T) - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
-
- addList(Map<String, Object>, String, List<? extends Map<String, Object>>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- addList(Map<String, Object>, String, T[]) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- addLong(Map<String, Object>, String, long) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- addLongs(Map<String, Object>, String, long...) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- addMessage(Message) - Method in class org.apache.beam.sdk.io.jms.JmsCheckpointMark
-
- addNameFilter(MetricNameFilter) - Method in class org.apache.beam.sdk.metrics.MetricsFilter.Builder
-
- addNull(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- addObject(Map<String, Object>, String, Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- addOutput(PCollection<?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
-
Adds a primitive output to this Dataflow step, producing the specified output PValue
,
including its Coder
if a TypedPValue
.
- addOverrideForClass(Class<?>, DataflowWorkerLoggingOptions.Level) - Method in class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
-
Overrides the default log level for the passed in class.
- addOverrideForName(String, DataflowWorkerLoggingOptions.Level) - Method in class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
-
Overrides the default log level for the passed in name.
- addOverrideForPackage(Package, DataflowWorkerLoggingOptions.Level) - Method in class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
-
Overrides the default log level for the passed in package.
- addProperties(Configuration, Properties) - Static method in class org.apache.beam.runners.apex.ApexYarnLauncher
-
Transfer the properties to the configuration object.
- addStep(PTransform<?, ?>, String) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
-
Adds a step to the Dataflow workflow for the given transform, with the given Dataflow step
type.
- addStep(String) - Method in class org.apache.beam.sdk.metrics.MetricsFilter.Builder
-
Add a step filter.
- addString(Map<String, Object>, String, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- addStringList(Map<String, Object>, String, List<String>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- addTraceFor(AbstractGoogleClient, String) - Method in class org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions.GoogleApiTracer
-
- addTraceFor(AbstractGoogleClientRequest<?>, String) - Method in class org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions.GoogleApiTracer
-
- advance() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-
- advance(String) - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
-
Advances the watermarks to the next-in-line watermarks.
- advance() - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
-
- advance() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
- advance() - Method in class org.apache.beam.sdk.io.Source.Reader
-
Advances the reader to the next valid record.
- advance() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
-
Advances the reader to the next valid record.
- advanceBy(Duration) - Static method in class org.apache.beam.sdk.testing.TestStream.ProcessingTimeEvent
-
For internal use only: no backwards compatibility guarantees.
- advanceImpl() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
-
- advanceImpl() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
Advances to the next record and returns true
, or returns false if there is no next
record.
- advanceNextBatchWatermarkToInfinity() - Method in class org.apache.beam.runners.spark.io.CreateStream
-
Advances the watermark in the next batch to the end-of-time.
- advanceProcessingTime(Duration) - Method in class org.apache.beam.sdk.testing.TestStream.Builder
-
Advance the processing time by the specified amount.
- advanceTo(Instant) - Static method in class org.apache.beam.sdk.testing.TestStream.WatermarkEvent
-
For internal use only: no backwards compatibility guarantees.
- advanceWatermark() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-
Advances the watermark.
- advanceWatermarkForNextBatch(Instant) - Method in class org.apache.beam.runners.spark.io.CreateStream
-
Advances the watermark in the next batch.
- advanceWatermarkTo(Instant) - Method in class org.apache.beam.sdk.testing.TestStream.Builder
-
Advance the watermark of this source to the specified instant.
- advanceWatermarkToInfinity() - Method in class org.apache.beam.sdk.testing.TestStream.Builder
-
Advance the watermark to infinity, completing this
TestStream
.
- AfterAll - Class in org.apache.beam.sdk.transforms.windowing
-
A composite
Trigger
that fires when all of its sub-triggers are ready.
- AfterEach - Class in org.apache.beam.sdk.transforms.windowing
-
A composite
Trigger
that executes its sub-triggers in order.
- AfterFirst - Class in org.apache.beam.sdk.transforms.windowing
-
A composite
Trigger
that fires once after at least one of its sub-triggers have fired.
- AfterPane - Class in org.apache.beam.sdk.transforms.windowing
-
A
Trigger
that fires at some point after a specified number of input elements have
arrived.
- AfterProcessingTime - Class in org.apache.beam.sdk.transforms.windowing
-
A
Trigger
trigger that fires at a specified point in processing time, relative to when
input first arrives.
- AfterSynchronizedProcessingTime - Class in org.apache.beam.sdk.transforms.windowing
-
FOR INTERNAL USE ONLY.
- afterTimeSinceNewOutput(ReadableDuration) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
-
- afterTimeSinceNewOutput(SerializableFunction<InputT, ReadableDuration>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
-
- afterTotalOf(ReadableDuration) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
-
- afterTotalOf(SerializableFunction<InputT, ReadableDuration>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
-
- AfterWatermark - Class in org.apache.beam.sdk.transforms.windowing
-
AfterWatermark
triggers fire based on progress of the system watermark.
- AfterWatermark.AfterWatermarkEarlyAndLate - Class in org.apache.beam.sdk.transforms.windowing
-
- AfterWatermark.FromEndOfWindow - Class in org.apache.beam.sdk.transforms.windowing
-
A watermark trigger targeted relative to the end of the window.
- AggAccumParam - Class in org.apache.beam.runners.spark.aggregators
-
Aggregator accumulator param.
- AggAccumParam() - Constructor for class org.apache.beam.runners.spark.aggregators.AggAccumParam
-
- AggregationAccumulator() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamAggregationTransforms.AggregationAccumulator
-
- AggregationAccumulatorCoder(List<Coder>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamAggregationTransforms.AggregationAccumulatorCoder
-
- AggregationAdaptor(List<AggregateCall>, BeamRecordSqlType) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamAggregationTransforms.AggregationAdaptor
-
- AggregationGroupByKeyFn(int, ImmutableBitSet) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamAggregationTransforms.AggregationGroupByKeyFn
-
- AggregatorMetric - Class in org.apache.beam.runners.spark.metrics
-
- AggregatorMetricSource - Class in org.apache.beam.runners.spark.metrics
-
- AggregatorMetricSource(String, NamedAggregators) - Constructor for class org.apache.beam.runners.spark.metrics.AggregatorMetricSource
-
- AggregatorsAccumulator - Class in org.apache.beam.runners.spark.aggregators
-
For resilience, Accumulators
are required to be wrapped in a Singleton.
- AggregatorsAccumulator() - Constructor for class org.apache.beam.runners.spark.aggregators.AggregatorsAccumulator
-
- AggregatorsAccumulator.AccumulatorCheckpointingSparkListener - Class in org.apache.beam.runners.spark.aggregators
-
- align(Duration) - Method in interface org.apache.beam.sdk.state.Timer
-
- alignedTo(Duration, Instant) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
-
Aligns timestamps to the smallest multiple of period
since the offset
greater
than the timestamp.
- alignedTo(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
-
Aligns the time to be the smallest multiple of period
greater than the epoch
boundary (aka new Instant(0)
).
- alignTo(Duration, Instant) - Static method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform
-
For internal use only; no backwards-compatibility guarantees.
- alignTo(Duration) - Static method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform
-
For internal use only; no backwards-compatibility guarantees.
- AlignTo() - Constructor for class org.apache.beam.sdk.transforms.windowing.TimestampTransform.AlignTo
-
- ALL_CONTEXTS - Static variable in class org.apache.beam.sdk.testing.CoderProperties
-
All the contexts, for use in test cases.
- ALL_KEYS - Static variable in class org.apache.beam.sdk.io.range.ByteKeyRange
-
The range of all keys, with empty start and end keys.
- allMatches(String) - Static method in class org.apache.beam.sdk.transforms.Regex
-
- allMatches(Pattern) - Static method in class org.apache.beam.sdk.transforms.Regex
-
- AllMatches(Pattern) - Constructor for class org.apache.beam.sdk.transforms.Regex.AllMatches
-
- allOf(Watch.Growth.TerminationCondition<InputT, FirstStateT>, Watch.Growth.TerminationCondition<InputT, SecondStateT>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
-
- allowsDynamicSplitting() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-
- allowsDynamicSplitting() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
-
- allowsDynamicSplitting() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
Whether this reader should allow dynamic splitting of the offset ranges.
- AlwaysPassMatcher() - Constructor for class org.apache.beam.sdk.testing.TestPipelineOptions.AlwaysPassMatcher
-
- AlwaysPassMatcherFactory() - Constructor for class org.apache.beam.sdk.testing.TestPipelineOptions.AlwaysPassMatcherFactory
-
- alwaysRetry() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy
-
Always retry all failures.
- AmqpIO - Class in org.apache.beam.sdk.io.amqp
-
AmqpIO supports AMQP 1.0 protocol using the Apache QPid Proton-J library.
- AmqpIO.Read - Class in org.apache.beam.sdk.io.amqp
-
A
PTransform
to read/receive messages using AMQP 1.0 protocol.
- AmqpIO.Write - Class in org.apache.beam.sdk.io.amqp
-
A
PTransform
to send messages using AMQP 1.0 protocol.
- AmqpMessageCoder - Class in org.apache.beam.sdk.io.amqp
-
A coder for AMQP message.
- AmqpMessageCoder() - Constructor for class org.apache.beam.sdk.io.amqp.AmqpMessageCoder
-
- AmqpMessageCoderProviderRegistrar - Class in org.apache.beam.sdk.io.amqp
-
- AmqpMessageCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.amqp.AmqpMessageCoderProviderRegistrar
-
- and(TupleTag<V>, List<V>) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
-
Returns a new
CoGbkResult
based on this, with the given tag and given data
added to it.
- and(TupleTag<V>, PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
-
Returns a new KeyedPCollectionTuple<K>
that is the same as this,
appended with the given PCollection.
- and(PCollection.IsBounded) - Method in enum org.apache.beam.sdk.values.PCollection.IsBounded
-
Returns the composed IsBounded property.
- and(PCollection<T>) - Method in class org.apache.beam.sdk.values.PCollectionList
-
- and(Iterable<PCollection<T>>) - Method in class org.apache.beam.sdk.values.PCollectionList
-
- and(TupleTag<T>, PCollection<T>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
-
- and(TupleTag<?>) - Method in class org.apache.beam.sdk.values.TupleTagList
-
- and(List<TupleTag<?>>) - Method in class org.apache.beam.sdk.values.TupleTagList
-
- any(long) - Static method in class org.apache.beam.sdk.transforms.Sample
-
Sample#any(long)
takes a PCollection<T>
and a limit, and
produces a new PCollection<T>
containing up to limit
elements of the input PCollection
.
- ApexPipelineOptions - Interface in org.apache.beam.runners.apex
-
Options that configure the Apex pipeline.
- ApexRunner - Class in org.apache.beam.runners.apex
-
A
PipelineRunner
that translates the
pipeline to an Apex DAG and executes it on an Apex cluster.
- ApexRunner(ApexPipelineOptions) - Constructor for class org.apache.beam.runners.apex.ApexRunner
-
- ApexRunner.CreateApexPCollectionView<ElemT,ViewT> - Class in org.apache.beam.runners.apex
-
- ApexRunnerRegistrar - Class in org.apache.beam.runners.apex
-
- ApexRunnerRegistrar.Options - Class in org.apache.beam.runners.apex
-
- ApexRunnerRegistrar.Runner - Class in org.apache.beam.runners.apex
-
- ApexRunnerResult - Class in org.apache.beam.runners.apex
-
Result of executing a
Pipeline
with Apex in embedded mode.
- ApexRunnerResult(DAG, Launcher.AppHandle) - Constructor for class org.apache.beam.runners.apex.ApexRunnerResult
-
- ApexYarnLauncher - Class in org.apache.beam.runners.apex
-
Proxy to launch the YARN application through the hadoop script to run in the
pre-configured environment (class path, configuration, native libraries etc.).
- ApexYarnLauncher() - Constructor for class org.apache.beam.runners.apex.ApexYarnLauncher
-
- ApexYarnLauncher.LaunchParams - Class in org.apache.beam.runners.apex
-
Launch parameters that will be serialized and passed to the child process.
- ApexYarnLauncher.ProcessWatcher - Class in org.apache.beam.runners.apex
-
Starts a command and waits for it to complete.
- applicableTo(PCollection<?>) - Static method in class org.apache.beam.sdk.transforms.GroupByKey
-
- ApplicationNameOptions - Interface in org.apache.beam.sdk.options
-
Options that allow setting the application name.
- apply(WindowFunction.Context<T>) - Method in class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator.GearpumpWindowFn
-
- apply(InputT) - Method in interface org.apache.beam.sdk.coders.DelegateCoder.CodingFunction
-
- apply(BeamRecord) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamAggregationTransforms.AggregationGroupByKeyFn
-
- apply(BeamRecord) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamAggregationTransforms.WindowTimestampFn
-
- apply(BeamRecord) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamJoinTransforms.ExtractJoinFields
-
- apply(KV<BeamRecord, KV<BeamRecord, BeamRecord>>) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamJoinTransforms.JoinParts2WholeRow
-
- apply(BeamRecord) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSetOperatorsTransforms.BeamSqlRow2KvFn
-
- apply(PTransform<? super PBegin, OutputT>) - Method in class org.apache.beam.sdk.Pipeline
-
- apply(String, PTransform<? super PBegin, OutputT>) - Method in class org.apache.beam.sdk.Pipeline
-
- apply(T) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert.MatcherCheckerFn
-
- apply(Statement, Description) - Method in class org.apache.beam.sdk.testing.TestPipeline
-
- apply(double, double) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
-
Applies the binary operation to the two operands, returning the result.
- apply(V, V) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
-
Applies the binary operation to the two operands, returning the result.
- apply(int, int) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
-
Applies the binary operation to the two operands, returning the result.
- apply(long, long) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
-
Applies the binary operation to the two operands, returning the result.
- apply(Iterable<? extends InputT>) - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
-
Applies this CombineFn
to a collection of input values
to produce a combined output value.
- apply(Iterable<? extends InputT>, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
-
Applies this CombineFnWithContext
to a collection of input values to produce a
combined output value.
- apply(PTransform<KeyedPCollectionTuple<K>, OutputT>) - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
-
- apply(String, PTransform<KeyedPCollectionTuple<K>, OutputT>) - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
-
Applies the given
PTransform
to this input
KeyedPCollectionTuple
and returns
its
OutputT
.
- apply(InputT) - Method in interface org.apache.beam.sdk.transforms.SerializableFunction
-
Returns the result of invoking this function on the given input.
- apply(InputT) - Method in class org.apache.beam.sdk.transforms.SimpleFunction
-
- apply(PrimitiveViewT) - Method in class org.apache.beam.sdk.transforms.ViewFn
-
A function to adapt a primitive view type to a desired view type.
- apply(InputT, Instant) - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.PollFn
-
- apply(PTransform<? super PBegin, OutputT>) - Method in class org.apache.beam.sdk.values.PBegin
-
- apply(String, PTransform<? super PBegin, OutputT>) - Method in class org.apache.beam.sdk.values.PBegin
-
Applies the given
PTransform
to this
PBegin
, using
name
to identify
this specific application of the transform.
- apply(PTransform<? super PCollection<T>, OutputT>) - Method in class org.apache.beam.sdk.values.PCollection
-
- apply(String, PTransform<? super PCollection<T>, OutputT>) - Method in class org.apache.beam.sdk.values.PCollection
-
Applies the given
PTransform
to this input
PCollection
,
using
name
to identify this specific application of the transform.
- apply(PTransform<PCollectionList<T>, OutputT>) - Method in class org.apache.beam.sdk.values.PCollectionList
-
- apply(String, PTransform<PCollectionList<T>, OutputT>) - Method in class org.apache.beam.sdk.values.PCollectionList
-
Applies the given
PTransform
to this input
PCollectionList
,
using
name
to identify this specific application of the transform.
- apply(PTransform<PCollectionTuple, OutputT>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
-
- apply(String, PTransform<PCollectionTuple, OutputT>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
-
- apply(Iterable<WindowedValue<T>>) - Method in class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn
-
Deprecated.
- apply(Iterable<WindowedValue<T>>) - Method in class org.apache.beam.sdk.values.PCollectionViews.ListViewFn
-
Deprecated.
- apply(Iterable<WindowedValue<KV<K, V>>>) - Method in class org.apache.beam.sdk.values.PCollectionViews.MapViewFn
-
Deprecated.
Input iterable must actually be Iterable<WindowedValue<KV<K, V>>>
.
- apply(Iterable<WindowedValue<KV<K, V>>>) - Method in class org.apache.beam.sdk.values.PCollectionViews.MultimapViewFn
-
Deprecated.
- apply(Iterable<WindowedValue<T>>) - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn
-
Deprecated.
- applyTransform(InputT, PTransform<? super InputT, OutputT>) - Static method in class org.apache.beam.sdk.Pipeline
-
For internal use only; no backwards-compatibility guarantees.
- applyTransform(String, InputT, PTransform<? super InputT, OutputT>) - Static method in class org.apache.beam.sdk.Pipeline
-
For internal use only; no backwards-compatibility guarantees.
- ApproximateQuantiles - Class in org.apache.beam.sdk.transforms
-
PTransform
s for getting an idea of a PCollection
's
data distribution using approximate N
-tiles (e.g.
- ApproximateQuantiles.ApproximateQuantilesCombineFn<T,ComparatorT extends java.util.Comparator<T> & java.io.Serializable> - Class in org.apache.beam.sdk.transforms
-
The ApproximateQuantilesCombineFn
combiner gives an idea
of the distribution of a collection of values using approximate
N
-tiles.
- ApproximateUnique - Class in org.apache.beam.sdk.transforms
-
PTransform
s for estimating the number of distinct elements
in a PCollection
, or the number of distinct values
associated with each key in a PCollection
of KV
s.
- ApproximateUnique() - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique
-
- ApproximateUnique.ApproximateUniqueCombineFn<T> - Class in org.apache.beam.sdk.transforms
-
CombineFn
that computes an estimate of the number of
distinct values that were combined.
- ApproximateUnique.ApproximateUniqueCombineFn.LargestUnique - Class in org.apache.beam.sdk.transforms
-
A heap utility class to efficiently track the largest added elements.
- ApproximateUniqueCombineFn(long, Coder<T>) - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn
-
- array() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
-
Returns the backing array.
- as(Class<T>) - Method in interface org.apache.beam.sdk.options.PipelineOptions
-
Transforms this object into an object of type <T>
saving each property
that has been manipulated.
- as(Class<T>) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
-
Creates and returns an object that implements <T>
.
- as(Class<T>) - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
-
Creates and returns an object that implements <T>
using the values configured on
this builder during construction.
- asCloudObject(Coder<?>) - Static method in class org.apache.beam.runners.dataflow.util.CloudObjects
-
- asInputStream(int, int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
-
Returns an InputStream
wrapper which supplies the portion of this backing byte buffer
starting at offset
and up to length
bytes.
- asIterable() - Static method in class org.apache.beam.sdk.transforms.View
-
- AsJsons<InputT> - Class in org.apache.beam.sdk.extensions.jackson
-
PTransform
for serializing objects to JSON
Strings
.
- asList() - Static method in class org.apache.beam.sdk.transforms.View
-
- asMap() - Method in class org.apache.beam.sdk.transforms.display.DisplayData
-
- asMap() - Static method in class org.apache.beam.sdk.transforms.View
-
- asMultimap() - Static method in class org.apache.beam.sdk.transforms.View
-
- asOutputReference(PValue, AppliedPTransform<?, ?, ?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
-
Encode a PValue reference as an output reference.
- asOutputStream() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
-
Returns an output stream which writes to the backing buffer from the current position.
- asPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription
-
Returns the string representation of this subscription as a path used in the Cloud Pub/Sub
API.
- asPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic
-
Returns the string representation of this topic as a path used in the Cloud Pub/Sub
API.
- ASSERTION_ERROR - Static variable in class org.apache.beam.runners.apex.ApexRunner
-
TODO: this isn't thread safe and may cause issues when tests run in parallel
Holds any most resent assertion error that was raised while processing elements.
- assertionError() - Method in class org.apache.beam.sdk.testing.SuccessOrFailure
-
- assertSourcesEqualReferenceSource(BoundedSource<T>, List<? extends BoundedSource<T>>, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-
Given a reference Source
and a list of Source
s, assert that the union of
the records read from the list of sources is equal to the records read from the reference
source.
- assertSplitAtFractionBehavior(BoundedSource<T>, int, double, SourceTestUtils.ExpectedSplitOutcome, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-
- assertSplitAtFractionExhaustive(BoundedSource<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-
Asserts that for each possible start position,
BoundedSource.BoundedReader#splitAtFraction
at every interesting fraction (halfway
between two fractions that differ by at least one item) can be called successfully and the
results are consistent if a split succeeds.
- assertSplitAtFractionFails(BoundedSource<T>, int, double, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-
Asserts that the source
's reader fails to splitAtFraction(fraction)
after reading numItemsToReadBeforeSplit
items.
- assertSplitAtFractionSucceedsAndConsistent(BoundedSource<T>, int, double, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-
Verifies some consistency properties of
BoundedSource.BoundedReader#splitAtFraction
on the given source.
- assertUnstartedReaderReadsSameAsItsSource(BoundedSource.BoundedReader<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-
Assert that a Reader
returns a Source
that, when read from, produces the same
records as the reader.
- assign(BoundedWindow, Instant) - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
-
- AssignContext() - Constructor for class org.apache.beam.sdk.transforms.windowing.WindowFn.AssignContext
-
- assignedWindows(WindowFn<T, W>, long) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
-
- assignedWindowsWithValue(WindowFn<T, W>, TimestampedValue<T>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
-
- assignsToOneWindow() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-
- assignsToOneWindow() - Method in class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
-
- assignsToOneWindow() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-
- assignsToOneWindow() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
-
Returns true if this
WindowFn
always assigns an element to exactly one window.
- assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-
- assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-
- assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-
- assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-
- assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
-
Returns the single window to which elements with this timestamp belong.
- assignWindows(WindowFn<Object, GlobalWindow>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-
- assignWindows(WindowFn<Object, W>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
-
- assignWindows(WindowFn<T, W>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
-
- assignWindows(WindowFn<Object, IntervalWindow>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-
- assignWindows(WindowFn<Object, IntervalWindow>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-
- assignWindows(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-
- assignWindows(WindowFn<T, W>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
-
Given a timestamp and element, returns the set of windows into which it
should be placed.
- asSingleton() - Static method in class org.apache.beam.sdk.transforms.View
-
- asSingletonView() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
-
Returns a
PTransform
that produces a
PCollectionView
whose elements are the result of combining elements per-window in
the input
PCollection
.
- asV1Beta1Path() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription
-
- asV1Beta1Path() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic
-
- asV1Beta2Path() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription
-
- asV1Beta2Path() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic
-
- atMinimumTimestamp(V) - Static method in class org.apache.beam.sdk.values.TimestampedValue
-
- AtomicCoder<T> - Class in org.apache.beam.sdk.coders
-
A
Coder
that has no component
Coders
or other configuration.
- AtomicCoder() - Constructor for class org.apache.beam.sdk.coders.AtomicCoder
-
- AtomicLongFactory() - Constructor for class org.apache.beam.sdk.options.PipelineOptions.AtomicLongFactory
-
- attached() - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
-
- attempted() - Method in interface org.apache.beam.sdk.metrics.MetricResult
-
Return the value of this metric across all attempts of executing all parts of the pipeline.
- autoCastField(int, Object) - Static method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamTableUtils
-
- AvailableParallelismFactory() - Constructor for class org.apache.beam.runners.direct.DirectOptions.AvailableParallelismFactory
-
- AvroCoder<T> - Class in org.apache.beam.sdk.coders
-
A
Coder
using Avro binary format.
- AvroCoder(Class<T>, Schema) - Constructor for class org.apache.beam.sdk.coders.AvroCoder
-
- AvroIO - Class in org.apache.beam.sdk.io
-
- AvroIO.Parse<T> - Class in org.apache.beam.sdk.io
-
- AvroIO.ParseAll<T> - Class in org.apache.beam.sdk.io
-
- AvroIO.Read<T> - Class in org.apache.beam.sdk.io
-
- AvroIO.ReadAll<T> - Class in org.apache.beam.sdk.io
-
- AvroIO.TypedWrite<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io
-
- AvroIO.Write<T> - Class in org.apache.beam.sdk.io
-
- AvroReader(AvroSource<T>) - Constructor for class org.apache.beam.sdk.io.AvroSource.AvroReader
-
Reads Avro records of type T
from the specified source.
- AvroSource<T> - Class in org.apache.beam.sdk.io
-
Do not use in pipelines directly: most users should use
AvroIO.Read
.
- AvroSource.AvroReader<T> - Class in org.apache.beam.sdk.io
-
- awaitTermination(Duration) - Method in class org.apache.beam.runners.spark.SparkPipelineResult
-
- awaitTermination(Duration) - Method in class org.apache.beam.runners.spark.SparkRunnerDebugger.DebugSparkPipelineResult
-
- AWSClientsProvider - Interface in org.apache.beam.sdk.io.kinesis
-
Provides instances of AWS clients.
- BACKLOG_UNKNOWN - Static variable in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
-
Constant representing an unknown amount of backlog.
- backlogBytes() - Static method in class org.apache.beam.sdk.metrics.SourceMetrics
-
Gauge for source backlog in bytes.
- backlogBytesOfSplit(String) - Static method in class org.apache.beam.sdk.metrics.SourceMetrics
-
Gauge for source split backlog in bytes.
- backlogElements() - Static method in class org.apache.beam.sdk.metrics.SourceMetrics
-
Gauge for source backlog in elements.
- backlogElementsOfSplit(String) - Static method in class org.apache.beam.sdk.metrics.SourceMetrics
-
Gauge for source split backlog in elements.
- bag() - Static method in class org.apache.beam.sdk.state.StateSpecs
-
Create a
StateSpec
for a
BagState
, optimized for adding values frequently
and occasionally retrieving all the values that have been added.
- bag(Coder<T>) - Static method in class org.apache.beam.sdk.state.StateSpecs
-
- BagState<T> - Interface in org.apache.beam.sdk.state
-
- BaseBeamTable - Class in org.apache.beam.sdk.extensions.sql.impl.schema
-
Each IO in Beam has one table schema, by extending
BaseBeamTable
.
- BaseBeamTable(BeamRecordSqlType) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.BaseBeamTable
-
- BatchStatefulParDoOverrides - Class in org.apache.beam.runners.dataflow
-
PTransformOverrideFactories
that expands to correctly implement
stateful
ParDo
using window-unaware
BatchViewOverrides.GroupByKeyAndSortValuesOnly
to linearize
processing per key.
- BatchStatefulParDoOverrides() - Constructor for class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides
-
- BatchStatefulParDoOverrides.BatchStatefulDoFn<K,V,OutputT> - Class in org.apache.beam.runners.dataflow
-
A key-preserving
DoFn
that explodes an iterable that has been grouped by key and
window.
- BEAM_REL_DATATYPE_SYSTEM - Static variable in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelDataTypeSystem
-
- BeamAggregationRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
-
- BeamAggregationRel(RelOptCluster, RelTraitSet, RelNode, boolean, ImmutableBitSet, List<ImmutableBitSet>, List<AggregateCall>, WindowFn, Trigger, int, Duration) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamAggregationRel
-
- BeamAggregationRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
-
Rule to detect the window/trigger settings.
- BeamAggregationRule(Class<? extends Aggregate>, Class<? extends Project>, RelBuilderFactory) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rule.BeamAggregationRule
-
- BeamAggregationRule(RelOptRuleOperand, String) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rule.BeamAggregationRule
-
- BeamAggregationTransforms - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
Collections of PTransform
and DoFn
used to perform GROUP-BY operation.
- BeamAggregationTransforms() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamAggregationTransforms
-
- BeamAggregationTransforms.AggregationAccumulator - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
A class to holder varied accumulator objects.
- BeamAggregationTransforms.AggregationAccumulatorCoder - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
- BeamAggregationTransforms.AggregationAdaptor - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
An adaptor class to invoke Calcite UDAF instances in Beam CombineFn
.
- BeamAggregationTransforms.AggregationGroupByKeyFn - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
extract group-by fields.
- BeamAggregationTransforms.MergeAggregationRecord - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
Merge KV to single record.
- BeamAggregationTransforms.WindowTimestampFn - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
Assign event timestamp.
- BeamFilterRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
-
BeamRelNode to replace a Filter
node.
- BeamFilterRel(RelOptCluster, RelTraitSet, RelNode, RexNode) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamFilterRel
-
- BeamFilterRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
-
- BeamIntersectRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
-
BeamRelNode
to replace a Intersect
node.
- BeamIntersectRel(RelOptCluster, RelTraitSet, List<RelNode>, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIntersectRel
-
- BeamIntersectRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
-
ConverterRule
to replace Intersect
with BeamIntersectRel
.
- BeamIOSinkRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
-
BeamRelNode to replace a TableModify
node.
- BeamIOSinkRel(RelOptCluster, RelTraitSet, RelOptTable, Prepare.CatalogReader, RelNode, Operation, List<String>, List<RexNode>, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
-
- BeamIOSinkRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
-
- BeamIOSourceRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
-
BeamRelNode to replace a TableScan
node.
- BeamIOSourceRel(RelOptCluster, RelTraitSet, RelOptTable) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
-
- BeamIOSourceRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
-
- BeamIOType - Enum in org.apache.beam.sdk.extensions.sql.impl.schema
-
Type as a source IO, determined whether it's a STREAMING process, or batch
process.
- BeamJoinRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
-
BeamRelNode
to replace a Join
node.
- BeamJoinRel(RelOptCluster, RelTraitSet, RelNode, RelNode, RexNode, Set<CorrelationId>, JoinRelType) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
-
- BeamJoinRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
-
ConverterRule
to replace Join
with BeamJoinRel
.
- BeamJoinTransforms - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
Collections of PTransform
and DoFn
used to perform JOIN operation.
- BeamJoinTransforms() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamJoinTransforms
-
- BeamJoinTransforms.ExtractJoinFields - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
A SimpleFunction
to extract join fields from the specified row.
- BeamJoinTransforms.JoinParts2WholeRow - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
A SimpleFunction
to combine two rows into one.
- BeamJoinTransforms.SideInputJoinDoFn - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
A DoFn
which implement the sideInput-JOIN.
- BeamKafkaCSVTable - Class in org.apache.beam.sdk.extensions.sql.impl.schema.kafka
-
A Kafka topic that saves records as CSV format.
- BeamKafkaCSVTable(BeamRecordSqlType, String, List<String>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.kafka.BeamKafkaCSVTable
-
- BeamKafkaCSVTable(BeamRecordSqlType, String, List<String>, CSVFormat) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.kafka.BeamKafkaCSVTable
-
- BeamKafkaCSVTable.CsvRecorderDecoder - Class in org.apache.beam.sdk.extensions.sql.impl.schema.kafka
-
A PTransform to convert
KV<byte[], byte[]>
to
BeamRecord
.
- BeamKafkaCSVTable.CsvRecorderEncoder - Class in org.apache.beam.sdk.extensions.sql.impl.schema.kafka
-
A PTransform to convert
BeamRecord
to
KV<byte[], byte[]>
.
- BeamKafkaTable - Class in org.apache.beam.sdk.extensions.sql.impl.schema.kafka
-
BeamKafkaTable
represent a Kafka topic, as source or target.
- BeamKafkaTable(BeamRecordSqlType) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.kafka.BeamKafkaTable
-
- BeamKafkaTable(BeamRecordSqlType, String, List<String>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.kafka.BeamKafkaTable
-
- BeamLogicalConvention - Enum in org.apache.beam.sdk.extensions.sql.impl.rel
-
Convertion for Beam SQL.
- BeamMinusRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
-
BeamRelNode
to replace a Minus
node.
- BeamMinusRel(RelOptCluster, RelTraitSet, List<RelNode>, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamMinusRel
-
- BeamMinusRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
-
ConverterRule
to replace Minus
with BeamMinusRel
.
- BeamPCollectionTable - Class in org.apache.beam.sdk.extensions.sql.impl.schema
-
BeamPCollectionTable
converts a PCollection<BeamSqlRow>
as a virtual table,
then a downstream query can query directly.
- BeamPCollectionTable(BeamRecordSqlType) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.BeamPCollectionTable
-
- BeamPCollectionTable(PCollection<BeamRecord>, BeamRecordSqlType) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.BeamPCollectionTable
-
- BeamProjectRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
-
BeamRelNode to replace a Project
node.
- BeamProjectRel(RelOptCluster, RelTraitSet, RelNode, List<? extends RexNode>, RelDataType) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamProjectRel
-
projects: RexLiteral
, RexInputRef
, RexCall
.
- BeamProjectRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
-
- BeamQueryPlanner - Class in org.apache.beam.sdk.extensions.sql.impl.planner
-
The core component to handle through a SQL statement, from explain execution plan,
to generate a Beam pipeline.
- BeamQueryPlanner(SchemaPlus) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.planner.BeamQueryPlanner
-
- BeamRecord - Class in org.apache.beam.sdk.values
-
- BeamRecord(BeamRecordType, List<Object>) - Constructor for class org.apache.beam.sdk.values.BeamRecord
-
Creates a BeamRecord.
- BeamRecord(BeamRecordType, Object...) - Constructor for class org.apache.beam.sdk.values.BeamRecord
-
- BeamRecordCoder - Class in org.apache.beam.sdk.coders
-
- BeamRecordSqlType - Class in org.apache.beam.sdk.extensions.sql
-
- BeamRecordSqlType(List<String>, List<Coder>) - Constructor for class org.apache.beam.sdk.extensions.sql.BeamRecordSqlType
-
- BeamRecordType - Class in org.apache.beam.sdk.values
-
- BeamRecordType(List<String>, List<Coder>) - Constructor for class org.apache.beam.sdk.values.BeamRecordType
-
- BeamRelDataTypeSystem - Class in org.apache.beam.sdk.extensions.sql.impl.planner
-
customized data type in Beam.
- BeamRelDataTypeSystem() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelDataTypeSystem
-
- BeamRelNode - Interface in org.apache.beam.sdk.extensions.sql.impl.rel
-
- BeamRuleSets - Class in org.apache.beam.sdk.extensions.sql.impl.planner
-
- BeamRuleSets() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRuleSets
-
- BeamSetOperatorRelBase - Class in org.apache.beam.sdk.extensions.sql.impl.rel
-
Delegate for Set operators: BeamUnionRel
, BeamIntersectRel
and BeamMinusRel
.
- BeamSetOperatorRelBase(BeamRelNode, BeamSetOperatorRelBase.OpType, List<RelNode>, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSetOperatorRelBase
-
- BeamSetOperatorRelBase.OpType - Enum in org.apache.beam.sdk.extensions.sql.impl.rel
-
Set operator type.
- BeamSetOperatorsTransforms - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
Collections of PTransform
and DoFn
used to perform Set operations.
- BeamSetOperatorsTransforms() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSetOperatorsTransforms
-
- BeamSetOperatorsTransforms.BeamSqlRow2KvFn - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
Transform a BeamSqlRow
to a KV<BeamSqlRow, BeamSqlRow>
.
- BeamSetOperatorsTransforms.SetOperatorFilteringDoFn - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
Filter function used for Set operators.
- BeamSortRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
-
BeamRelNode
to replace a Sort
node.
- BeamSortRel(RelOptCluster, RelTraitSet, RelNode, RelCollation, RexNode, RexNode) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSortRel
-
- BeamSortRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
-
ConverterRule
to replace Sort
with BeamSortRel
.
- BeamSparkRunnerRegistrator - Class in org.apache.beam.runners.spark.coders
-
Custom KryoRegistrator
s for Beam's Spark runner needs.
- BeamSparkRunnerRegistrator() - Constructor for class org.apache.beam.runners.spark.coders.BeamSparkRunnerRegistrator
-
- BeamSql - Class in org.apache.beam.sdk.extensions.sql
-
BeamSql
is the DSL interface of BeamSQL.
- BeamSql() - Constructor for class org.apache.beam.sdk.extensions.sql.BeamSql
-
- BeamSql.QueryTransform - Class in org.apache.beam.sdk.extensions.sql
-
A
PTransform
representing an execution plan for a SQL query.
- BeamSql.SimpleQueryTransform - Class in org.apache.beam.sdk.extensions.sql
-
A
PTransform
representing an execution plan for a SQL query referencing
a single table.
- BeamSqlAbsExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'ABS' function.
- BeamSqlAbsExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlAbsExpression
-
- BeamSqlAcosExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'ACOS' function.
- BeamSqlAcosExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlAcosExpression
-
- BeamSqlAndExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.logical
-
BeamSqlExpression
for 'AND' operation.
- BeamSqlAndExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.logical.BeamSqlAndExpression
-
- BeamSqlArithmeticExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic
-
Base class for all arithmetic operators.
- BeamSqlArithmeticExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlArithmeticExpression
-
- BeamSqlArithmeticExpression(List<BeamSqlExpression>, SqlTypeName) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlArithmeticExpression
-
- BeamSqlAsinExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'ASIN' function.
- BeamSqlAsinExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlAsinExpression
-
- BeamSqlAtan2Expression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
- BeamSqlAtan2Expression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlAtan2Expression
-
- BeamSqlAtanExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'ATAN' function.
- BeamSqlAtanExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlAtanExpression
-
- BeamSqlCaseExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator
-
BeamSqlCaseExpression
represents CASE, NULLIF, COALESCE in SQL.
- BeamSqlCaseExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlCaseExpression
-
- BeamSqlCastExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator
-
Base class to support 'CAST' operations for all SqlTypeName
.
- BeamSqlCastExpression(List<BeamSqlExpression>, SqlTypeName) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlCastExpression
-
- BeamSqlCeilExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'CEIL' function.
- BeamSqlCeilExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlCeilExpression
-
- BeamSqlCharLengthExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string
-
'CHAR_LENGTH' operator.
- BeamSqlCharLengthExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlCharLengthExpression
-
- BeamSqlCli - Class in org.apache.beam.sdk.extensions.sql.impl
-
BeamSqlCli
provides methods to execute Beam SQL with an interactive client.
- BeamSqlCli() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.BeamSqlCli
-
- BeamSqlCompareExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison
-
- BeamSqlCompareExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlCompareExpression
-
- BeamSqlConcatExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string
-
String concat operator.
- BeamSqlConcatExpression(List<BeamSqlExpression>, SqlTypeName) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlConcatExpression
-
- BeamSqlConcatExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlConcatExpression
-
- BeamSqlCosExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'COS' function.
- BeamSqlCosExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlCosExpression
-
- BeamSqlCotExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'COT' function.
- BeamSqlCotExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlCotExpression
-
- BeamSqlCurrentDateExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date
-
BeamSqlExpression
for CURRENT_DATE and LOCALTIME.
- BeamSqlCurrentDateExpression() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date.BeamSqlCurrentDateExpression
-
- BeamSqlCurrentTimeExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date
-
BeamSqlExpression
for LOCALTIME and CURRENT_TIME.
- BeamSqlCurrentTimeExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date.BeamSqlCurrentTimeExpression
-
- BeamSqlCurrentTimestampExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date
-
BeamSqlExpression
for LOCALTIMESTAMP and CURRENT_TIMESTAMP.
- BeamSqlCurrentTimestampExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date.BeamSqlCurrentTimestampExpression
-
- BeamSqlDateCeilExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date
-
BeamSqlExpression
for CEIL(date).
- BeamSqlDateCeilExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date.BeamSqlDateCeilExpression
-
- BeamSqlDateFloorExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date
-
BeamSqlExpression
for FLOOR(date).
- BeamSqlDateFloorExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date.BeamSqlDateFloorExpression
-
- BeamSqlDegreesExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'DEGREES' function.
- BeamSqlDegreesExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlDegreesExpression
-
- BeamSqlDivideExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic
-
'/' operator.
- BeamSqlDivideExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlDivideExpression
-
- BeamSqlEnv - Class in org.apache.beam.sdk.extensions.sql.impl
-
- BeamSqlEnv() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
-
- BeamSqlEqualsExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison
-
BeamSqlExpression
for =
operation.
- BeamSqlEqualsExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlEqualsExpression
-
- BeamSqlExpExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'EXP' function.
- BeamSqlExpExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlExpExpression
-
- BeamSqlExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator
-
BeamSqlExpression
is an equivalent expression in BeamSQL, of RexNode
in Calcite.
- BeamSqlExpression() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlExpression
-
- BeamSqlExpression(List<BeamSqlExpression>, SqlTypeName) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlExpression
-
- BeamSqlExpressionExecutor - Interface in org.apache.beam.sdk.extensions.sql.impl.interpreter
-
BeamSqlExpressionExecutor
fills the gap between relational
expressions in Calcite SQL and executable code.
- BeamSqlExtractExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date
-
BeamSqlExpression
for EXTRACT.
- BeamSqlExtractExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date.BeamSqlExtractExpression
-
- BeamSqlFilterFn - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
- BeamSqlFilterFn(String, BeamSqlExpressionExecutor) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSqlFilterFn
-
- BeamSqlFloorExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'FLOOR' function.
- BeamSqlFloorExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlFloorExpression
-
- BeamSqlFnExecutor - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter
-
- BeamSqlFnExecutor(BeamRelNode) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.BeamSqlFnExecutor
-
- BeamSqlGreaterThanExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison
-
BeamSqlExpression
for >
operation.
- BeamSqlGreaterThanExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlGreaterThanExpression
-
- BeamSqlGreaterThanOrEqualsExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison
-
BeamSqlExpression
for >=
operation.
- BeamSqlGreaterThanOrEqualsExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlGreaterThanOrEqualsExpression
-
- BeamSqlInitCapExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string
-
'INITCAP' operator.
- BeamSqlInitCapExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlInitCapExpression
-
- BeamSqlInputRefExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator
-
An primitive operation for direct field extraction.
- BeamSqlInputRefExpression(SqlTypeName, int) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlInputRefExpression
-
- BeamSqlIsNotNullExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison
-
BeamSqlExpression
for 'IS NOT NULL' operation.
- BeamSqlIsNotNullExpression(BeamSqlExpression) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlIsNotNullExpression
-
- BeamSqlIsNullExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison
-
BeamSqlExpression
for 'IS NULL' operation.
- BeamSqlIsNullExpression(BeamSqlExpression) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlIsNullExpression
-
- BeamSqlLessThanExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison
-
BeamSqlExpression
for <
operation.
- BeamSqlLessThanExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlLessThanExpression
-
- BeamSqlLessThanOrEqualsExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison
-
BeamSqlExpression
for <=
operation.
- BeamSqlLessThanOrEqualsExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlLessThanOrEqualsExpression
-
- BeamSqlLnExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'LN' function.
- BeamSqlLnExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlLnExpression
-
- BeamSqlLogExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'Log10' function.
- BeamSqlLogExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlLogExpression
-
- BeamSqlLogicalExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.logical
-
BeamSqlExpression
for Logical operators.
- BeamSqlLogicalExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.logical.BeamSqlLogicalExpression
-
- BeamSqlLowerExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string
-
'LOWER' operator.
- BeamSqlLowerExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlLowerExpression
-
- BeamSqlMathBinaryExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
Base class for all binary functions such as
POWER, MOD, RAND_INTEGER, ATAN2, ROUND, TRUNCATE.
- BeamSqlMathBinaryExpression(List<BeamSqlExpression>, SqlTypeName) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlMathBinaryExpression
-
- BeamSqlMathUnaryExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
Base class for all unary functions such as
ABS, SQRT, LN, LOG10, EXP, CEIL, FLOOR, RAND, ACOS,
ASIN, ATAN, COS, COT, DEGREES, RADIANS, SIGN, SIN, TAN.
- BeamSqlMathUnaryExpression(List<BeamSqlExpression>, SqlTypeName) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlMathUnaryExpression
-
- BeamSqlMinusExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic
-
'-' operator.
- BeamSqlMinusExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlMinusExpression
-
- BeamSqlModExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic
-
'%' operator.
- BeamSqlModExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlModExpression
-
- BeamSqlMultiplyExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic
-
'*' operator.
- BeamSqlMultiplyExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlMultiplyExpression
-
- BeamSqlNotEqualsExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison
-
BeamSqlExpression
for <>
operation.
- BeamSqlNotEqualsExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlNotEqualsExpression
-
- BeamSqlNotExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.logical
-
BeamSqlExpression
for logical operator: NOT.
- BeamSqlNotExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.logical.BeamSqlNotExpression
-
- BeamSqlOrExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.logical
-
BeamSqlExpression
for 'OR' operation.
- BeamSqlOrExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.logical.BeamSqlOrExpression
-
- BeamSqlOutputToConsoleFn - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
A test PTransform to display output in console.
- BeamSqlOutputToConsoleFn(String) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSqlOutputToConsoleFn
-
- BeamSqlOverlayExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string
-
'OVERLAY' operator.
- BeamSqlOverlayExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlOverlayExpression
-
- BeamSqlPiExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
Base class for the PI function.
- BeamSqlPiExpression() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlPiExpression
-
- BeamSqlPlusExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic
-
'+' operator.
- BeamSqlPlusExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlPlusExpression
-
- BeamSqlPositionExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string
-
String position operator.
- BeamSqlPositionExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlPositionExpression
-
- BeamSqlPowerExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathBinaryExpression
for 'POWER' function.
- BeamSqlPowerExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlPowerExpression
-
- BeamSqlPrimitive<T> - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator
-
- BeamSqlProjectFn - Class in org.apache.beam.sdk.extensions.sql.impl.transform
-
- BeamSqlProjectFn(String, BeamSqlExpressionExecutor, BeamRecordSqlType) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSqlProjectFn
-
- BeamSqlRadiansExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'RADIANS' function.
- BeamSqlRadiansExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlRadiansExpression
-
- BeamSqlRandExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'RAND([seed])' function.
- BeamSqlRandExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlRandExpression
-
- BeamSqlRandIntegerExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'RAND_INTEGER([seed, ] numeric)'
function.
- BeamSqlRandIntegerExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlRandIntegerExpression
-
- BeamSqlRecordHelper - Class in org.apache.beam.sdk.extensions.sql
-
- BeamSqlRecordHelper() - Constructor for class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper
-
- BeamSqlRecordHelper.BooleanCoder - Class in org.apache.beam.sdk.extensions.sql
-
Coder
for Java type
Boolean
.
- BeamSqlRecordHelper.DateCoder - Class in org.apache.beam.sdk.extensions.sql
-
Coder
for Java type
Date
, it's stored as
Long
.
- BeamSqlRecordHelper.DoubleCoder - Class in org.apache.beam.sdk.extensions.sql
-
Coder
for Java type
Double
, it's stored as
BigDecimal
.
- BeamSqlRecordHelper.FloatCoder - Class in org.apache.beam.sdk.extensions.sql
-
Coder
for Java type
Float
, it's stored as
BigDecimal
.
- BeamSqlRecordHelper.ShortCoder - Class in org.apache.beam.sdk.extensions.sql
-
Coder
for Java type
Short
.
- BeamSqlRecordHelper.TimeCoder - Class in org.apache.beam.sdk.extensions.sql
-
Coder
for Java type
GregorianCalendar
, it's stored as
Long
.
- BeamSqlReinterpretExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator
-
BeamSqlExpression
for REINTERPRET.
- BeamSqlReinterpretExpression(List<BeamSqlExpression>, SqlTypeName) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlReinterpretExpression
-
- BeamSqlRoundExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathBinaryExpression
for 'ROUND' function.
- BeamSqlRoundExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlRoundExpression
-
- beamSqlRow2CsvLine(BeamRecord, CSVFormat) - Static method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamTableUtils
-
- BeamSqlRow2KvFn() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSetOperatorsTransforms.BeamSqlRow2KvFn
-
- beamSqlRowType - Variable in class org.apache.beam.sdk.extensions.sql.impl.schema.BaseBeamTable
-
- beamSqlRowType - Variable in class org.apache.beam.sdk.extensions.sql.impl.schema.text.BeamTextCSVTableIOReader
-
- beamSqlRowType - Variable in class org.apache.beam.sdk.extensions.sql.impl.schema.text.BeamTextCSVTableIOWriter
-
- BeamSqlSignExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'SIGN' function.
- BeamSqlSignExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlSignExpression
-
- BeamSqlSinExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'SIN' function.
- BeamSqlSinExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlSinExpression
-
- BeamSqlStringUnaryExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string
-
Base class for all string unary operators.
- BeamSqlStringUnaryExpression(List<BeamSqlExpression>, SqlTypeName) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlStringUnaryExpression
-
- BeamSqlSubstringExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string
-
'SUBSTRING' operator.
- BeamSqlSubstringExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlSubstringExpression
-
- BeamSqlTable - Interface in org.apache.beam.sdk.extensions.sql.impl.schema
-
This interface defines a Beam Sql Table.
- BeamSqlTanExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathUnaryExpression
for 'TAN' function.
- BeamSqlTanExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlTanExpression
-
- BeamSqlTrimExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string
-
Trim operator.
- BeamSqlTrimExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlTrimExpression
-
- BeamSqlTruncateExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
BeamSqlMathBinaryExpression
for 'TRUNCATE' function.
- BeamSqlTruncateExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlTruncateExpression
-
- BeamSqlUdf - Interface in org.apache.beam.sdk.extensions.sql
-
Interface to create a UDF in Beam SQL.
- BeamSqlUdfExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator
-
invoke a UDF function.
- BeamSqlUdfExpression(Method, List<BeamSqlExpression>, SqlTypeName) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlUdfExpression
-
- BeamSqlUpperExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string
-
'UPPER' operator.
- BeamSqlUpperExpression(List<BeamSqlExpression>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string.BeamSqlUpperExpression
-
- BeamSqlWindowEndExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator
-
BeamSqlExpression
for HOP_END
, TUMBLE_END
, SESSION_END
operation.
- BeamSqlWindowEndExpression() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlWindowEndExpression
-
- BeamSqlWindowExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator
-
BeamSqlExpression
for HOP
, TUMBLE
, SESSION
operation.
- BeamSqlWindowExpression(List<BeamSqlExpression>, SqlTypeName) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlWindowExpression
-
- BeamSqlWindowStartExpression - Class in org.apache.beam.sdk.extensions.sql.impl.interpreter.operator
-
BeamSqlExpression
for HOP_START
, TUMBLE_START
,
SESSION_START
operation.
- BeamSqlWindowStartExpression() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlWindowStartExpression
-
- BeamTableUtils - Class in org.apache.beam.sdk.extensions.sql.impl.schema
-
Utility methods for working with BeamTable
.
- BeamTableUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.BeamTableUtils
-
- BeamTextCSVTable - Class in org.apache.beam.sdk.extensions.sql.impl.schema.text
-
BeamTextCSVTable
is a BeamTextTable
which formatted in CSV.
- BeamTextCSVTable(BeamRecordSqlType, String) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.text.BeamTextCSVTable
-
CSV table with DEFAULT
format.
- BeamTextCSVTable(BeamRecordSqlType, String, CSVFormat) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.text.BeamTextCSVTable
-
- BeamTextCSVTableIOReader - Class in org.apache.beam.sdk.extensions.sql.impl.schema.text
-
IOReader for BeamTextCSVTable
.
- BeamTextCSVTableIOReader(BeamRecordSqlType, String, CSVFormat) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.text.BeamTextCSVTableIOReader
-
- BeamTextCSVTableIOWriter - Class in org.apache.beam.sdk.extensions.sql.impl.schema.text
-
IOWriter for BeamTextCSVTable
.
- BeamTextCSVTableIOWriter(BeamRecordSqlType, String, CSVFormat) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.text.BeamTextCSVTableIOWriter
-
- BeamTextTable - Class in org.apache.beam.sdk.extensions.sql.impl.schema.text
-
BeamTextTable
represents a text file/directory(backed by TextIO
).
- BeamTextTable(BeamRecordSqlType, String) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.text.BeamTextTable
-
- BeamUnionRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
-
- BeamUnionRel(RelOptCluster, RelTraitSet, List<RelNode>, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnionRel
-
- BeamUnionRel(RelInput) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnionRel
-
- BeamUnionRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
-
A
ConverterRule
to replace
org.apache.calcite.rel.core.Union
with
BeamUnionRule
.
- BeamValuesRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
-
BeamRelNode
to replace a Values
node.
- BeamValuesRel(RelOptCluster, RelDataType, ImmutableList<ImmutableList<RexLiteral>>, RelTraitSet) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamValuesRel
-
- BeamValuesRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
-
ConverterRule
to replace Values
with BeamValuesRel
.
- begin() - Method in class org.apache.beam.sdk.Pipeline
-
Returns a
PBegin
owned by this Pipeline.
- beginningOnDay(int) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-
- beginningOnDay(int, int) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-
- BigDecimalCoder - Class in org.apache.beam.sdk.coders
-
- bigdecimals() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-
- BigEndianIntegerCoder - Class in org.apache.beam.sdk.coders
-
- BigEndianLongCoder - Class in org.apache.beam.sdk.coders
-
- BigIntegerCoder - Class in org.apache.beam.sdk.coders
-
- bigintegers() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-
- BigQueryCoderProviderRegistrar - Class in org.apache.beam.sdk.io.gcp.bigquery
-
- BigQueryCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryCoderProviderRegistrar
-
- BigQueryFileToTable - Class in org.apache.beam.runners.dataflow
-
An example that stresses the BigQuery sink at scale.
- BigQueryFileToTable() - Constructor for class org.apache.beam.runners.dataflow.BigQueryFileToTable
-
- BigQueryHelpers - Class in org.apache.beam.sdk.io.gcp.bigquery
-
A set of helper functions and classes used by
BigQueryIO
.
- BigQueryHelpers() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
-
- BigQueryIO - Class in org.apache.beam.sdk.io.gcp.bigquery
-
- BigQueryIO.Read - Class in org.apache.beam.sdk.io.gcp.bigquery
-
- BigQueryIO.TypedRead<T> - Class in org.apache.beam.sdk.io.gcp.bigquery
-
- BigQueryIO.Write<T> - Class in org.apache.beam.sdk.io.gcp.bigquery
-
- BigQueryIO.Write.CreateDisposition - Enum in org.apache.beam.sdk.io.gcp.bigquery
-
An enumeration type for the BigQuery create disposition strings.
- BigQueryIO.Write.Method - Enum in org.apache.beam.sdk.io.gcp.bigquery
-
Determines the method used to insert data in BigQuery.
- BigQueryIO.Write.WriteDisposition - Enum in org.apache.beam.sdk.io.gcp.bigquery
-
An enumeration type for the BigQuery write disposition strings.
- BigQueryOptions - Interface in org.apache.beam.sdk.io.gcp.bigquery
-
Properties needed when using Google BigQuery with the Apache Beam SDK.
- BigtableIO - Class in org.apache.beam.sdk.io.gcp.bigtable
-
Transforms
for reading from and writing to Google Cloud Bigtable.
- BigtableIO.Read - Class in org.apache.beam.sdk.io.gcp.bigtable
-
A
PTransform
that reads from Google Cloud Bigtable.
- BigtableIO.Write - Class in org.apache.beam.sdk.io.gcp.bigtable
-
A
PTransform
that writes to Google Cloud Bigtable.
- BinaryCombineDoubleFn() - Constructor for class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
-
- BinaryCombineFn() - Constructor for class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
-
- BinaryCombineIntegerFn() - Constructor for class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
-
- BinaryCombineLongFn() - Constructor for class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
-
- bind(String, StateBinder) - Method in interface org.apache.beam.sdk.state.StateSpec
-
For internal use only; no backwards-compatibility guarantees.
- bindBag(String, StateSpec<BagState<T>>, Coder<T>) - Method in interface org.apache.beam.sdk.state.StateBinder
-
- bindCombining(String, StateSpec<CombiningState<InputT, AccumT, OutputT>>, Coder<AccumT>, Combine.CombineFn<InputT, AccumT, OutputT>) - Method in interface org.apache.beam.sdk.state.StateBinder
-
- bindCombiningWithContext(String, StateSpec<CombiningState<InputT, AccumT, OutputT>>, Coder<AccumT>, CombineWithContext.CombineFnWithContext<InputT, AccumT, OutputT>) - Method in interface org.apache.beam.sdk.state.StateBinder
-
- bindMap(String, StateSpec<MapState<KeyT, ValueT>>, Coder<KeyT>, Coder<ValueT>) - Method in interface org.apache.beam.sdk.state.StateBinder
-
- bindSet(String, StateSpec<SetState<T>>, Coder<T>) - Method in interface org.apache.beam.sdk.state.StateBinder
-
- bindValue(String, StateSpec<ValueState<T>>, Coder<T>) - Method in interface org.apache.beam.sdk.state.StateBinder
-
- bindWatermark(String, StateSpec<WatermarkHoldState>, TimestampCombiner) - Method in interface org.apache.beam.sdk.state.StateBinder
-
- BitSetCoder - Class in org.apache.beam.sdk.coders
-
Coder for BitSet
.
- Block() - Constructor for class org.apache.beam.sdk.io.BlockBasedSource.Block
-
- BlockBasedReader(BlockBasedSource<T>) - Constructor for class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
-
- BlockBasedSource<T> - Class in org.apache.beam.sdk.io
-
A
BlockBasedSource
is a
FileBasedSource
where a file consists of blocks of
records.
- BlockBasedSource(String, EmptyMatchTreatment, long) - Constructor for class org.apache.beam.sdk.io.BlockBasedSource
-
Creates a BlockBasedSource
based on a file name or pattern.
- BlockBasedSource(String, long) - Constructor for class org.apache.beam.sdk.io.BlockBasedSource
-
- BlockBasedSource(ValueProvider<String>, long) - Constructor for class org.apache.beam.sdk.io.BlockBasedSource
-
- BlockBasedSource(ValueProvider<String>, EmptyMatchTreatment, long) - Constructor for class org.apache.beam.sdk.io.BlockBasedSource
-
- BlockBasedSource(MatchResult.Metadata, long, long, long) - Constructor for class org.apache.beam.sdk.io.BlockBasedSource
-
Creates a BlockBasedSource
for a single file.
- BlockBasedSource.Block<T> - Class in org.apache.beam.sdk.io
-
A Block
represents a block of records that can be read.
- BlockBasedSource.BlockBasedReader<T> - Class in org.apache.beam.sdk.io
-
- BooleanCoder - Class in org.apache.beam.sdk.coders
-
- BooleanCoder() - Constructor for class org.apache.beam.sdk.coders.BooleanCoder
-
- booleans() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-
- Bounded(SparkContext, BoundedSource<T>, SerializablePipelineOptions, String) - Constructor for class org.apache.beam.runners.spark.io.SourceRDD.Bounded
-
- BoundedReader() - Constructor for class org.apache.beam.sdk.io.BoundedSource.BoundedReader
-
- BoundedReadFromUnboundedSource<T> - Class in org.apache.beam.sdk.io
-
PTransform
that reads a bounded amount of data from an
UnboundedSource
,
specified as one or both of a maximum number of elements or a maximum period of time to read.
- BoundedSource<T> - Class in org.apache.beam.sdk.io
-
A
Source
that reads a finite amount of input and, because of that, supports
some additional operations.
- BoundedSource() - Constructor for class org.apache.beam.sdk.io.BoundedSource
-
- BoundedSource.BoundedReader<T> - Class in org.apache.beam.sdk.io
-
A Reader
that reads a bounded amount of input and supports some additional
operations, such as progress estimation and dynamic work rebalancing.
- BoundedSourceWrapper<T> - Class in org.apache.beam.runners.gearpump.translators.io
-
wrapper over BoundedSource for Gearpump DataSource API.
- BoundedSourceWrapper(BoundedSource<T>, PipelineOptions) - Constructor for class org.apache.beam.runners.gearpump.translators.io.BoundedSourceWrapper
-
- BoundedWindow - Class in org.apache.beam.sdk.transforms.windowing
-
A
BoundedWindow
represents window information assigned to data elements.
- BoundedWindow() - Constructor for class org.apache.beam.sdk.transforms.windowing.BoundedWindow
-
- boundedWindowToGearpumpWindow(BoundedWindow) - Static method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils
-
- broadcast(JavaSparkContext) - Method in class org.apache.beam.runners.spark.util.SideInputBroadcast
-
- BufferedExternalSorter - Class in org.apache.beam.sdk.extensions.sorter
-
Sorter
that will use in memory sorting until the values can't fit into memory and will
then fall back to external sorting.
- BufferedExternalSorter.Options - Class in org.apache.beam.sdk.extensions.sorter
-
Contains configuration for the sorter.
- build() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions.Builder
-
- build() - Method in class org.apache.beam.sdk.io.fs.CreateOptions.StandardCreateOptions.Builder
-
- build() - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
-
- build() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig.Builder
-
- build() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction.Builder
-
- build() - Method in class org.apache.beam.sdk.metrics.MetricsFilter.Builder
-
- buildBeamPipeline(PCollectionTuple, BeamSqlEnv) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamAggregationRel
-
- buildBeamPipeline(PCollectionTuple, BeamSqlEnv) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamFilterRel
-
- buildBeamPipeline(PCollectionTuple, BeamSqlEnv) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIntersectRel
-
- buildBeamPipeline(PCollectionTuple, BeamSqlEnv) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
-
Note that BeamIOSinkRel
returns the input PCollection,
which is the persisted PCollection.
- buildBeamPipeline(PCollectionTuple, BeamSqlEnv) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
-
- buildBeamPipeline(PCollectionTuple, BeamSqlEnv) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
-
- buildBeamPipeline(PCollectionTuple, BeamSqlEnv) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamMinusRel
-
- buildBeamPipeline(PCollectionTuple, BeamSqlEnv) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamProjectRel
-
- buildBeamPipeline(PCollectionTuple, BeamSqlEnv) - Method in interface org.apache.beam.sdk.extensions.sql.impl.rel.BeamRelNode
-
A
BeamRelNode
is a recursive structure, the
BeamQueryPlanner
visits it with a DFS(Depth-First-Search)
algorithm.
- buildBeamPipeline(PCollectionTuple, BeamSqlEnv) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSetOperatorRelBase
-
- buildBeamPipeline(PCollectionTuple, BeamSqlEnv) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSortRel
-
- buildBeamPipeline(PCollectionTuple, BeamSqlEnv) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnionRel
-
- buildBeamPipeline(PCollectionTuple, BeamSqlEnv) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamValuesRel
-
- builder() - Static method in class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions
-
- Builder() - Constructor for class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions.Builder
-
- Builder() - Constructor for class org.apache.beam.sdk.io.fs.CreateOptions.Builder
-
- builder() - Static method in class org.apache.beam.sdk.io.fs.CreateOptions.StandardCreateOptions
-
- Builder() - Constructor for class org.apache.beam.sdk.io.fs.CreateOptions.StandardCreateOptions.Builder
-
- builder() - Static method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata
-
- Builder() - Constructor for class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
-
- Builder() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig.Builder
-
- Builder() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction.Builder
-
- builder() - Static method in class org.apache.beam.sdk.metrics.MetricsFilter
-
- Builder() - Constructor for class org.apache.beam.sdk.metrics.MetricsFilter.Builder
-
- buildIOReader(Pipeline) - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamPCollectionTable
-
- buildIOReader(Pipeline) - Method in interface org.apache.beam.sdk.extensions.sql.impl.schema.BeamSqlTable
-
create a PCollection<BeamSqlRow>
from source.
- buildIOReader(Pipeline) - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.kafka.BeamKafkaTable
-
- buildIOReader(Pipeline) - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.text.BeamTextCSVTable
-
- buildIOWriter() - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamPCollectionTable
-
- buildIOWriter() - Method in interface org.apache.beam.sdk.extensions.sql.impl.schema.BeamSqlTable
-
create a IO.write()
instance to write to target.
- buildIOWriter() - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.kafka.BeamKafkaTable
-
- buildIOWriter() - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.text.BeamTextCSVTable
-
- buildOutputFilenames(Iterable<FileBasedSink.FileResult<DestinationT>>) - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-
- buildTemporaryFilename(ResourceId, String) - Static method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-
Constructs a temporary file resource given the temporary directory and a filename.
- by(PredicateT) - Static method in class org.apache.beam.sdk.transforms.Filter
-
Returns a PTransform
that takes an input
PCollection<T>
and returns a PCollection<T>
with
elements that satisfy the given predicate.
- ByteArray - Class in org.apache.beam.runners.spark.util
-
Serializable byte array.
- ByteArray(byte[]) - Constructor for class org.apache.beam.runners.spark.util.ByteArray
-
- ByteArrayCoder - Class in org.apache.beam.sdk.coders
-
- ByteCoder - Class in org.apache.beam.sdk.coders
-
A
ByteCoder
encodes
Byte
values in 1 byte using Java serialization.
- ByteKey - Class in org.apache.beam.sdk.io.range
-
A class representing a key consisting of an array of bytes.
- ByteKeyRange - Class in org.apache.beam.sdk.io.range
-
A class representing a range of
ByteKeys
.
- ByteKeyRangeTracker - Class in org.apache.beam.sdk.io.range
-
- bytes() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-
- bytesRead() - Static method in class org.apache.beam.sdk.metrics.SourceMetrics
-
Counter of bytes read by a source.
- bytesReadBySplit(String) - Static method in class org.apache.beam.sdk.metrics.SourceMetrics
-
Counter of bytes read by a source split.
- ByteStringCoder - Class in org.apache.beam.sdk.extensions.protobuf
-
A
Coder
for
ByteString
objects based on their encoded Protocol Buffer form.
- bytesWritten() - Static method in class org.apache.beam.sdk.metrics.SinkMetrics
-
Counter of bytes written to a sink.
- calc(BigDecimal, BigDecimal) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlArithmeticExpression
-
- calc(BigDecimal, BigDecimal) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlDivideExpression
-
- calc(BigDecimal, BigDecimal) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlMinusExpression
-
- calc(BigDecimal, BigDecimal) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlModExpression
-
- calc(BigDecimal, BigDecimal) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlMultiplyExpression
-
- calc(BigDecimal, BigDecimal) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlPlusExpression
-
- CalciteUtils - Class in org.apache.beam.sdk.extensions.sql.impl.utils
-
Utility methods for Calcite related operations.
- CalciteUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlAbsExpression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlAcosExpression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlAsinExpression
-
- calculate(BeamSqlPrimitive, BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlAtan2Expression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlAtanExpression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlCeilExpression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlCosExpression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlCotExpression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlDegreesExpression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlExpExpression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlFloorExpression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlLnExpression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlLogExpression
-
- calculate(BeamSqlPrimitive, BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlMathBinaryExpression
-
The base method for implementation of math binary functions.
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlMathUnaryExpression
-
For the operands of other type SqlTypeName#NUMERIC_TYPES
.
- calculate(BeamSqlPrimitive, BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlPowerExpression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlRadiansExpression
-
- calculate(BeamSqlPrimitive, BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlRoundExpression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlSignExpression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlSinExpression
-
- calculate(BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlTanExpression
-
- calculate(BeamSqlPrimitive, BeamSqlPrimitive) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlTruncateExpression
-
- CalendarWindows - Class in org.apache.beam.sdk.transforms.windowing
-
A collection of
WindowFn
s that windows values into calendar-based
windows such as spans of days, months, or years.
- CalendarWindows() - Constructor for class org.apache.beam.sdk.transforms.windowing.CalendarWindows
-
- CalendarWindows.DaysWindows - Class in org.apache.beam.sdk.transforms.windowing
-
A
WindowFn
that windows elements into periods measured by days.
- CalendarWindows.MonthsWindows - Class in org.apache.beam.sdk.transforms.windowing
-
A
WindowFn
that windows elements into periods measured by months.
- CalendarWindows.YearsWindows - Class in org.apache.beam.sdk.transforms.windowing
-
A
WindowFn
that windows elements into periods measured by years.
- cancel() - Method in class org.apache.beam.runners.apex.ApexRunnerResult
-
- cancel() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-
- cancel() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
-
- cancel() - Method in class org.apache.beam.runners.direct.DirectRunner.DirectPipelineResult
-
- cancel() - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
-
- cancel() - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
-
- cancel() - Method in class org.apache.beam.runners.gearpump.GearpumpPipelineResult
-
- cancel(JobApi.CancelJobRequest, StreamObserver<JobApi.CancelJobResponse>) - Method in class org.apache.beam.runners.reference.job.ReferenceRunnerJobService
-
- cancel() - Method in class org.apache.beam.runners.spark.SparkPipelineResult
-
- cancel() - Method in interface org.apache.beam.sdk.PipelineResult
-
Cancels the pipeline execution.
- canConvertConvention(Convention) - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-
- CannotProvideCoderException - Exception in org.apache.beam.sdk.coders
-
- CannotProvideCoderException(String) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
-
- CannotProvideCoderException(String, CannotProvideCoderException.ReasonCode) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
-
- CannotProvideCoderException(String, Throwable) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
-
- CannotProvideCoderException(String, Throwable, CannotProvideCoderException.ReasonCode) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
-
- CannotProvideCoderException(Throwable) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
-
- CannotProvideCoderException(Throwable, CannotProvideCoderException.ReasonCode) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
-
- CannotProvideCoderException.ReasonCode - Enum in org.apache.beam.sdk.coders
-
Indicates the reason that
Coder
inference failed.
- canStopPolling(Instant, StateT) - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
-
- CassandraIO - Class in org.apache.beam.sdk.io.cassandra
-
An IO to read from Apache Cassandra.
- CassandraIO.Read<T> - Class in org.apache.beam.sdk.io.cassandra
-
- CassandraIO.Write<T> - Class in org.apache.beam.sdk.io.cassandra
-
- CassandraService<T> - Interface in org.apache.beam.sdk.io.cassandra
-
An interface for real or fake implementations of Cassandra.
- CassandraService.Writer<T> - Interface in org.apache.beam.sdk.io.cassandra
-
Writer for an entity.
- CassandraServiceImpl<T> - Class in org.apache.beam.sdk.io.cassandra
-
An implementation of the
CassandraService
that actually use a Cassandra instance.
- CassandraServiceImpl() - Constructor for class org.apache.beam.sdk.io.cassandra.CassandraServiceImpl
-
- CassandraServiceImpl.TokenRange - Class in org.apache.beam.sdk.io.cassandra
-
Represent a token range in Cassandra instance, wrapping the partition count, size and token
range.
- CassandraServiceImpl.WriterImpl<T> - Class in org.apache.beam.sdk.io.cassandra
-
Writer storing an entity into Apache Cassandra database.
- characters() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-
- checkDone() - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
-
- checkDone() - Method in interface org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker
-
Called by the runner after DoFn.ProcessElement
returns.
- checkpoint() - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
-
- checkpoint() - Method in interface org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker
-
Signals that the current
DoFn.ProcessElement
call should terminate as soon as possible:
after this method returns, the tracker MUST refuse all future claim calls, and
RestrictionTracker.checkDone()
MUST succeed.
- classesToTranslators() - Method in interface org.apache.beam.runners.dataflow.util.CoderCloudObjectTranslatorRegistrar
-
- classesToTranslators() - Method in class org.apache.beam.runners.dataflow.util.DefaultCoderCloudObjectTranslatorRegistrar
-
- classNamesToTranslators() - Method in interface org.apache.beam.runners.dataflow.util.CoderCloudObjectTranslatorRegistrar
-
- classNamesToTranslators() - Method in class org.apache.beam.runners.dataflow.util.DefaultCoderCloudObjectTranslatorRegistrar
-
- CLASSPATH_SCHEME - Static variable in class org.apache.beam.runners.apex.ApexRunner
-
- cleanup() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
-
- clear() - Static method in class org.apache.beam.runners.spark.aggregators.AggregatorsAccumulator
-
- clear() - Static method in class org.apache.beam.runners.spark.metrics.MetricsAccumulator
-
- clear() - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
-
- clear() - Method in interface org.apache.beam.sdk.state.State
-
Clear out the state location.
- clearCache() - Static method in class org.apache.beam.runners.spark.io.MicrobatchSource
-
- clearOutputElements() - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
Clears the record of the elements output so far to the main output.
- clearOutputElements(TupleTag<T>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
Clears the record of the elements output so far to the output with the given tag.
- clone() - Method in class org.apache.beam.runners.dataflow.util.CloudObject
-
- clone() - Method in class org.apache.beam.runners.flink.metrics.MetricsAccumulator
-
- close() - Method in class org.apache.beam.runners.gearpump.translators.io.GearpumpSource
-
- close() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-
- close() - Method in interface org.apache.beam.sdk.extensions.sql.impl.interpreter.BeamSqlExpressionExecutor
-
- close() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.BeamSqlFnExecutor
-
- close() - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSqlFilterFn
-
- close() - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSqlProjectFn
-
- close() - Method in class org.apache.beam.sdk.io.cassandra.CassandraServiceImpl.WriterImpl
-
- close() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
-
Closes the channel and returns the bundle result.
- close() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
-
Closes any ReadableByteChannel
created for the current reader.
- close() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerAccessor
-
- close() - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantDeserializer
-
- close() - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantSerializer
-
- close() - Method in class org.apache.beam.sdk.io.Source.Reader
-
Closes the reader.
- close() - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
- CloudDebuggerOptions - Interface in org.apache.beam.runners.dataflow.options
-
Options for controlling Cloud Debugger.
- CloudObject - Class in org.apache.beam.runners.dataflow.util
-
A representation of an arbitrary Java object to be instantiated by Dataflow
workers.
- cloudObjectClassName() - Method in interface org.apache.beam.runners.dataflow.util.CloudObjectTranslator
-
- CloudObjects - Class in org.apache.beam.runners.dataflow.util
-
- CloudObjectTranslator<T> - Interface in org.apache.beam.runners.dataflow.util
-
A translator that takes an object and creates a
CloudObject
which can be converted back
to the original object.
- CloudResourceManagerOptions - Interface in org.apache.beam.sdk.extensions.gcp.options
-
Properties needed when using Google CloudResourceManager with the Apache Beam SDK.
- CO_GBK_RESULT_SCHEMA - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- Coder<T> - Class in org.apache.beam.sdk.coders
-
A
Coder<T>
defines how to encode and decode values of type
T
into
byte streams.
- Coder() - Constructor for class org.apache.beam.sdk.coders.Coder
-
- Coder.Context - Class in org.apache.beam.sdk.coders
-
- Coder.NonDeterministicException - Exception in org.apache.beam.sdk.coders
-
Exception thrown by
Coder.verifyDeterministic()
if the encoding is
not deterministic, including details of why the encoding is not deterministic.
- CoderCloudObjectTranslatorRegistrar - Interface in org.apache.beam.runners.dataflow.util
-
Coder
authors have the ability to automatically have their
Coder
registered with
the Dataflow Runner by creating a
ServiceLoader
entry and a concrete implementation of
this interface.
- coderConsistentWithEquals(Coder<T>, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that for the given Coder<T>
and values of
type T
, the values are equal if and only if the
encoded bytes are equal.
- coderConsistentWithEqualsInContext(Coder<T>, Coder.Context, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that for the given Coder<T>
, Coder.Context
, and
values of type T
, the values are equal if and only if the
encoded bytes are equal, in any Coder.Context
.
- coderDecodeEncodeContentsEqual(Coder<CollectionT>, CollectionT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that for the given Coder<Collection<T>>
,
and value of type Collection<T>
, encoding followed by decoding yields an
equal value of type Collection<T>
, in any Coder.Context
.
- coderDecodeEncodeContentsEqualInContext(Coder<CollectionT>, Coder.Context, CollectionT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that for the given Coder<Collection<T>>
,
and value of type Collection<T>
, encoding followed by decoding yields an
equal value of type Collection<T>
, in the given Coder.Context
.
- coderDecodeEncodeContentsInSameOrder(Coder<IterableT>, IterableT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that for the given Coder<Collection<T>>
,
and value of type Collection<T>
, encoding followed by decoding yields an
equal value of type Collection<T>
, in any Coder.Context
.
- coderDecodeEncodeContentsInSameOrderInContext(Coder<IterableT>, Coder.Context, IterableT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that for the given Coder<Iterable<T>>
,
and value of type Iterable<T>
, encoding followed by decoding yields an
equal value of type Collection<T>
, in the given Coder.Context
.
- coderDecodeEncodeEqual(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that for the given Coder<T>
,
and value of type T
, encoding followed by decoding yields an
equal value of type T
, in any Coder.Context
.
- coderDecodeEncodeEqualInContext(Coder<T>, Coder.Context, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that for the given Coder<T>
, Coder.Context
,
and value of type T
, encoding followed by decoding yields an
equal value of type T
.
- coderDecodesBase64(Coder<T>, String, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
- coderDecodesBase64(Coder<T>, List<String>, List<T>) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
- coderDecodesBase64ContentsEqual(Coder<IterableT>, String, IterableT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
- coderDecodesBase64ContentsEqual(Coder<IterableT>, List<String>, List<IterableT>) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
- coderDeterministic(Coder<T>, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that for the given Coder<T>
, and values of
type T
, if the values are equal then the encoded bytes are equal, in any
Coder.Context
.
- coderDeterministicInContext(Coder<T>, Coder.Context, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that for the given Coder<T>
, Coder.Context
, and values of
type T
, if the values are equal then the encoded bytes are equal.
- coderEncodesBase64(Coder<T>, T, String) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
- coderEncodesBase64(Coder<T>, List<T>, List<String>) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
- CoderException - Exception in org.apache.beam.sdk.coders
-
An Exception
thrown if there is a problem encoding or decoding a value.
- CoderException(String) - Constructor for exception org.apache.beam.sdk.coders.CoderException
-
- CoderException(String, Throwable) - Constructor for exception org.apache.beam.sdk.coders.CoderException
-
- CoderException(Throwable) - Constructor for exception org.apache.beam.sdk.coders.CoderException
-
- coderFor(TypeDescriptor<T>, List<? extends Coder<?>>) - Method in class org.apache.beam.sdk.coders.CoderProvider
-
Returns a Coder<T>
to use for values of a particular type, given the Coders for each of
the type's generic parameter types.
- coderFromCloudObject(CloudObject) - Static method in class org.apache.beam.runners.dataflow.util.CloudObjects
-
- CoderHelpers - Class in org.apache.beam.runners.spark.coders
-
Serialization utility class.
- CoderProperties - Class in org.apache.beam.sdk.testing
-
Properties for use in
Coder
tests.
- CoderProperties() - Constructor for class org.apache.beam.sdk.testing.CoderProperties
-
- CoderProperties.TestElementByteSizeObserver - Class in org.apache.beam.sdk.testing
-
An ElementByteSizeObserver
that records the observed element sizes for testing
purposes.
- CoderProvider - Class in org.apache.beam.sdk.coders
-
- CoderProvider() - Constructor for class org.apache.beam.sdk.coders.CoderProvider
-
- CoderProviderRegistrar - Interface in org.apache.beam.sdk.coders
-
Coder
creators have the ability to automatically have their
coders
registered with this SDK by creating a
ServiceLoader
entry
and a concrete implementation of this interface.
- CoderProviders - Class in org.apache.beam.sdk.coders
-
Static utility methods for creating and working with
CoderProvider
s.
- CoderRegistry - Class in org.apache.beam.sdk.coders
-
- coderSerializable(Coder<T>) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that the given Coder<T>
can be correctly serialized and
deserialized.
- CoGbkResult - Class in org.apache.beam.sdk.transforms.join
-
- CoGbkResult(CoGbkResultSchema, Iterable<RawUnionValue>) - Constructor for class org.apache.beam.sdk.transforms.join.CoGbkResult
-
- CoGbkResult(CoGbkResultSchema, Iterable<RawUnionValue>, int) - Constructor for class org.apache.beam.sdk.transforms.join.CoGbkResult
-
- CoGbkResult.CoGbkResultCoder - Class in org.apache.beam.sdk.transforms.join
-
- CoGbkResultSchema - Class in org.apache.beam.sdk.transforms.join
-
- CoGbkResultSchema(TupleTagList) - Constructor for class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
-
Builds a schema from a tuple of TupleTag<?>
s.
- CoGroupByKey<K> - Class in org.apache.beam.sdk.transforms.join
-
- CollectionCoder<T> - Class in org.apache.beam.sdk.coders
-
- CollectionCoder(Coder<T>) - Constructor for class org.apache.beam.sdk.coders.CollectionCoder
-
- Combine - Class in org.apache.beam.sdk.transforms
-
PTransform
s for combining PCollection
elements
globally and per-key.
- combine(Iterable<? extends Instant>) - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
-
- combine(Instant...) - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
-
- Combine.AccumulatingCombineFn<InputT,AccumT extends Combine.AccumulatingCombineFn.Accumulator<InputT,AccumT,OutputT>,OutputT> - Class in org.apache.beam.sdk.transforms
-
- Combine.AccumulatingCombineFn.Accumulator<InputT,AccumT,OutputT> - Interface in org.apache.beam.sdk.transforms
-
The type of mutable accumulator values used by this
AccumulatingCombineFn
.
- Combine.BinaryCombineDoubleFn - Class in org.apache.beam.sdk.transforms
-
An abstract subclass of
Combine.CombineFn
for implementing combiners that are more
easily and efficiently expressed as binary operations on
double
s.
- Combine.BinaryCombineFn<V> - Class in org.apache.beam.sdk.transforms
-
An abstract subclass of
Combine.CombineFn
for implementing combiners that are more
easily expressed as binary operations.
- Combine.BinaryCombineIntegerFn - Class in org.apache.beam.sdk.transforms
-
An abstract subclass of
Combine.CombineFn
for implementing combiners that are more
easily and efficiently expressed as binary operations on
int
s
- Combine.BinaryCombineLongFn - Class in org.apache.beam.sdk.transforms
-
An abstract subclass of
Combine.CombineFn
for implementing combiners that are more
easily and efficiently expressed as binary operations on
long
s.
- Combine.CombineFn<InputT,AccumT,OutputT> - Class in org.apache.beam.sdk.transforms
-
A CombineFn<InputT, AccumT, OutputT>
specifies how to combine a
collection of input values of type InputT
into a single
output value of type OutputT
.
- Combine.Globally<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
-
Combine.Globally<InputT, OutputT>
takes a
PCollection<InputT>
and returns a
PCollection<OutputT>
whose elements are the result of
combining all the elements in each window of the input
PCollection
,
using a specified
CombineFn<InputT, AccumT, OutputT>
.
- Combine.GloballyAsSingletonView<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
-
Combine.GloballyAsSingletonView<InputT, OutputT>
takes a
PCollection<InputT>
and returns a
PCollectionView<OutputT>
whose elements are the result of
combining all the elements in each window of the input
PCollection
,
using a specified
CombineFn<InputT, AccumT, OutputT>
.
- Combine.GroupedValues<K,InputT,OutputT> - Class in org.apache.beam.sdk.transforms
-
GroupedValues<K, InputT, OutputT>
takes a
PCollection<KV<K, Iterable<InputT>>>
,
such as the result of
GroupByKey
, applies a specified
CombineFn<InputT, AccumT, OutputT>
to each of the input
KV<K,
Iterable<InputT>>
elements to produce a combined output
KV<K, OutputT>
element, and
returns a
PCollection<KV<K, OutputT>>
containing all the combined output elements.
- Combine.Holder<V> - Class in org.apache.beam.sdk.transforms
-
Holds a single value value of type V
which may or may not be present.
- Combine.IterableCombineFn<V> - Class in org.apache.beam.sdk.transforms
-
- Combine.PerKey<K,InputT,OutputT> - Class in org.apache.beam.sdk.transforms
-
PerKey<K, InputT, OutputT>
takes a
PCollection<KV<K, InputT>>
, groups it by key, applies a
combining function to the InputT
values associated with each
key to produce a combined OutputT
value, and returns a
PCollection<KV<K, OutputT>>
representing a map from each
distinct key of the input PCollection
to the corresponding
combined value.
- Combine.PerKeyWithHotKeyFanout<K,InputT,OutputT> - Class in org.apache.beam.sdk.transforms
-
- Combine.SimpleCombineFn<V> - Class in org.apache.beam.sdk.transforms
-
Deprecated.
- CombineFn() - Constructor for class org.apache.beam.sdk.transforms.Combine.CombineFn
-
- combineFn() - Static method in class org.apache.beam.sdk.transforms.Count
-
- combineFn() - Static method in class org.apache.beam.sdk.transforms.Latest
-
- combineFn(int) - Static method in class org.apache.beam.sdk.transforms.Sample
-
- CombineFnBase - Class in org.apache.beam.sdk.transforms
-
For internal use only; no backwards-compatibility guarantees.
- CombineFnBase() - Constructor for class org.apache.beam.sdk.transforms.CombineFnBase
-
- CombineFnBase.GlobalCombineFn<InputT,AccumT,OutputT> - Interface in org.apache.beam.sdk.transforms
-
For internal use only; no backwards-compatibility guarantees.
- CombineFns - Class in org.apache.beam.sdk.transforms
-
Static utility methods that create combine function instances.
- CombineFns() - Constructor for class org.apache.beam.sdk.transforms.CombineFns
-
- CombineFns.CoCombineResult - Class in org.apache.beam.sdk.transforms
-
A tuple of outputs produced by a composed combine functions.
- CombineFns.ComposeCombineFnBuilder - Class in org.apache.beam.sdk.transforms
-
- CombineFns.ComposedCombineFn<DataT> - Class in org.apache.beam.sdk.transforms
-
- CombineFns.ComposedCombineFnWithContext<DataT> - Class in org.apache.beam.sdk.transforms
-
- CombineFnTester - Class in org.apache.beam.sdk.testing
-
- CombineFnTester() - Constructor for class org.apache.beam.sdk.testing.CombineFnTester
-
- CombineFnWithContext() - Constructor for class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
-
- CombineWithContext - Class in org.apache.beam.sdk.transforms
-
This class contains combine functions that have access to PipelineOptions
and side inputs
through CombineWithContext.Context
.
- CombineWithContext() - Constructor for class org.apache.beam.sdk.transforms.CombineWithContext
-
- CombineWithContext.CombineFnWithContext<InputT,AccumT,OutputT> - Class in org.apache.beam.sdk.transforms
-
A combine function that has access to PipelineOptions
and side inputs through
CombineWithContext.Context
.
- CombineWithContext.Context - Class in org.apache.beam.sdk.transforms
-
Information accessible to all methods in CombineFnWithContext
and KeyedCombineFnWithContext
.
- CombineWithContext.RequiresContextInternal - Interface in org.apache.beam.sdk.transforms
-
An internal interface for signaling that a GloballyCombineFn
or a PerKeyCombineFn
needs to access CombineWithContext.Context
.
- combining(Combine.CombineFn<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
-
- combining(CombineWithContext.CombineFnWithContext<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
-
For internal use only; no backwards compatibility guarantees
- combining(Coder<AccumT>, Combine.CombineFn<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
-
Identical to #combining(CombineFn)
, but with an accumulator coder explicitly supplied.
- combining(Coder<AccumT>, CombineWithContext.CombineFnWithContext<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
-
For internal use only; no backwards compatibility guarantees
- combiningFromInputInternal(Coder<InputT>, Combine.CombineFn<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
-
For internal use only; no backwards-compatibility guarantees.
- CombiningState<InputT,AccumT,OutputT> - Interface in org.apache.beam.sdk.state
-
A
ReadableState
cell defined by a
Combine.CombineFn
, accepting multiple input values,
combining them as specified into accumulators, and producing a single output value.
- commitManifest(ArtifactApi.CommitManifestRequest, StreamObserver<ArtifactApi.CommitManifestResponse>) - Method in class org.apache.beam.artifact.local.LocalFileSystemArtifactStagerService
-
- committed() - Method in interface org.apache.beam.sdk.metrics.MetricResult
-
Return the value of this metric across all successfully completed parts of the pipeline.
- commonPrefixLength(RandomAccessData, RandomAccessData) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.UnsignedLexicographicalComparator
-
Compute the length of the common prefix of the two provided sets of bytes.
- compact(AccumT) - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
-
Returns an accumulator that represents the same logical value as the
input accumulator, but may have a more compact representation.
- compact(List<V>) - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
-
- compact(Object[]) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
-
- compact(Object[], CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
-
- compact(AccumT, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
-
Returns an accumulator that represents the same logical value as the
input accumulator, but may have a more compact representation.
- compare(JobMessage, JobMessage) - Method in class org.apache.beam.runners.dataflow.util.MonitoringUtil.TimeStampComparator
-
- compare(RandomAccessData, RandomAccessData) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.UnsignedLexicographicalComparator
-
- compare(RandomAccessData, RandomAccessData, int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.UnsignedLexicographicalComparator
-
Compare the two sets of bytes starting at the given offset.
- compare(CharSequence, CharSequence) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlCompareExpression
-
Compare between String values, mapping to SqlTypeName#VARCHAR
.
- compare(Boolean, Boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlCompareExpression
-
Compare between Boolean values, mapping to SqlTypeName#BOOLEAN
.
- compare(Number, Number) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlCompareExpression
-
Compare between Number values, including SqlTypeName#BIGINT
,
SqlTypeName#DECIMAL
, SqlTypeName#DOUBLE
, SqlTypeName#FLOAT
,
SqlTypeName#INTEGER
, SqlTypeName#SMALLINT
and SqlTypeName#TINYINT
.
- compare(CharSequence, CharSequence) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlEqualsExpression
-
- compare(Boolean, Boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlEqualsExpression
-
- compare(Number, Number) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlEqualsExpression
-
- compare(CharSequence, CharSequence) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlGreaterThanExpression
-
- compare(Boolean, Boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlGreaterThanExpression
-
- compare(Number, Number) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlGreaterThanExpression
-
- compare(CharSequence, CharSequence) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlGreaterThanOrEqualsExpression
-
- compare(Boolean, Boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlGreaterThanOrEqualsExpression
-
- compare(Number, Number) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlGreaterThanOrEqualsExpression
-
- compare(CharSequence, CharSequence) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlLessThanExpression
-
- compare(Boolean, Boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlLessThanExpression
-
- compare(Number, Number) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlLessThanExpression
-
- compare(CharSequence, CharSequence) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlLessThanOrEqualsExpression
-
- compare(Boolean, Boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlLessThanOrEqualsExpression
-
- compare(Number, Number) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlLessThanOrEqualsExpression
-
- compare(CharSequence, CharSequence) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlNotEqualsExpression
-
- compare(Boolean, Boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlNotEqualsExpression
-
- compare(Number, Number) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison.BeamSqlNotEqualsExpression
-
- compare(T, T) - Method in class org.apache.beam.sdk.transforms.Top.Largest
-
Deprecated.
- compare(T, T) - Method in class org.apache.beam.sdk.transforms.Top.Natural
-
- compare(T, T) - Method in class org.apache.beam.sdk.transforms.Top.Reversed
-
- compare(T, T) - Method in class org.apache.beam.sdk.transforms.Top.Smallest
-
Deprecated.
- compare(KV<K, V>, KV<K, V>) - Method in class org.apache.beam.sdk.values.KV.OrderByKey
-
- compare(KV<K, V>, KV<K, V>) - Method in class org.apache.beam.sdk.values.KV.OrderByValue
-
- compareTo(ByteArray) - Method in class org.apache.beam.runners.spark.util.ByteArray
-
- compareTo(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKey
-
ByteKey
implements
Comparable<ByteKey>
by comparing the
arrays in lexicographic order.
- compareTo(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
-
- compileBeamPipeline(String, Pipeline, BeamSqlEnv) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamQueryPlanner
-
compileBeamPipeline
translate a SQL statement to executed as Beam data flow,
which is linked with the given pipeline
.
- compilePipeline(String, BeamSqlEnv) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlCli
-
- compilePipeline(String, Pipeline, BeamSqlEnv) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlCli
-
- complete(List<TimestampedValue<OutputT>>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
-
Constructs a
Watch.Growth.PollResult
with the given outputs and declares that there will be no
new outputs for the current input.
- complete(Instant, List<OutputT>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
-
- COMPONENT_ENCODINGS - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- compose() - Static method in class org.apache.beam.sdk.transforms.CombineFns
-
- ComposeCombineFnBuilder() - Constructor for class org.apache.beam.sdk.transforms.CombineFns.ComposeCombineFnBuilder
-
- CompositeSource - Class in org.apache.beam.runners.spark.metrics
-
Composite source made up of several MetricRegistry
instances.
- CompositeSource(String, MetricRegistry...) - Constructor for class org.apache.beam.runners.spark.metrics.CompositeSource
-
- CompressedReader(CompressedSource<T>, FileBasedSource.FileBasedReader<T>) - Constructor for class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-
Create a CompressedReader
from a CompressedSource
and delegate reader.
- CompressedSource<T> - Class in org.apache.beam.sdk.io
-
A Source that reads from compressed files.
- CompressedSource.CompressedReader<T> - Class in org.apache.beam.sdk.io
-
- CompressedSource.CompressionMode - Enum in org.apache.beam.sdk.io
-
- CompressedSource.DecompressingChannelFactory - Interface in org.apache.beam.sdk.io
-
Factory interface for creating channels that decompress the content of an underlying channel.
- Compression - Enum in org.apache.beam.sdk.io
-
Various compression types for reading/writing files.
- compute(Partition, TaskContext) - Method in class org.apache.beam.runners.spark.io.SourceRDD.Bounded
-
- compute(Partition, TaskContext) - Method in class org.apache.beam.runners.spark.io.SourceRDD.Unbounded
-
- ConfigurationLocator() - Constructor for class org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptions.ConfigurationLocator
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantSerializer
-
- configure() - Static method in class org.apache.beam.sdk.transforms.windowing.Window
-
Returns a new builder for a
Window
transform for setting windowing parameters other
than the windowing function.
- connect() - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
-
Connect to the Redis instance.
- ConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-
- ConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.ConnectionConfiguration
-
- ConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
-
- ConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.solr.SolrIO.ConnectionConfiguration
-
- connectToSpanner() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-
- consistentWithEquals() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
-
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
-
Returns
true
if this
Coder
is injective with respect to
Object.equals(java.lang.Object)
.
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
-
Returns
true
if this
Coder
is injective with respect to
Object.equals(java.lang.Object)
.
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
-
Returns
true
if this
Coder
is injective with respect to
Object.equals(java.lang.Object)
.
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
-
Returns
true
if this
Coder
is injective with respect to
Object.equals(java.lang.Object)
.
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BitSetCoder
-
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BooleanCoder
-
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.ByteCoder
-
Returns
true
if this
Coder
is injective with respect to
Object.equals(java.lang.Object)
.
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.Coder
-
Returns
true
if this
Coder
is injective with respect to
Object.equals(java.lang.Object)
.
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.DoubleCoder
-
Returns
true
if this
Coder
is injective with respect to
Object.equals(java.lang.Object)
.
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.DurationCoder
-
Returns
true
if this
Coder
is injective with respect to
Object.equals(java.lang.Object)
.
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.InstantCoder
-
Returns
true
if this
Coder
is injective with respect to
Object.equals(java.lang.Object)
.
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.KvCoder
-
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
-
LengthPrefixCoder
is consistent with equals if the nested Coder
is.
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.NullableCoder
-
NullableCoder
is consistent with equals if the nested Coder
is.
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
-
Returns
true
if this
Coder
is injective with respect to
Object.equals(java.lang.Object)
.
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.VarIntCoder
-
Returns
true
if this
Coder
is injective with respect to
Object.equals(java.lang.Object)
.
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.VarLongCoder
-
Returns
true
if this
Coder
is injective with respect to
Object.equals(java.lang.Object)
.
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.VoidCoder
-
Returns
true
if this
Coder
is injective with respect to
Object.equals(java.lang.Object)
.
- consistentWithEquals() - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
-
Returns
true
if this
Coder
is injective with respect to
Object.equals(java.lang.Object)
.
- consistentWithEquals() - Method in class org.apache.beam.sdk.io.fs.MetadataCoder
-
- consistentWithEquals() - Method in class org.apache.beam.sdk.io.fs.ResourceIdCoder
-
- consistentWithEquals() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
-
- consistentWithEquals() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
-
- consistentWithEquals() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
-
- ConsoleIO - Class in org.apache.beam.runners.spark.io
-
Print to console.
- ConsoleIO.Write - Class in org.apache.beam.runners.spark.io
-
Write on the console.
- ConsoleIO.Write.Unbound<T> - Class in org.apache.beam.runners.spark.io
-
- constant(FileBasedSink.FilenamePolicy, SerializableFunction<UserT, OutputT>) - Static method in class org.apache.beam.sdk.io.DynamicFileDestinations
-
- constant(FileBasedSink.FilenamePolicy) - Static method in class org.apache.beam.sdk.io.DynamicFileDestinations
-
A specialization of #constant(FilenamePolicy, SerializableFunction)
for the case where
UserT and OutputT are the same type and the format function is the identity.
- constant(OutT) - Static method in class org.apache.beam.sdk.transforms.SerializableFunctions
-
- constantDestinations(FileBasedSink.FilenamePolicy, Schema, Map<String, Object>, CodecFactory, SerializableFunction<UserT, OutputT>) - Static method in class org.apache.beam.sdk.io.AvroIO
-
- contains(PCollectionView<T>) - Method in class org.apache.beam.runners.spark.util.SparkSideInputReader
-
- contains(T) - Method in interface org.apache.beam.sdk.state.SetState
-
Returns true if this set contains the specified element.
- contains(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
-
Returns whether this window contains the given window.
- containsInAnyOrder(T...) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
-
Asserts that the iterable in question contains the provided elements.
- containsInAnyOrder(Iterable<T>) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
-
Asserts that the iterable in question contains the provided elements.
- containsInAnyOrder(T...) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-
Checks that the Iterable
contains the expected elements, in any order.
- containsInAnyOrder(Iterable<T>) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-
Checks that the Iterable
contains the expected elements, in any order.
- containsKey(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
-
Returns
true
if the specified
ByteKey
is contained within this range.
- Context(boolean) - Constructor for class org.apache.beam.sdk.coders.Coder.Context
-
Deprecated.
- Context(TableDataInsertAllResponse.InsertErrors) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy.Context
-
- Context() - Constructor for class org.apache.beam.sdk.transforms.CombineWithContext.Context
-
- continuously(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.FileIO.Match
-
- continuously(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.FileIO.MatchAll
-
- continuously(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
-
Continuously watches for new files at the given interval until the given termination
condition is reached, where the input to the condition is the filepattern.
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamFilterRule
-
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIntersectRule
-
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIOSinkRule
-
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIOSourceRule
-
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamJoinRule
-
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamMinusRule
-
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamProjectRule
-
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSortRule
-
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamUnionRule
-
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamValuesRule
-
- convertToArgs(PipelineOptions) - Static method in class org.apache.beam.sdk.testing.TestPipeline
-
- convertToBagSpecInternal(StateSpec<CombiningState<InputT, AccumT, OutputT>>) - Static method in class org.apache.beam.sdk.state.StateSpecs
-
For internal use only; no backwards-compatibility guarantees.
- convertToBeamRel(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamQueryPlanner
-
It parses and validate the input query, then convert into a
BeamRelNode
tree.
- convertToFileResourceIfPossible(String) - Static method in class org.apache.beam.sdk.io.FileBasedSink
-
This is a helper function for turning a user-provided output filename prefix and converting it
into a
ResourceId
for writing output files.
- copy() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
-
Returns a copy of this RandomAccessData.
- copy(RelTraitSet, RelNode, boolean, ImmutableBitSet, List<ImmutableBitSet>, List<AggregateCall>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamAggregationRel
-
- copy(RelTraitSet, RelNode, RexNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamFilterRel
-
- copy(RelTraitSet, List<RelNode>, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIntersectRel
-
- copy(RelTraitSet, List<RelNode>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
-
- copy(RelTraitSet, RexNode, RelNode, RelNode, JoinRelType, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
-
- copy(RelTraitSet, List<RelNode>, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamMinusRel
-
- copy(RelTraitSet, RelNode, List<RexNode>, RelDataType) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamProjectRel
-
- copy(RelTraitSet, RelNode, RelCollation, RexNode, RexNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSortRel
-
- copy(RelTraitSet, List<RelNode>, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnionRel
-
- copy(List<ResourceIdT>, List<ResourceIdT>) - Method in class org.apache.beam.sdk.io.FileSystem
-
Copies a List
of file-like resources from one location to another.
- copy(List<ResourceId>, List<ResourceId>, MoveOptions...) - Static method in class org.apache.beam.sdk.io.FileSystems
-
Copies a List
of file-like resources from one location to another.
- copyFrom(ByteBuffer) - Static method in class org.apache.beam.sdk.io.range.ByteKey
-
Creates a new
ByteKey
backed by a copy of the data remaining in the specified
ByteBuffer
.
- copyFrom(byte[]) - Static method in class org.apache.beam.sdk.io.range.ByteKey
-
Creates a new
ByteKey
backed by a copy of the specified
byte[]
.
- count() - Method in class org.apache.beam.sdk.metrics.DistributionResult
-
- Count - Class in org.apache.beam.sdk.transforms
-
- countAsserts(Pipeline) - Static method in class org.apache.beam.sdk.testing.PAssert
-
- Counter - Interface in org.apache.beam.sdk.metrics
-
A metric that reports a single long value and can be incremented or decremented.
- counter(String, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
-
Create a metric that can be incremented and decremented, and is aggregated by taking the sum.
- counter(Class<?>, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
-
Create a metric that can be incremented and decremented, and is aggregated by taking the sum.
- CounterMark(long, Instant) - Constructor for class org.apache.beam.sdk.io.CountingSource.CounterMark
-
Creates a checkpoint mark reflecting the last emitted value.
- counters() - Method in interface org.apache.beam.sdk.metrics.MetricQueryResults
-
Return the metric results for the counters that matched the filter.
- CountingSource - Class in org.apache.beam.sdk.io
-
- CountingSource.CounterMark - Class in org.apache.beam.sdk.io
-
The checkpoint for an unbounded
CountingSource
is simply the last value produced.
- CrashingRunner - Class in org.apache.beam.sdk.testing
-
- CrashingRunner() - Constructor for class org.apache.beam.sdk.testing.CrashingRunner
-
- create(DataflowPipelineOptions) - Static method in class org.apache.beam.runners.dataflow.DataflowClient
-
- create(PCollectionView<?>, Coder<T>) - Static method in class org.apache.beam.runners.dataflow.DataflowRunner.StreamingPCollectionViewWriterFn
-
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions.DataflowClientFactory
-
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions.StagerFactory
-
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowPipelineOptions.StagingLocationFactory
-
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions.WorkerHarnessContainerImageFactory
-
- create(PipelineOptions) - Method in class org.apache.beam.runners.direct.DirectOptions.AvailableParallelismFactory
-
- create(PipelineOptions) - Method in class org.apache.beam.runners.flink.DefaultParallelismFactory
-
- create(boolean) - Static method in class org.apache.beam.runners.flink.TestFlinkRunner
-
- create() - Static method in class org.apache.beam.runners.reference.job.ReferenceRunnerJobService
-
- create(PipelineOptions) - Method in class org.apache.beam.runners.spark.SparkContextOptions.EmptyListenersList
-
- create(PipelineOptions) - Method in class org.apache.beam.runners.spark.SparkPipelineOptions.TmpCheckpointDirFactory
-
- create() - Static method in class org.apache.beam.runners.spark.SparkRunner
-
Creates and returns a new SparkRunner with default options.
- create(SparkPipelineOptions) - Static method in class org.apache.beam.runners.spark.SparkRunner
-
Creates and returns a new SparkRunner with specified options.
- create(PipelineOptions) - Method in class org.apache.beam.runners.spark.TestSparkPipelineOptions.DefaultStopPipelineWatermarkFactory
-
- create(byte[], Coder<T>) - Static method in class org.apache.beam.runners.spark.util.SideInputBroadcast
-
- create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.DefaultProjectFactory
-
- create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpTempLocationFactory
-
- create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpUserCredentialsFactory
-
- create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcsOptions.ExecutorServiceFactory
-
- create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcsOptions.PathValidatorFactory
-
- create(BufferedExternalSorter.Options) - Static method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter
-
- create(BufferedExternalSorter.Options) - Static method in class org.apache.beam.sdk.extensions.sorter.SortValues
-
Returns a
SortValues<PrimaryKeyT, SecondaryKeyT, ValueT>
PTransform
.
- create(List<String>, List<Integer>) - Static method in class org.apache.beam.sdk.extensions.sql.BeamRecordSqlType
-
- create(String[], String, String) - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-
Creates a new Elasticsearch connection configuration.
- create(WritableByteChannel) - Method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
-
Deprecated.
- create(WritableByteChannel) - Method in interface org.apache.beam.sdk.io.FileBasedSink.WritableByteChannelFactory
-
- create(EmptyMatchTreatment) - Static method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
-
- create(ResourceIdT, CreateOptions) - Method in class org.apache.beam.sdk.io.FileSystem
-
Returns a write channel for the given ResourceIdT
.
- create(ResourceId, String) - Static method in class org.apache.beam.sdk.io.FileSystems
-
- create(ResourceId, CreateOptions) - Static method in class org.apache.beam.sdk.io.FileSystems
-
- create(MatchResult.Status, List<MatchResult.Metadata>) - Static method in class org.apache.beam.sdk.io.fs.MatchResult
-
- create(MatchResult.Status, IOException) - Static method in class org.apache.beam.sdk.io.fs.MatchResult
-
- create(Mutation, Mutation...) - Static method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
-
Creates a new group.
- create(Mutation, Iterable<Mutation>) - Static method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
-
- create() - Static method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-
- create() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-
- create(Timestamp) - Static method in class org.apache.beam.sdk.io.gcp.spanner.Transaction
-
- create(PipelineOptions) - Method in class org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptions.ConfigurationLocator
-
- create(DataSource) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
-
- create(String, String) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
-
- create(String, String) - Static method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
-
Describe a connection configuration to the MQTT broker.
- create(String, String, String) - Static method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
-
Describe a connection configuration to the MQTT broker.
- create() - Static method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
-
- create(String, int) - Static method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
-
- create(String) - Static method in class org.apache.beam.sdk.io.solr.SolrIO.ConnectionConfiguration
-
Creates a new Solr connection configuration.
- create(long, long, long, long) - Static method in class org.apache.beam.sdk.metrics.DistributionResult
-
- create(long, Instant) - Static method in class org.apache.beam.sdk.metrics.GaugeResult
-
- create(PipelineOptions) - Method in interface org.apache.beam.sdk.options.DefaultValueFactory
-
- create(PipelineOptions) - Method in class org.apache.beam.sdk.options.PipelineOptions.AtomicLongFactory
-
- create(PipelineOptions) - Method in class org.apache.beam.sdk.options.PipelineOptions.DirectRunner
-
- create(PipelineOptions) - Method in class org.apache.beam.sdk.options.PipelineOptions.JobNameFactory
-
- create(PipelineOptions) - Method in class org.apache.beam.sdk.options.PipelineOptions.UserAgentFactory
-
- create() - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
-
Creates and returns an object that implements
PipelineOptions
using the values
configured on this builder during construction.
- create() - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
-
- create() - Static method in class org.apache.beam.sdk.Pipeline
-
- create(PipelineOptions) - Static method in class org.apache.beam.sdk.Pipeline
-
- create() - Static method in class org.apache.beam.sdk.testing.TestPipeline
-
Creates and returns a new test pipeline.
- create(PipelineOptions) - Method in class org.apache.beam.sdk.testing.TestPipelineOptions.AlwaysPassMatcherFactory
-
- create(Coder<T>) - Static method in class org.apache.beam.sdk.testing.TestStream
-
- create(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
-
Returns an approximate quantiles combiner with the given
compareFn
and desired number of quantiles.
- create(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
-
- create(int, ComparatorT, long, double) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
-
Creates an approximate quantiles combiner with the given
compareFn
and desired number of quantiles.
- Create<T> - Class in org.apache.beam.sdk.transforms
-
Create<T>
takes a collection of elements of type T
known when the pipeline is constructed and returns a
PCollection<T>
containing the elements.
- Create() - Constructor for class org.apache.beam.sdk.transforms.Create
-
- create() - Static method in class org.apache.beam.sdk.transforms.Distinct
-
Returns a Distinct<T>
PTransform
.
- create() - Static method in class org.apache.beam.sdk.transforms.GroupByKey
-
Returns a GroupByKey<K, V>
PTransform
.
- create() - Static method in class org.apache.beam.sdk.transforms.join.CoGroupByKey
-
Returns a CoGroupByKey<K>
PTransform
.
- create() - Static method in class org.apache.beam.sdk.transforms.Keys
-
Returns a Keys<K>
PTransform
.
- create() - Static method in class org.apache.beam.sdk.transforms.KvSwap
-
Returns a KvSwap<K, V>
PTransform
.
- create() - Static method in class org.apache.beam.sdk.transforms.Values
-
Returns a Values<V>
PTransform
.
- Create.OfValueProvider<T> - Class in org.apache.beam.sdk.transforms
-
- Create.TimestampedValues<T> - Class in org.apache.beam.sdk.transforms
-
A PTransform
that creates a PCollection
whose elements have
associated timestamps.
- Create.Values<T> - Class in org.apache.beam.sdk.transforms
-
A PTransform
that creates a PCollection
from a set of in-memory objects.
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamAggregationTransforms.AggregationAdaptor
-
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
-
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn
-
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
-
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
-
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
-
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
-
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
-
Returns a new, mutable accumulator value, representing the accumulation of zero input values.
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
-
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
-
- createAccumulator(CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
-
- createAccumulator(CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
-
Returns a new, mutable accumulator value, representing the accumulation of zero input values.
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
-
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
-
- createContextual(DeserializationContext, BeanProperty) - Method in class org.apache.beam.sdk.options.ValueProvider.Deserializer
-
- CreateDataflowView<ElemT,ViewT> - Class in org.apache.beam.runners.dataflow
-
- createDecompressingChannel(ReadableByteChannel) - Method in enum org.apache.beam.sdk.io.CompressedSource.CompressionMode
-
Deprecated.
- createDecompressingChannel(ReadableByteChannel) - Method in interface org.apache.beam.sdk.io.CompressedSource.DecompressingChannelFactory
-
Given a channel, create a channel that decompresses the content read from the channel.
- createDefault() - Static method in class org.apache.beam.sdk.coders.CoderRegistry
-
Creates a CoderRegistry containing registrations for all standard coders part of the core Java
Apache Beam SDK and also any registrations provided by
coder registrars
.
- createForSubrangeOfFile(String, long, long) - Method in class org.apache.beam.sdk.io.AvroSource
-
- createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.AvroSource
-
- createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.BlockBasedSource
-
Creates a BlockBasedSource
for the specified range in a single file.
- createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.CompressedSource
-
Creates a CompressedSource
for a subrange of a file.
- createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.FileBasedSource
-
Creates and returns a new FileBasedSource
of the same type as the current
FileBasedSource
backed by a given file and an offset range.
- createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.xml.XmlSource
-
- CreateGearpumpPCollectionViewTranslator<ElemT,ViewT> - Class in org.apache.beam.runners.gearpump.translators
-
CreateGearpumpPCollectionView bridges input stream to down stream
transforms.
- CreateGearpumpPCollectionViewTranslator() - Constructor for class org.apache.beam.runners.gearpump.translators.CreateGearpumpPCollectionViewTranslator
-
- createInputFormatInstance() - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.HadoopInputFormatBoundedSource
-
Creates instance of InputFormat class.
- createJar(File, File) - Static method in class org.apache.beam.runners.apex.ApexYarnLauncher
-
Create a jar file from the given directory.
- createJob(Job) - Method in class org.apache.beam.runners.dataflow.DataflowClient
-
Creates the Dataflow Job
.
- CreateOptions - Class in org.apache.beam.sdk.io.fs
-
An abstract class that contains common configuration options for creating resources.
- CreateOptions() - Constructor for class org.apache.beam.sdk.io.fs.CreateOptions
-
- CreateOptions.Builder<BuilderT extends CreateOptions.Builder<BuilderT>> - Class in org.apache.beam.sdk.io.fs
-
- CreateOptions.StandardCreateOptions - Class in org.apache.beam.sdk.io.fs
-
A standard configuration options with builder.
- CreateOptions.StandardCreateOptions.Builder - Class in org.apache.beam.sdk.io.fs
-
- createPane(boolean, boolean, PaneInfo.Timing) - Static method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
-
- createPane(boolean, boolean, PaneInfo.Timing, long, long) - Static method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
-
Factory method to create a
PaneInfo
with the specified parameters.
- createPrimitiveOutputInternal(Pipeline, WindowingStrategy<?, ?>, PCollection.IsBounded, Coder<T>) - Static method in class org.apache.beam.sdk.values.PCollection
-
For internal use only; no backwards-compatibility guarantees.
- createProcessContext(ValueInSingleWindow<InputT>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
- createReader(PipelineOptions) - Method in class org.apache.beam.runners.gearpump.translators.io.BoundedSourceWrapper
-
- createReader(PipelineOptions) - Method in class org.apache.beam.runners.gearpump.translators.io.GearpumpSource
-
- createReader(PipelineOptions) - Method in class org.apache.beam.runners.gearpump.translators.io.UnboundedSourceWrapper
-
- createReader(PipelineOptions, UnboundedSource.CheckpointMark) - Method in class org.apache.beam.runners.gearpump.translators.io.ValuesSource
-
- createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.BoundedSource
-
- createReader(CassandraIO.CassandraSource<T>) - Method in interface org.apache.beam.sdk.io.cassandra.CassandraService
-
- createReader(CassandraIO.CassandraSource<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraServiceImpl
-
- createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
-
- createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
-
- createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.HadoopInputFormatBoundedSource
-
- createReader(PipelineOptions, JmsCheckpointMark) - Method in class org.apache.beam.sdk.io.jms.JmsIO.UnboundedJmsSource
-
- createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
-
- createReader(PipelineOptions, CheckpointMarkT) - Method in class org.apache.beam.sdk.io.UnboundedSource
-
- createRunner(ReadyCheckingSideInputReader) - Method in class org.apache.beam.runners.gearpump.translators.utils.DoFnRunnerFactory
-
- createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.AvroSource
-
- createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.BlockBasedSource
-
Creates a BlockBasedReader
.
- createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.CompressedSource
-
Creates a FileBasedReader
to read a single file.
- createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
-
Creates and returns an instance of a FileBasedReader
implementation for the current
source assuming the source represents a single file.
- createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.xml.XmlSource
-
- createSourceForSubrange(long, long) - Method in class org.apache.beam.sdk.io.FileBasedSource
-
- createSourceForSubrange(long, long) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-
- CreateStream<T> - Class in org.apache.beam.runners.spark.io
-
Create an input stream from Queue.
- createStructuralValues(Coder<T>, List<T>) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-
Testing utilities below depend on standard assertions and matchers to compare elements read by
sources.
- CreateTables<DestinationT> - Class in org.apache.beam.sdk.io.gcp.bigquery
-
Creates any tables needed before performing streaming writes to the tables.
- CreateTables(BigQueryIO.Write.CreateDisposition, DynamicDestinations<?, DestinationT>) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.CreateTables
-
- createTransaction() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO
-
Returns a transform that creates a batch transaction.
- CreateTransaction() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
-
- createWriteOperation() - Method in class org.apache.beam.sdk.io.FileBasedSink
-
- createWriter(CassandraIO.Write<T>) - Method in interface org.apache.beam.sdk.io.cassandra.CassandraService
-
- createWriter(CassandraIO.Write<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraServiceImpl
-
- createWriter() - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-
- CredentialFactory - Interface in org.apache.beam.sdk.extensions.gcp.auth
-
Construct an oauth credential to be used by the SDK and the SDK workers.
- csvFormat - Variable in class org.apache.beam.sdk.extensions.sql.impl.schema.text.BeamTextCSVTableIOReader
-
- csvFormat - Variable in class org.apache.beam.sdk.extensions.sql.impl.schema.text.BeamTextCSVTableIOWriter
-
- csvLine2BeamSqlRow(CSVFormat, String, BeamRecordSqlType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamTableUtils
-
- CsvRecorderDecoder(BeamRecordSqlType, CSVFormat) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.kafka.BeamKafkaCSVTable.CsvRecorderDecoder
-
- CsvRecorderEncoder(BeamRecordSqlType, CSVFormat) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.kafka.BeamKafkaCSVTable.CsvRecorderEncoder
-
- CsvSink - Class in org.apache.beam.runners.spark.metrics.sink
-
- CsvSink(Properties, MetricRegistry, SecurityManager) - Constructor for class org.apache.beam.runners.spark.metrics.sink.CsvSink
-
- ctxt - Variable in class org.apache.beam.runners.spark.SparkRunner.Evaluator
-
- current() - Method in interface org.apache.beam.runners.spark.aggregators.NamedAggregators.State
-
- currentEventTime() - Method in interface org.apache.beam.sdk.state.Timers
-
Returns the current event time.
- currentInputWatermarkTime() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-
- currentOutputWatermarkTime() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-
- currentProcessingTime() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-
- currentProcessingTime() - Method in interface org.apache.beam.sdk.state.Timers
-
Returns the current processing time.
- currentRestriction() - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
-
- currentRestriction() - Method in interface org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker
-
Returns a restriction accurately describing the full range of work the current DoFn.ProcessElement
call will do, including already completed work.
- currentSynchronizedProcessingTime() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-
- currentSynchronizedProcessingTime() - Method in interface org.apache.beam.sdk.state.Timers
-
Returns the current synchronized processing time or null
if unknown.
- CUSTOM_SOURCE_FORMAT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- CustomCoder<T> - Class in org.apache.beam.sdk.coders
-
- CustomCoder() - Constructor for class org.apache.beam.sdk.coders.CustomCoder
-
- DataflowClient - Class in org.apache.beam.runners.dataflow
-
Wrapper around the generated Dataflow
client to provide common functionality.
- DataflowClientFactory() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions.DataflowClientFactory
-
- DataflowJobAlreadyExistsException - Exception in org.apache.beam.runners.dataflow
-
An exception that is thrown if the unique job name constraint of the Dataflow
service is broken because an existing job with the same job name is currently active.
- DataflowJobAlreadyExistsException(DataflowPipelineJob, String) - Constructor for exception org.apache.beam.runners.dataflow.DataflowJobAlreadyExistsException
-
Create a new
DataflowJobAlreadyExistsException
with the specified
DataflowPipelineJob
and message.
- DataflowJobAlreadyUpdatedException - Exception in org.apache.beam.runners.dataflow
-
An exception that is thrown if the existing job has already been updated within the Dataflow
service and is no longer able to be updated.
- DataflowJobAlreadyUpdatedException(DataflowPipelineJob, String) - Constructor for exception org.apache.beam.runners.dataflow.DataflowJobAlreadyUpdatedException
-
Create a new
DataflowJobAlreadyUpdatedException
with the specified
DataflowPipelineJob
and message.
- DataflowJobException - Exception in org.apache.beam.runners.dataflow
-
- DataflowPipelineDebugOptions - Interface in org.apache.beam.runners.dataflow.options
-
Internal.
- DataflowPipelineDebugOptions.DataflowClientFactory - Class in org.apache.beam.runners.dataflow.options
-
Returns the default Dataflow client built from the passed in PipelineOptions.
- DataflowPipelineDebugOptions.StagerFactory - Class in org.apache.beam.runners.dataflow.options
-
- DataflowPipelineJob - Class in org.apache.beam.runners.dataflow
-
A DataflowPipelineJob represents a job submitted to Dataflow using
DataflowRunner
.
- DataflowPipelineJob(DataflowClient, String, DataflowPipelineOptions, Map<AppliedPTransform<?, ?, ?>, String>) - Constructor for class org.apache.beam.runners.dataflow.DataflowPipelineJob
-
Constructs the job.
- DataflowPipelineOptions - Interface in org.apache.beam.runners.dataflow.options
-
- DataflowPipelineOptions.StagingLocationFactory - Class in org.apache.beam.runners.dataflow.options
-
- DataflowPipelineRegistrar - Class in org.apache.beam.runners.dataflow
-
- DataflowPipelineRegistrar.Options - Class in org.apache.beam.runners.dataflow
-
- DataflowPipelineRegistrar.Runner - Class in org.apache.beam.runners.dataflow
-
- DataflowPipelineTranslator - Class in org.apache.beam.runners.dataflow
-
- DataflowPipelineTranslator.JobSpecification - Class in org.apache.beam.runners.dataflow
-
The result of a job translation.
- DataflowPipelineWorkerPoolOptions - Interface in org.apache.beam.runners.dataflow.options
-
Options that are used to configure the Dataflow pipeline worker pool.
- DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType - Enum in org.apache.beam.runners.dataflow.options
-
Type of autoscaling algorithm to use.
- DataflowPipelineWorkerPoolOptions.WorkerHarnessContainerImageFactory - Class in org.apache.beam.runners.dataflow.options
-
Returns the default Docker container image that executes Dataflow worker harness, residing in
Google Container Registry.
- DataflowProfilingAgentConfiguration() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowProfilingOptions.DataflowProfilingAgentConfiguration
-
- DataflowProfilingOptions - Interface in org.apache.beam.runners.dataflow.options
-
Options for controlling profiling of pipeline execution.
- DataflowProfilingOptions.DataflowProfilingAgentConfiguration - Class in org.apache.beam.runners.dataflow.options
-
Configuration the for profiling agent.
- DataflowRunner - Class in org.apache.beam.runners.dataflow
-
A
PipelineRunner
that executes the operations in the pipeline by first translating them
to the Dataflow representation using the
DataflowPipelineTranslator
and then submitting
them to a Dataflow service for execution.
- DataflowRunner(DataflowPipelineOptions) - Constructor for class org.apache.beam.runners.dataflow.DataflowRunner
-
- DataflowRunner.StreamingPCollectionViewWriterFn<T> - Class in org.apache.beam.runners.dataflow
-
- DataflowRunnerHooks - Class in org.apache.beam.runners.dataflow
-
An instance of this class can be passed to the
DataflowRunner
to add user defined hooks to be
invoked at various times during pipeline execution.
- DataflowRunnerHooks() - Constructor for class org.apache.beam.runners.dataflow.DataflowRunnerHooks
-
- DataflowRunnerInfo - Class in org.apache.beam.runners.dataflow
-
- DataflowServiceException - Exception in org.apache.beam.runners.dataflow
-
Signals there was an error retrieving information about a job from the Cloud Dataflow Service.
- DataflowTemplateJob - Class in org.apache.beam.runners.dataflow.util
-
- DataflowTemplateJob() - Constructor for class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
-
- DataflowTransport - Class in org.apache.beam.runners.dataflow.util
-
Helpers for cloud communication.
- DataflowTransport() - Constructor for class org.apache.beam.runners.dataflow.util.DataflowTransport
-
- DataflowWorkerHarnessOptions - Interface in org.apache.beam.runners.dataflow.options
-
Options that are used exclusively within the Dataflow worker harness.
- DataflowWorkerLoggingOptions - Interface in org.apache.beam.runners.dataflow.options
-
Options that are used to control logging configuration on the Dataflow worker.
- DataflowWorkerLoggingOptions.Level - Enum in org.apache.beam.runners.dataflow.options
-
The set of log levels that can be used on the Dataflow worker.
- DataflowWorkerLoggingOptions.WorkerLogLevelOverrides - Class in org.apache.beam.runners.dataflow.options
-
Defines a log level override for a specific class, package, or name.
- DataSourceConfiguration() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
-
- DatastoreIO - Class in org.apache.beam.sdk.io.gcp.datastore
-
- DatastoreV1 - Class in org.apache.beam.sdk.io.gcp.datastore
-
- DatastoreV1.DeleteEntity - Class in org.apache.beam.sdk.io.gcp.datastore
-
A
PTransform
that deletes
Entities
from Cloud Datastore.
- DatastoreV1.DeleteKey - Class in org.apache.beam.sdk.io.gcp.datastore
-
A
PTransform
that deletes
Entities
associated with the given
Keys
from Cloud Datastore.
- DatastoreV1.Read - Class in org.apache.beam.sdk.io.gcp.datastore
-
A
PTransform
that reads the result rows of a Cloud Datastore query as
Entity
objects.
- DatastoreV1.Write - Class in org.apache.beam.sdk.io.gcp.datastore
-
A
PTransform
that writes
Entity
objects to Cloud Datastore.
- days(int) - Static method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows
-
Returns a
WindowFn
that windows elements into periods measured by days.
- dec() - Method in interface org.apache.beam.sdk.metrics.Counter
-
- dec(long) - Method in interface org.apache.beam.sdk.metrics.Counter
-
- decode(InputStream) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.AvroCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.BeamRecordCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.BitSetCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.BitSetCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.BooleanCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.ByteCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.Coder
-
Decodes a value of type T
from the given input stream in
the given context.
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.Coder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.DelegateCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.DelegateCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.DoubleCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.DurationCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.InstantCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.KvCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.KvCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.MapCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.MapCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.NullableCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.NullableCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.SerializableCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.ShardedKeyCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.VarIntCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.VarLongCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.coders.VoidCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.BooleanCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.DateCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.DoubleCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.FloatCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.ShortCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.TimeCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamAggregationTransforms.AggregationAccumulatorCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.io.amqp.AmqpMessageCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.ParamsCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.io.fs.MetadataCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.io.fs.ResourceIdCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV2
-
- decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessagePayloadOnlyCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessagePayloadOnlyCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.io.hadoop.WritableCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.io.ReadableFileCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.io.xml.JAXBCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.io.xml.JAXBCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo.PaneInfoCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
-
- decode(InputStream) - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
-
- decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
-
- decodePane(byte) - Static method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
-
- decodeToIterable(List<T>) - Method in class org.apache.beam.sdk.coders.CollectionCoder
-
Builds an instance of IterableT
, this coder's associated Iterable
-like
subtype, from a list of decoded elements.
- decodeToIterable(List<T>) - Method in class org.apache.beam.sdk.coders.IterableCoder
-
- decodeToIterable(List<T>) - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
-
Builds an instance of IterableT
, this coder's associated Iterable
-like
subtype, from a list of decoded elements.
- decodeToIterable(List<T>) - Method in class org.apache.beam.sdk.coders.ListCoder
-
- decodeToIterable(List<T>) - Method in class org.apache.beam.sdk.coders.SetCoder
-
Builds an instance of IterableT
, this coder's associated Iterable
-like
subtype, from a list of decoded elements.
- deduceOutputType(SqlTypeName, SqlTypeName) - Static method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlArithmeticExpression
-
- Default - Annotation Type in org.apache.beam.sdk.options
-
Default
represents a set of annotations that can be used to annotate getter properties
on
PipelineOptions
with information representing the default value to be returned
if no value is specified.
- Default.Boolean - Annotation Type in org.apache.beam.sdk.options
-
This represents that the default of the option is the specified boolean primitive value.
- Default.Byte - Annotation Type in org.apache.beam.sdk.options
-
This represents that the default of the option is the specified byte primitive value.
- Default.Character - Annotation Type in org.apache.beam.sdk.options
-
This represents that the default of the option is the specified char primitive value.
- Default.Class - Annotation Type in org.apache.beam.sdk.options
-
This represents that the default of the option is the specified Class
value.
- Default.Double - Annotation Type in org.apache.beam.sdk.options
-
This represents that the default of the option is the specified double primitive value.
- Default.Enum - Annotation Type in org.apache.beam.sdk.options
-
This represents that the default of the option is the specified enum.
- Default.Float - Annotation Type in org.apache.beam.sdk.options
-
This represents that the default of the option is the specified float primitive value.
- Default.InstanceFactory - Annotation Type in org.apache.beam.sdk.options
-
- Default.Integer - Annotation Type in org.apache.beam.sdk.options
-
This represents that the default of the option is the specified int primitive value.
- Default.Long - Annotation Type in org.apache.beam.sdk.options
-
This represents that the default of the option is the specified long primitive value.
- Default.Short - Annotation Type in org.apache.beam.sdk.options
-
This represents that the default of the option is the specified short primitive value.
- Default.String - Annotation Type in org.apache.beam.sdk.options
-
This represents that the default of the option is the specified String
value.
- DEFAULT_BYTE_ARRAY_CODER - Static variable in class org.apache.beam.sdk.io.TFRecordIO
-
The default coder, which returns each record of the input file as a byte array.
- DEFAULT_MAX_NUM_ELEMENTS - Static variable in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
-
The cost (in time and space) to compute quantiles to a given
accuracy is a function of the total number of elements in the
data set.
- DEFAULT_QUEUE_MAX_POLL_TIME - Static variable in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
- DEFAULT_QUEUE_POLL_TIME - Static variable in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
- DEFAULT_SCHEME - Static variable in class org.apache.beam.sdk.io.FileSystems
-
- DEFAULT_UNWINDOWED_SHARD_TEMPLATE - Static variable in class org.apache.beam.sdk.io.DefaultFilenamePolicy
-
The default sharding name template.
- DefaultCoder - Annotation Type in org.apache.beam.sdk.coders
-
The
DefaultCoder
annotation specifies a
Coder
class to handle encoding and
decoding instances of the annotated class.
- DefaultCoder.DefaultCoderProviderRegistrar - Class in org.apache.beam.sdk.coders
-
- DefaultCoderCloudObjectTranslatorRegistrar - Class in org.apache.beam.runners.dataflow.util
-
- DefaultCoderCloudObjectTranslatorRegistrar() - Constructor for class org.apache.beam.runners.dataflow.util.DefaultCoderCloudObjectTranslatorRegistrar
-
- DefaultCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.coders.DefaultCoder.DefaultCoderProviderRegistrar
-
- DefaultConcludeTransform() - Constructor for class org.apache.beam.sdk.testing.PAssert.DefaultConcludeTransform
-
- DefaultFilenamePolicy - Class in org.apache.beam.sdk.io
-
- DefaultFilenamePolicy.Params - Class in org.apache.beam.sdk.io
-
- DefaultFilenamePolicy.ParamsCoder - Class in org.apache.beam.sdk.io
-
- DefaultParallelismFactory - Class in org.apache.beam.runners.flink
-
- DefaultParallelismFactory() - Constructor for class org.apache.beam.runners.flink.DefaultParallelismFactory
-
- DefaultPipelineOptionsRegistrar - Class in org.apache.beam.sdk.options
-
- DefaultPipelineOptionsRegistrar() - Constructor for class org.apache.beam.sdk.options.DefaultPipelineOptionsRegistrar
-
- DefaultProjectFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.DefaultProjectFactory
-
- Defaults() - Constructor for class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
-
- DefaultStopPipelineWatermarkFactory() - Constructor for class org.apache.beam.runners.spark.TestSparkPipelineOptions.DefaultStopPipelineWatermarkFactory
-
- DefaultTrigger - Class in org.apache.beam.sdk.transforms.windowing
-
A trigger that is equivalent to Repeatedly.forever(AfterWatermark.pastEndOfWindow())
.
- defaultValue() - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
-
Returns the default value when there are no values added to the accumulator.
- defaultValue() - Method in interface org.apache.beam.sdk.transforms.CombineFnBase.GlobalCombineFn
-
Returns the default value when there are no values added to the accumulator.
- defaultValue() - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
-
- defaultValue() - Method in class org.apache.beam.sdk.transforms.View.AsSingleton
-
Returns the default value of this transform, or null if there isn't one.
- DefaultValueFactory<T> - Interface in org.apache.beam.sdk.options
-
- delay(Duration) - Static method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform
-
For internal use only; no backwards-compatibility guarantees.
- Delay() - Constructor for class org.apache.beam.sdk.transforms.windowing.TimestampTransform.Delay
-
- delegate() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.ParDoSingle
-
- delegate(HasDisplayData) - Method in interface org.apache.beam.sdk.transforms.display.DisplayData.Builder
-
Register display data from the specified component on behalf of the current component.
- DelegateCoder<T,IntermediateT> - Class in org.apache.beam.sdk.coders
-
A
DelegateCoder<T, IntermediateT>
wraps a
Coder
for
IntermediateT
and
encodes/decodes values of type
T
by converting
to/from
IntermediateT
and then encoding/decoding using the underlying
Coder<IntermediateT>
.
- DelegateCoder(Coder<IntermediateT>, DelegateCoder.CodingFunction<T, IntermediateT>, DelegateCoder.CodingFunction<IntermediateT, T>, TypeDescriptor<T>) - Constructor for class org.apache.beam.sdk.coders.DelegateCoder
-
- DelegateCoder.CodingFunction<InputT,OutputT> - Interface in org.apache.beam.sdk.coders
-
- delete(Collection<ResourceIdT>) - Method in class org.apache.beam.sdk.io.FileSystem
-
Deletes a collection of resources.
- delete(Collection<ResourceId>, MoveOptions...) - Static method in class org.apache.beam.sdk.io.FileSystems
-
Deletes a collection of resources.
- deleteEntity() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1
-
- deleteKey() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1
-
- deleteTimer(StateNamespace, String, TimeDomain) - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-
- deleteTimer(TimerInternals.TimerData) - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-
- deleteTimer(StateNamespace, String) - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-
- deleteTimer(Instant, TimeDomain) - Method in interface org.apache.beam.sdk.state.Timers
-
Removes the timer set in this context for the timestmap
and timeDomain
.
- dependsOnlyOnEarliestTimestamp() - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
-
Returns true
if the result of combination of many output timestamps actually depends
only on the earliest.
- dependsOnlyOnWindow() - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
-
Returns true
if the result does not depend on what outputs were combined but only
the window they are in.
- describeMismatchSafely(PipelineResult, Description) - Method in class org.apache.beam.sdk.testing.FileChecksumMatcher
-
- describeTo(Description) - Method in class org.apache.beam.sdk.testing.FileChecksumMatcher
-
- describeTo(Description) - Method in class org.apache.beam.sdk.testing.RegexMatcher
-
- describeTo(Description) - Method in class org.apache.beam.sdk.testing.TestPipelineOptions.AlwaysPassMatcher
-
- Description - Annotation Type in org.apache.beam.sdk.options
-
Descriptions are used to generate human readable output when the --help
command is specified.
- deserialize(String, byte[]) - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantDeserializer
-
- deserialize(JsonParser, DeserializationContext) - Method in class org.apache.beam.sdk.options.ValueProvider.Deserializer
-
- deserializeTimers(Collection<byte[]>, TimerInternals.TimerDataCoder) - Static method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-
- detect(String) - Static method in enum org.apache.beam.sdk.io.Compression
-
- detectClassPathResourcesToStage(ClassLoader) - Static method in class org.apache.beam.runners.dataflow.DataflowRunner
-
Attempts to detect all the resources the class loader has access to.
- detectClassPathResourcesToStage(ClassLoader) - Static method in class org.apache.beam.runners.flink.FlinkRunner
-
Attempts to detect all the resources the class loader has access to.
- DirectOptions - Interface in org.apache.beam.runners.direct
-
- DirectOptions.AvailableParallelismFactory - Class in org.apache.beam.runners.direct
-
- DIRECTORY_CONTAINER - Static variable in class org.apache.beam.sdk.io.ShardNameTemplate
-
Shard is a file within a directory.
- DirectRegistrar - Class in org.apache.beam.runners.direct
-
- DirectRegistrar.Options - Class in org.apache.beam.runners.direct
-
- DirectRegistrar.Runner - Class in org.apache.beam.runners.direct
-
- DirectRunner - Class in org.apache.beam.runners.direct
-
- DirectRunner() - Constructor for class org.apache.beam.sdk.options.PipelineOptions.DirectRunner
-
- DirectRunner.DirectPipelineResult - Class in org.apache.beam.runners.direct
-
- DirectTestOptions - Interface in org.apache.beam.runners.direct
-
Internal-only options for tweaking the behavior of the
DirectRunner
in ways that users
should never do.
- DISALLOW_COMBINER_LIFTING - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- discardingFiredPanes() - Method in class org.apache.beam.sdk.transforms.windowing.Window
-
Returns a new Window
PTransform
that uses the registered WindowFn and
Triggering behavior, and that discards elements in a pane after they are triggered.
- dispatchBag(Coder<?>) - Method in interface org.apache.beam.sdk.state.StateSpec.Cases
-
- dispatchBag(Coder<?>) - Method in class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
-
- dispatchCombining(Combine.CombineFn<?, ?, ?>, Coder<?>) - Method in interface org.apache.beam.sdk.state.StateSpec.Cases
-
- dispatchCombining(Combine.CombineFn<?, ?, ?>, Coder<?>) - Method in class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
-
- dispatchDefault() - Method in class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
-
- dispatchMap(Coder<?>, Coder<?>) - Method in interface org.apache.beam.sdk.state.StateSpec.Cases
-
- dispatchMap(Coder<?>, Coder<?>) - Method in class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
-
- dispatchSet(Coder<?>) - Method in interface org.apache.beam.sdk.state.StateSpec.Cases
-
- dispatchSet(Coder<?>) - Method in class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
-
- dispatchValue(Coder<?>) - Method in interface org.apache.beam.sdk.state.StateSpec.Cases
-
- dispatchValue(Coder<?>) - Method in class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
-
- DISPLAY_DATA - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- DisplayData - Class in org.apache.beam.sdk.transforms.display
-
Static display data associated with a pipeline component.
- DisplayData.Builder - Interface in org.apache.beam.sdk.transforms.display
-
Utility to build up display data from a component and its included
subcomponents.
- DisplayData.Identifier - Class in org.apache.beam.sdk.transforms.display
-
Unique identifier for a display data item within a component.
- DisplayData.Item - Class in org.apache.beam.sdk.transforms.display
-
Items
are the unit of display data.
- DisplayData.ItemSpec<T> - Class in org.apache.beam.sdk.transforms.display
-
- DisplayData.Path - Class in org.apache.beam.sdk.transforms.display
-
Structured path of registered display data within a component hierarchy.
- DisplayData.Type - Enum in org.apache.beam.sdk.transforms.display
-
Display data type.
- distance(long, long) - Static method in class org.apache.beam.sdk.io.cassandra.CassandraServiceImpl
-
Measure distance between two tokens.
- Distinct<T> - Class in org.apache.beam.sdk.transforms
-
Distinct<T>
takes a PCollection<T>
and
returns a PCollection<T>
that has all distinct elements of the
input.
- Distinct() - Constructor for class org.apache.beam.sdk.transforms.Distinct
-
- Distinct.WithRepresentativeValues<T,IdT> - Class in org.apache.beam.sdk.transforms
-
- Distribution - Interface in org.apache.beam.sdk.metrics
-
A metric that reports information about the distribution of reported values.
- distribution(String, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
-
Create a metric that records various statistics about the distribution of reported values.
- distribution(Class<?>, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
-
Create a metric that records various statistics about the distribution of reported values.
- DistributionResult - Class in org.apache.beam.sdk.metrics
-
- DistributionResult() - Constructor for class org.apache.beam.sdk.metrics.DistributionResult
-
- distributions() - Method in interface org.apache.beam.sdk.metrics.MetricQueryResults
-
Return the metric results for the distributions that matched the filter.
- doChecks(PAssert.PAssertionSite, ActualT, SerializableFunction<ActualT, Void>) - Static method in class org.apache.beam.sdk.testing.PAssert
-
- DoFn<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
-
The argument to
ParDo
providing the code to use to process
elements of the input
PCollection
.
- DoFn() - Constructor for class org.apache.beam.sdk.transforms.DoFn
-
- DoFn.BoundedPerElement - Annotation Type in org.apache.beam.sdk.transforms
-
- DoFn.FinishBundle - Annotation Type in org.apache.beam.sdk.transforms
-
Annotation for the method to use to finish processing a batch of elements.
- DoFn.FinishBundleContext - Class in org.apache.beam.sdk.transforms
-
- DoFn.GetInitialRestriction - Annotation Type in org.apache.beam.sdk.transforms
-
Annotation for the method that maps an element to an initial restriction for a
splittable DoFn
.
- DoFn.GetRestrictionCoder - Annotation Type in org.apache.beam.sdk.transforms
-
Annotation for the method that returns the coder to use for the restriction of a
splittable DoFn
.
- DoFn.NewTracker - Annotation Type in org.apache.beam.sdk.transforms
-
- DoFn.OnTimer - Annotation Type in org.apache.beam.sdk.transforms
-
Annotation for registering a callback for a timer.
- DoFn.OnTimerContext - Class in org.apache.beam.sdk.transforms
-
- DoFn.OutputReceiver<T> - Interface in org.apache.beam.sdk.transforms
-
Receives values of the given type.
- DoFn.ProcessContext - Class in org.apache.beam.sdk.transforms
-
- DoFn.ProcessContinuation - Class in org.apache.beam.sdk.transforms
-
When used as a return value of
DoFn.ProcessElement
, indicates whether there is more work to
be done for the current element.
- DoFn.ProcessElement - Annotation Type in org.apache.beam.sdk.transforms
-
Annotation for the method to use for processing elements.
- DoFn.Setup - Annotation Type in org.apache.beam.sdk.transforms
-
Annotation for the method to use to prepare an instance for processing bundles of elements.
- DoFn.SplitRestriction - Annotation Type in org.apache.beam.sdk.transforms
-
Annotation for the method that splits restriction of a
splittable DoFn
into multiple parts to
be processed in parallel.
- DoFn.StartBundle - Annotation Type in org.apache.beam.sdk.transforms
-
Annotation for the method to use to prepare an instance for processing a batch of elements.
- DoFn.StartBundleContext - Class in org.apache.beam.sdk.transforms
-
- DoFn.StateId - Annotation Type in org.apache.beam.sdk.transforms
-
Annotation for declaring and dereferencing state cells.
- DoFn.Teardown - Annotation Type in org.apache.beam.sdk.transforms
-
Annotation for the method to use to clean up this instance after processing bundles of
elements.
- DoFn.TimerId - Annotation Type in org.apache.beam.sdk.transforms
-
Annotation for declaring and dereferencing timers.
- DoFn.UnboundedPerElement - Annotation Type in org.apache.beam.sdk.transforms
-
- DoFn.WindowedContext - Class in org.apache.beam.sdk.transforms
-
Information accessible to all methods in this
DoFn
where the context is in some window.
- DoFnFunction<InputT,OutputT> - Class in org.apache.beam.runners.gearpump.translators.functions
-
Gearpump
FlatMapFunction
wrapper over Beam
DoFn
.
- DoFnFunction(GearpumpPipelineOptions, DoFn<InputT, OutputT>, WindowingStrategy<?, ?>, Collection<PCollectionView<?>>, Map<String, PCollectionView<?>>, TupleTag<OutputT>, List<TupleTag<?>>) - Constructor for class org.apache.beam.runners.gearpump.translators.functions.DoFnFunction
-
- DoFnRunnerFactory<InputT,OutputT> - Class in org.apache.beam.runners.gearpump.translators.utils
-
a serializable SimpleDoFnRunner
.
- DoFnRunnerFactory(GearpumpPipelineOptions, DoFn<InputT, OutputT>, Collection<PCollectionView<?>>, DoFnRunners.OutputManager, TupleTag<OutputT>, List<TupleTag<?>>, StepContext, WindowingStrategy<?, ?>) - Constructor for class org.apache.beam.runners.gearpump.translators.utils.DoFnRunnerFactory
-
- DoFnRunnerWithMetricsUpdate<InputT,OutputT> - Class in org.apache.beam.runners.flink.metrics
-
DoFnRunner
decorator which registers
MetricsContainerImpl
.
- DoFnRunnerWithMetricsUpdate(String, DoFnRunner<InputT, OutputT>, RuntimeContext) - Constructor for class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
-
- DoFnTester<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
-
A harness for unit-testing a
DoFn
.
- DoFnTester.CloningBehavior - Enum in org.apache.beam.sdk.transforms
-
When a
DoFnTester
should clone the
DoFn
under test and how it should manage
the lifecycle of the
DoFn
.
- DoubleCoder - Class in org.apache.beam.sdk.coders
-
A
DoubleCoder
encodes
Double
values in 8 bytes using Java serialization.
- doubles() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-
- doublesGlobally() - Static method in class org.apache.beam.sdk.transforms.Max
-
Returns a PTransform
that takes an input PCollection<Double>
and returns a
PCollection<Double>
whose contents is the maximum of the input PCollection
's
elements, or Double.NEGATIVE_INFINITY
if there are no elements.
- doublesGlobally() - Static method in class org.apache.beam.sdk.transforms.Min
-
Returns a PTransform
that takes an input PCollection<Double>
and returns a
PCollection<Double>
whose contents is the minimum of the input PCollection
's
elements, or Double.POSITIVE_INFINITY
if there are no elements.
- doublesGlobally() - Static method in class org.apache.beam.sdk.transforms.Sum
-
Returns a PTransform
that takes an input
PCollection<Double>
and returns a
PCollection<Double>
whose contents is the sum of the
input PCollection
's elements, or
0
if there are no elements.
- doublesPerKey() - Static method in class org.apache.beam.sdk.transforms.Max
-
Returns a PTransform
that takes an input PCollection<KV<K, Double>>
and returns
a PCollection<KV<K, Double>>
that contains an output element mapping each distinct key
in the input PCollection
to the maximum of the values associated with that key in the
input PCollection
.
- doublesPerKey() - Static method in class org.apache.beam.sdk.transforms.Min
-
Returns a PTransform
that takes an input PCollection<KV<K, Double>>
and returns
a PCollection<KV<K, Double>>
that contains an output element mapping each distinct key
in the input PCollection
to the minimum of the values associated with that key in the
input PCollection
.
- doublesPerKey() - Static method in class org.apache.beam.sdk.transforms.Sum
-
Returns a PTransform
that takes an input
PCollection<KV<K, Double>>
and returns a
PCollection<KV<K, Double>>
that contains an output
element mapping each distinct key in the input
PCollection
to the sum of the values associated with
that key in the input PCollection
.
- DurationCoder - Class in org.apache.beam.sdk.coders
-
- DynamicAvroDestinations<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io
-
- DynamicAvroDestinations() - Constructor for class org.apache.beam.sdk.io.DynamicAvroDestinations
-
- DynamicDestinations() - Constructor for class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
-
- DynamicDestinations<T,DestinationT> - Class in org.apache.beam.sdk.io.gcp.bigquery
-
This class provides the most general way of specifying dynamic BigQuery table destinations.
- DynamicDestinations() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
-
- DynamicFileDestinations - Class in org.apache.beam.sdk.io
-
- DynamicFileDestinations() - Constructor for class org.apache.beam.sdk.io.DynamicFileDestinations
-
- failure(PAssert.PAssertionSite, Throwable) - Static method in class org.apache.beam.sdk.testing.SuccessOrFailure
-
- FAILURE_COUNTER - Static variable in class org.apache.beam.sdk.testing.PAssert
-
- fewKeys() - Method in class org.apache.beam.sdk.transforms.GroupByKey
-
Returns whether it groups just few keys.
- fieldTypes - Variable in class org.apache.beam.sdk.extensions.sql.BeamRecordSqlType
-
- FileBasedReader(FileBasedSource<T>) - Constructor for class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
-
Subclasses should not perform IO operations at the constructor.
- FileBasedSink<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io
-
Abstract class for file-based output.
- FileBasedSink(ValueProvider<ResourceId>, FileBasedSink.DynamicDestinations<?, DestinationT, OutputT>) - Constructor for class org.apache.beam.sdk.io.FileBasedSink
-
Construct a
FileBasedSink
with the given temp directory, producing uncompressed files.
- FileBasedSink(ValueProvider<ResourceId>, FileBasedSink.DynamicDestinations<?, DestinationT, OutputT>, FileBasedSink.WritableByteChannelFactory) - Constructor for class org.apache.beam.sdk.io.FileBasedSink
-
Construct a
FileBasedSink
with the given temp directory and output channel type.
- FileBasedSink(ValueProvider<ResourceId>, FileBasedSink.DynamicDestinations<?, DestinationT, OutputT>, Compression) - Constructor for class org.apache.beam.sdk.io.FileBasedSink
-
Construct a
FileBasedSink
with the given temp directory and output channel type.
- FileBasedSink.CompressionType - Enum in org.apache.beam.sdk.io
-
- FileBasedSink.DynamicDestinations<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io
-
- FileBasedSink.FilenamePolicy - Class in org.apache.beam.sdk.io
-
A naming policy for output files.
- FileBasedSink.FileResult<DestinationT> - Class in org.apache.beam.sdk.io
-
Result of a single bundle write.
- FileBasedSink.FileResultCoder<DestinationT> - Class in org.apache.beam.sdk.io
-
- FileBasedSink.OutputFileHints - Interface in org.apache.beam.sdk.io
-
Provides hints about how to generate output files, such as a suggested filename suffix (e.g.
- FileBasedSink.WritableByteChannelFactory - Interface in org.apache.beam.sdk.io
-
- FileBasedSink.WriteOperation<DestinationT,OutputT> - Class in org.apache.beam.sdk.io
-
Abstract operation that manages the process of writing to
FileBasedSink
.
- FileBasedSink.Writer<DestinationT,OutputT> - Class in org.apache.beam.sdk.io
-
- FileBasedSource<T> - Class in org.apache.beam.sdk.io
-
A common base class for all file-based
Source
s.
- FileBasedSource(ValueProvider<String>, EmptyMatchTreatment, long) - Constructor for class org.apache.beam.sdk.io.FileBasedSource
-
Create a FileBaseSource
based on a file or a file pattern specification, with the given
strategy for treating filepatterns that do not match any files.
- FileBasedSource(ValueProvider<String>, long) - Constructor for class org.apache.beam.sdk.io.FileBasedSource
-
- FileBasedSource(MatchResult.Metadata, long, long, long) - Constructor for class org.apache.beam.sdk.io.FileBasedSource
-
Create a FileBasedSource
based on a single file.
- FileBasedSource.FileBasedReader<T> - Class in org.apache.beam.sdk.io
-
A
reader
that implements code common to readers of
FileBasedSource
s.
- FileBasedSource.Mode - Enum in org.apache.beam.sdk.io
-
A given FileBasedSource
represents a file resource of one of these types.
- FileChecksumMatcher - Class in org.apache.beam.sdk.testing
-
Matcher to verify file checksum in E2E test.
- FileChecksumMatcher(String, String) - Constructor for class org.apache.beam.sdk.testing.FileChecksumMatcher
-
Constructor that uses default shard template.
- FileChecksumMatcher(String, String, Pattern) - Constructor for class org.apache.beam.sdk.testing.FileChecksumMatcher
-
Constructor using a custom shard template.
- FileChecksumMatcher(String, ShardedFile) - Constructor for class org.apache.beam.sdk.testing.FileChecksumMatcher
-
Constructor using an entirely custom ShardedFile
implementation.
- FileIO - Class in org.apache.beam.sdk.io
-
Transforms for working with files.
- FileIO() - Constructor for class org.apache.beam.sdk.io.FileIO
-
- FileIO.Match - Class in org.apache.beam.sdk.io
-
- FileIO.MatchAll - Class in org.apache.beam.sdk.io
-
- FileIO.MatchConfiguration - Class in org.apache.beam.sdk.io
-
Describes configuration for matching filepatterns, such as
EmptyMatchTreatment
and
continuous watching for matching files.
- FileIO.ReadableFile - Class in org.apache.beam.sdk.io
-
A utility class for accessing a potentially compressed file.
- FileIO.ReadMatches - Class in org.apache.beam.sdk.io
-
- FilenamePolicy() - Constructor for class org.apache.beam.sdk.io.FileBasedSink.FilenamePolicy
-
- filePattern - Variable in class org.apache.beam.sdk.extensions.sql.impl.schema.text.BeamTextTable
-
- filepattern(String) - Method in class org.apache.beam.sdk.io.FileIO.Match
-
Matches the given filepattern.
- filepattern(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.FileIO.Match
-
- FileResult(ResourceId, int, BoundedWindow, PaneInfo, DestinationT) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.FileResult
-
- FileResultCoder(Coder<BoundedWindow>, Coder<DestinationT>) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
-
- FileSystem<ResourceIdT extends ResourceId> - Class in org.apache.beam.sdk.io
-
File system interface in Beam.
- FileSystem() - Constructor for class org.apache.beam.sdk.io.FileSystem
-
- FileSystemRegistrar - Interface in org.apache.beam.sdk.io
-
- FileSystems - Class in org.apache.beam.sdk.io
-
- FileSystems() - Constructor for class org.apache.beam.sdk.io.FileSystems
-
- Filter<T> - Class in org.apache.beam.sdk.transforms
-
PTransform
s for filtering from a PCollection
the
elements satisfying a predicate, or satisfying an inequality with
a given value based on the elements' natural ordering.
- finalize(Iterable<FileBasedSink.FileResult<DestinationT>>) - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-
Finalizes writing by copying temporary output files to their final location.
- finalizeCheckpoint() - Method in class org.apache.beam.runners.spark.io.EmptyCheckpointMark
-
- finalizeCheckpoint() - Method in class org.apache.beam.sdk.io.CountingSource.CounterMark
-
- finalizeCheckpoint() - Method in class org.apache.beam.sdk.io.jms.JmsCheckpointMark
-
Acknowledge all outstanding message.
- finalizeCheckpoint() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark
-
- finalizeCheckpoint() - Method in interface org.apache.beam.sdk.io.UnboundedSource.CheckpointMark
-
Called by the system to signal that this checkpoint mark has been committed along with
all the records which have been read from the
UnboundedSource.UnboundedReader
since the
previous checkpoint was taken.
- find(String) - Static method in class org.apache.beam.sdk.transforms.Regex
-
- find(Pattern) - Static method in class org.apache.beam.sdk.transforms.Regex
-
- find(String, int) - Static method in class org.apache.beam.sdk.transforms.Regex
-
- find(Pattern, int) - Static method in class org.apache.beam.sdk.transforms.Regex
-
- find(String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
-
- find(Pattern, String) - Static method in class org.apache.beam.sdk.transforms.Regex
-
- Find(Pattern, int) - Constructor for class org.apache.beam.sdk.transforms.Regex.Find
-
- findAll(String) - Static method in class org.apache.beam.sdk.transforms.Regex
-
- findAll(Pattern) - Static method in class org.apache.beam.sdk.transforms.Regex
-
- FindAll(Pattern) - Constructor for class org.apache.beam.sdk.transforms.Regex.FindAll
-
- findIndexOfField(String) - Method in class org.apache.beam.sdk.values.BeamRecordType
-
Find the index of a given field.
- findKV(String, int, int) - Static method in class org.apache.beam.sdk.transforms.Regex
-
- findKV(Pattern, int, int) - Static method in class org.apache.beam.sdk.transforms.Regex
-
- findKV(String, String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
-
- findKV(Pattern, String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
-
- FindKV(Pattern, int, int) - Constructor for class org.apache.beam.sdk.transforms.Regex.FindKV
-
- FindName(Pattern, String) - Constructor for class org.apache.beam.sdk.transforms.Regex.FindName
-
- FindNameKV(Pattern, String, String) - Constructor for class org.apache.beam.sdk.transforms.Regex.FindNameKV
-
- findTable(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
-
- finishBundle() - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
-
- finishBundle() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write.PubsubBoundedWriter
-
- finishBundle() - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
- FinishBundleContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.FinishBundleContext
-
- finishSpecifying() - Method in interface org.apache.beam.sdk.state.StateSpec
-
For internal use only; no backwards-compatibility guarantees.
- finishSpecifying(PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PCollection
-
After building, finalizes this
PValue
to make it ready for
running.
- finishSpecifying(PInput, PTransform<?, ?>) - Method in interface org.apache.beam.sdk.values.PValue
-
After building, finalizes this
PValue
to make it ready for being used as an input to a
PTransform
.
- finishSpecifying(PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PValueBase
-
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
-
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.WriteFilesResult
-
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PCollection
-
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PCollectionList
-
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
-
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PDone
-
Does nothing; there is nothing to finish specifying.
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in interface org.apache.beam.sdk.values.POutput
-
As part of applying the producing
PTransform
, finalizes this output to make it ready
for being used as an input and for running.
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PValueBase
-
- finishWrite() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
-
- fixDefaults() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
Fixes all the defaults so that equals can be used to check that two strategies are the same,
regardless of the state of "defaulted-ness".
- fixedSizeGlobally(int) - Static method in class org.apache.beam.sdk.transforms.Sample
-
Returns a PTransform
that takes a PCollection<T>
, selects sampleSize
elements, uniformly at random, and returns a PCollection<Iterable<T>>
containing the
selected elements.
- fixedSizePerKey(int) - Static method in class org.apache.beam.sdk.transforms.Sample
-
Returns a PTransform
that takes an input PCollection<KV<K, V>>
and returns a
PCollection<KV<K, Iterable<V>>>
that contains an output element mapping each distinct
key in the input PCollection
to a sample of sampleSize
values associated with
that key in the input PCollection
, taken uniformly at random.
- FixedWindows - Class in org.apache.beam.sdk.transforms.windowing
-
A
WindowFn
that windows values into fixed-size timestamp-based windows.
- flatMap(List<TranslatorUtils.RawUnionValue>) - Method in class org.apache.beam.runners.gearpump.translators.functions.DoFnFunction
-
- flatMap(WindowedValue<T>) - Method in class org.apache.beam.runners.gearpump.translators.WindowAssignTranslator.AssignWindows
-
- FlatMapElements<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
-
PTransform
s for mapping a simple function that returns iterables over the elements of a
PCollection
and merging the results.
- Flatten - Class in org.apache.beam.sdk.transforms
-
Flatten<T>
takes multiple PCollection<T>
s bundled
into a PCollectionList<T>
and returns a single
PCollection<T>
containing all the elements in all the input
PCollection
s.
- Flatten() - Constructor for class org.apache.beam.sdk.transforms.Flatten
-
- Flatten.Iterables<T> - Class in org.apache.beam.sdk.transforms
-
FlattenIterables<T>
takes a PCollection<Iterable<T>>
and returns a
PCollection<T>
that contains all the elements from each iterable.
- Flatten.PCollections<T> - Class in org.apache.beam.sdk.transforms
-
- FlattenPCollectionsTranslator<T> - Class in org.apache.beam.runners.gearpump.translators
-
Flatten.FlattenPCollectionList is translated to Gearpump merge function.
- FlattenPCollectionsTranslator() - Constructor for class org.apache.beam.runners.gearpump.translators.FlattenPCollectionsTranslator
-
- FlattenWithHeterogeneousCoders - Interface in org.apache.beam.sdk.testing
-
- FlinkDetachedRunnerResult - Class in org.apache.beam.runners.flink
-
Result of a detached execution of a
Pipeline
with Flink.
- FlinkMetricContainer - Class in org.apache.beam.runners.flink.metrics
-
Helper class for holding a MetricsContainerImpl
and forwarding Beam metrics to
Flink accumulators and metrics.
- FlinkMetricContainer(RuntimeContext) - Constructor for class org.apache.beam.runners.flink.metrics.FlinkMetricContainer
-
- FlinkMetricContainer.FlinkDistributionGauge - Class in org.apache.beam.runners.flink.metrics
-
- FlinkMetricContainer.FlinkGauge - Class in org.apache.beam.runners.flink.metrics
-
- FlinkPipelineOptions - Interface in org.apache.beam.runners.flink
-
Options which can be used to configure a Flink PipelineRunner.
- FlinkRunner - Class in org.apache.beam.runners.flink
-
A
PipelineRunner
that executes the operations in the
pipeline by first translating them to a Flink Plan and then executing them either locally
or on a Flink cluster, depending on the configuration.
- FlinkRunnerRegistrar - Class in org.apache.beam.runners.flink
-
AutoService registrar - will register FlinkRunner and FlinkOptions
as possible pipeline runner services.
- FlinkRunnerRegistrar.Options - Class in org.apache.beam.runners.flink
-
Pipeline options registrar.
- FlinkRunnerRegistrar.Runner - Class in org.apache.beam.runners.flink
-
Pipeline runner registrar.
- FlinkRunnerResult - Class in org.apache.beam.runners.flink
-
Result of executing a
Pipeline
with Flink.
- FlinkTransformOverrides - Class in org.apache.beam.runners.flink
-
- FlinkTransformOverrides() - Constructor for class org.apache.beam.runners.flink.FlinkTransformOverrides
-
- floats() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-
- fold(KV<Instant, WindowedValue<KV<K, List<V>>>>, KV<Instant, WindowedValue<KV<K, V>>>) - Method in class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator.Merge
-
- forBoolean(Boolean) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
-
Constructs a CloudObject
representing the given value.
- forClass(Class<?>) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
-
Constructs a CloudObject
to be used for serializing an instance of
the supplied class for transport via the Dataflow API.
- forClassName(String) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
-
Constructs a CloudObject
to be used for serializing data to be
deserialized using the supplied class name the supplied class name for
transport via the Dataflow API.
- forCoder(TypeDescriptor<?>, Coder<?>) - Static method in class org.apache.beam.sdk.coders.CoderProviders
-
Creates a
CoderProvider
that always returns the
given coder for the specified type.
- forever(Trigger) - Static method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
-
Create a composite trigger that repeatedly executes the trigger repeated
, firing each
time it fires and ignoring any indications to finish.
- forFloat(Float) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
-
Constructs a CloudObject
representing the given value.
- forFloat(Double) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
-
Constructs a CloudObject
representing the given value.
- forInteger(Long) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
-
Constructs a CloudObject
representing the given value.
- forInteger(Integer) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
-
Constructs a CloudObject
representing the given value.
- forKnownType(Object) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
-
Constructs a CloudObject
representing the given value of a
well-known cloud object type.
- FORMAT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- formatRecord(UserT) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
-
Convert an input record type into the output type.
- formatTimestamp(Instant) - Static method in class org.apache.beam.sdk.transforms.windowing.BoundedWindow
-
- forNewInput(Instant, InputT) - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
-
Called by the
Watch
transform to create a new independent termination state for a
newly arrived
InputT
.
- forRegistry(MetricRegistry) - Static method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
-
- forStreamFromSources(List<Integer>, Map<Integer, GlobalWatermarkHolder.SparkWatermarks>) - Static method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-
Build the TimerInternals
according to the feeding streams.
- forString(String) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
-
Constructs a CloudObject
representing the given value.
- forTransformHierarchy(TransformHierarchy, PipelineOptions) - Static method in class org.apache.beam.sdk.Pipeline
-
- from(Map<String, String>) - Static method in class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
-
Expects a map keyed by logger Name
s with values representing Level
s.
- from(String) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
-
Reads from the given filename or filepattern.
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
-
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.AvroIO.Read
-
Reads from the given filename or filepattern.
- from(String) - Method in class org.apache.beam.sdk.io.AvroIO.Read
-
- from(ValueProvider<String>) - Static method in class org.apache.beam.sdk.io.AvroSource
-
Reads from the given file name or pattern ("glob").
- from(String) - Static method in class org.apache.beam.sdk.io.AvroSource
-
- from(FileBasedSource<T>) - Static method in class org.apache.beam.sdk.io.CompressedSource
-
Creates a CompressedSource
from an underlying FileBasedSource
.
- from(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
-
Reads a BigQuery table specified as "[project_id]:[dataset_id].[table_id]"
or
"[dataset_id].[table_id]"
for tables within the current project.
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
-
- from(TableReference) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
-
- from(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-
- from(TableReference) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-
- from(long) - Static method in class org.apache.beam.sdk.io.GenerateSequence
-
Specifies the minimum number to generate (inclusive).
- from(BoundedSource<T>) - Method in class org.apache.beam.sdk.io.Read.Builder
-
Returns a new Read.Bounded
PTransform
reading from the given
BoundedSource
.
- from(UnboundedSource<T, ?>) - Method in class org.apache.beam.sdk.io.Read.Builder
-
Returns a new Read.Unbounded
PTransform
reading from the given
UnboundedSource
.
- from(BoundedSource<T>) - Static method in class org.apache.beam.sdk.io.Read
-
Returns a new Read.Bounded
PTransform
reading from the given
BoundedSource
.
- from(UnboundedSource<T, ?>) - Static method in class org.apache.beam.sdk.io.Read
-
- from(String) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Read
-
Provide name of collection while reading from Solr.
- from(String) - Method in class org.apache.beam.sdk.io.TextIO.Read
-
Reads text files that reads from the file(s) with the given filename or filename pattern.
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.TextIO.Read
-
- from(String) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
-
Returns a transform for reading TFRecord files that reads from the file(s)
with the given filename or filename pattern.
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
-
- from(String) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
A
PTransform
that parses one or more files with the given filename
or filename pattern and returns a bounded
PCollection
containing
one element for each sequence of characters reported by Apache Tika SAX Parser.
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
- from(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
-
Reads a single XML file or a set of XML files defined by a Java "glob" file pattern.
- from(HasDisplayData) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
-
- fromArgs(String...) - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
-
Sets the command line arguments to parse when constructing the
PipelineOptions
.
- fromArgs(String...) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
-
Sets the command line arguments to parse when constructing the
PipelineOptions
.
- fromByteArray(byte[], Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
-
Utility method for deserializing a byte array using the specified coder.
- fromByteArrays(Collection<byte[]>, Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
-
Utility method for deserializing a Iterable of byte arrays using the specified coder.
- fromByteFunction(Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
-
A function wrapper for converting a byte array to an object.
- fromByteFunction(Coder<K>, Coder<V>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
-
A function wrapper for converting a byte array pair to a key-value pair.
- fromByteFunctionIterable(Coder<K>, Coder<V>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
-
A function wrapper for converting a byte array pair to a key-value pair, where
values are Iterable
.
- fromCanonical(Compression) - Static method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
-
Deprecated.
- fromCloudDuration(String) - Static method in class org.apache.beam.runners.dataflow.util.TimeUtil
-
Converts a Dataflow API duration string into a
Duration
.
- fromCloudObject(CloudObject) - Method in interface org.apache.beam.runners.dataflow.util.CloudObjectTranslator
-
Converts back into the original object from a provided
CloudObject
.
- fromCloudTime(String) - Static method in class org.apache.beam.runners.dataflow.util.TimeUtil
-
Converts a time value received via the Dataflow API into the corresponding
Instant
.
- fromFile(File, boolean) - Static method in class org.apache.beam.sdk.io.LocalResources
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.apex.ApexRunner
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.apex.TestApexRunner
-
- fromOptions(DataflowPipelineOptions) - Static method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator
-
Constructs a translator from the provided options.
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.dataflow.DataflowRunner
-
Construct a runner from the provided options.
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.dataflow.TestDataflowRunner
-
Constructs a runner from the provided options.
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.dataflow.util.GcsStager
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.direct.DirectRunner
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.flink.FlinkRunner
-
Construct a runner from the provided options.
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.flink.TestFlinkRunner
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.gearpump.GearpumpRunner
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.gearpump.TestGearpumpRunner
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.spark.SparkRunner
-
Creates and returns a new SparkRunner with specified options.
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.spark.SparkRunnerDebugger
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.spark.TestSparkRunner
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.extensions.gcp.auth.GcpCredentialFactory
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.extensions.gcp.auth.NoopCredentialFactory
-
- fromOptions(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsFileSystemRegistrar
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
-
- fromOptions(PipelineOptions) - Method in interface org.apache.beam.sdk.io.FileSystemRegistrar
-
- fromOptions(PipelineOptions) - Method in class org.apache.beam.sdk.io.hdfs.HadoopFileSystemRegistrar
-
- fromOptions(PipelineOptions) - Method in class org.apache.beam.sdk.io.LocalFileSystemRegistrar
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.PipelineRunner
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.testing.CrashingRunner
-
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.testing.TestPipeline
-
- fromParams(DefaultFilenamePolicy.Params) - Static method in class org.apache.beam.sdk.io.DefaultFilenamePolicy
-
- fromPath(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription
-
Creates a class representing a Pub/Sub subscription from the specified subscription path.
- fromPath(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic
-
Creates a class representing a Cloud Pub/Sub topic from the specified topic path.
- fromPath(Path, boolean) - Static method in class org.apache.beam.sdk.io.LocalResources
-
- fromQuery(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
-
Reads results received after executing the given query.
- fromQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
-
- fromQuery(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-
- fromQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-
- fromRawEvents(Coder<T>, List<TestStream.Event<T>>) - Static method in class org.apache.beam.sdk.testing.TestStream
-
For internal use only.
- FromRawUnionValue() - Constructor for class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils.FromRawUnionValue
-
- fromSerializableFunctionWithOutputType(SerializableFunction<InputT, OutputT>, TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.transforms.SimpleFunction
-
- fromSpec(Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
-
Constructs a CloudObject
by copying the supplied serialized object
spec, which must represent an SDK object serialized for transport via the
Dataflow API.
- fromStandardParameters(ValueProvider<ResourceId>, String, String, boolean) - Static method in class org.apache.beam.sdk.io.DefaultFilenamePolicy
-
- fromStaticMethods(Class<?>, Class<?>) - Static method in class org.apache.beam.sdk.coders.CoderProviders
-
Creates a
CoderProvider
from a class's
static <T> Coder<T> of(TypeDescriptor<T>, List<Coder<?>>
) method.
- fromString(String, boolean) - Static method in class org.apache.beam.sdk.io.LocalResources
-
- fromString(ValueProvider<String>, boolean) - Static method in class org.apache.beam.sdk.io.LocalResources
-
- fromSubscription(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
-
Reads from the given subscription.
- fromSubscription(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
-
- fromTopic(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
-
Creates and returns a transform for reading from a Cloud Pub/Sub topic.
- fromTopic(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
-
- fullOuterJoin(PCollection<KV<K, V1>>, PCollection<KV<K, V2>>, V1, V2) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
-
Full Outer Join of two collections of KV elements.
- Gauge - Interface in org.apache.beam.sdk.metrics
-
A metric that reports the latest value out of reported values.
- gauge(String, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
-
Create a metric that can have its new value set, and is aggregated by taking the last reported
value.
- gauge(Class<?>, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
-
Create a metric that can have its new value set, and is aggregated by taking the last reported
value.
- GaugeResult - Class in org.apache.beam.sdk.metrics
-
The result of a
Gauge
metric.
- GaugeResult() - Constructor for class org.apache.beam.sdk.metrics.GaugeResult
-
- GaugeResult.EmptyGaugeResult - Class in org.apache.beam.sdk.metrics
-
- gauges() - Method in interface org.apache.beam.sdk.metrics.MetricQueryResults
-
Return the metric results for the gauges that matched the filter.
- GcpCredentialFactory - Class in org.apache.beam.sdk.extensions.gcp.auth
-
Construct an oauth credential to be used by the SDK and the SDK workers.
- GcpCredentialFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.auth.GcpCredentialFactory
-
- GcpIoPipelineOptionsRegistrar - Class in org.apache.beam.sdk.io.gcp.common
-
A registrar containing the default GCP options.
- GcpIoPipelineOptionsRegistrar() - Constructor for class org.apache.beam.sdk.io.gcp.common.GcpIoPipelineOptionsRegistrar
-
- GcpOptions - Interface in org.apache.beam.sdk.extensions.gcp.options
-
Options used to configure Google Cloud Platform specific options such as the project
and credentials.
- GcpOptions.DefaultProjectFactory - Class in org.apache.beam.sdk.extensions.gcp.options
-
Attempts to infer the default project based upon the environment this application
is executing within.
- GcpOptions.GcpTempLocationFactory - Class in org.apache.beam.sdk.extensions.gcp.options
-
- GcpOptions.GcpUserCredentialsFactory - Class in org.apache.beam.sdk.extensions.gcp.options
-
Attempts to load the GCP credentials.
- GcpPipelineOptionsRegistrar - Class in org.apache.beam.sdk.extensions.gcp.options
-
A registrar containing the default GCP options.
- GcpPipelineOptionsRegistrar() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcpPipelineOptionsRegistrar
-
- GcpTempLocationFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpTempLocationFactory
-
- GcpUserCredentialsFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpUserCredentialsFactory
-
- GcsCreateOptions - Class in org.apache.beam.sdk.extensions.gcp.storage
-
An abstract class that contains common configuration options for creating resources.
- GcsCreateOptions() - Constructor for class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions
-
- GcsCreateOptions.Builder - Class in org.apache.beam.sdk.extensions.gcp.storage
-
- GcsFileSystemRegistrar - Class in org.apache.beam.sdk.extensions.gcp.storage
-
AutoService
registrar for the GcsFileSystem
.
- GcsFileSystemRegistrar() - Constructor for class org.apache.beam.sdk.extensions.gcp.storage.GcsFileSystemRegistrar
-
- GcsOptions - Interface in org.apache.beam.sdk.extensions.gcp.options
-
Options used to configure Google Cloud Storage.
- GcsOptions.ExecutorServiceFactory - Class in org.apache.beam.sdk.extensions.gcp.options
-
Returns the default ExecutorService
to use within the Apache Beam SDK.
- GcsOptions.PathValidatorFactory - Class in org.apache.beam.sdk.extensions.gcp.options
-
- GcsPathValidator - Class in org.apache.beam.sdk.extensions.gcp.storage
-
- GcsResourceId - Class in org.apache.beam.sdk.extensions.gcp.storage
-
ResourceId
implementation for Google Cloud Storage.
- GcsStager - Class in org.apache.beam.runners.dataflow.util
-
Utility class for staging files to GCS.
- gcsUploadBufferSizeBytes() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions
-
The buffer size (in bytes) to use when uploading files to GCS.
- GearpumpPipelineOptions - Interface in org.apache.beam.runners.gearpump
-
Options that configure the Gearpump pipeline.
- GearpumpPipelineResult - Class in org.apache.beam.runners.gearpump
-
Result of executing a
Pipeline
with Gearpump.
- GearpumpPipelineResult(ClientContext, RunningApplication) - Constructor for class org.apache.beam.runners.gearpump.GearpumpPipelineResult
-
- GearpumpPipelineTranslator - Class in org.apache.beam.runners.gearpump.translators
-
- GearpumpPipelineTranslator(TranslationContext) - Constructor for class org.apache.beam.runners.gearpump.translators.GearpumpPipelineTranslator
-
- GearpumpRunner - Class in org.apache.beam.runners.gearpump
-
A
PipelineRunner
that executes the operations in the
pipeline by first translating them to Gearpump Stream DSL
and then executing them on a Gearpump cluster.
- GearpumpRunner(GearpumpPipelineOptions) - Constructor for class org.apache.beam.runners.gearpump.GearpumpRunner
-
- GearpumpRunnerRegistrar - Class in org.apache.beam.runners.gearpump
-
- GearpumpRunnerRegistrar.Options - Class in org.apache.beam.runners.gearpump
-
- GearpumpRunnerRegistrar.Runner - Class in org.apache.beam.runners.gearpump
-
- GearpumpSource<T> - Class in org.apache.beam.runners.gearpump.translators.io
-
- GearpumpWindowFn(boolean) - Constructor for class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator.GearpumpWindowFn
-
- GenerateSequence - Class in org.apache.beam.sdk.io
-
A
PTransform
that produces longs starting from the given value, and either up to the
given limit or until
Long.MAX_VALUE
/ until the given time elapses.
- GenerateSequence() - Constructor for class org.apache.beam.sdk.io.GenerateSequence
-
- get() - Static method in class org.apache.beam.runners.spark.io.EmptyCheckpointMark
-
- get(Long) - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
-
- get(PCollectionView<T>, BoundedWindow) - Method in class org.apache.beam.runners.spark.util.SparkSideInputReader
-
- get() - Method in class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
-
- get() - Method in interface org.apache.beam.sdk.options.ValueProvider
-
- get() - Method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
-
- get() - Method in class org.apache.beam.sdk.options.ValueProvider.RuntimeValueProvider
-
- get() - Method in class org.apache.beam.sdk.options.ValueProvider.StaticValueProvider
-
- get(K) - Method in interface org.apache.beam.sdk.state.MapState
-
A deferred lookup.
- get(TupleTag<V>) - Method in class org.apache.beam.sdk.transforms.CombineFns.CoCombineResult
-
Returns the value represented by the given
TupleTag
.
- get(int) - Method in class org.apache.beam.sdk.values.PCollectionList
-
Returns the
PCollection
at the given index (origin zero).
- get(TupleTag<T>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
-
- get(int) - Method in class org.apache.beam.sdk.values.TupleTagList
-
Returns the
TupleTag
at the given index (origin zero).
- getAccum() - Method in interface org.apache.beam.sdk.state.CombiningState
-
Read the merged accumulator for this state cell.
- getAccumulatorCoder(CoderRegistry, Coder<BeamRecord>) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamAggregationTransforms.AggregationAdaptor
-
- getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
-
- getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn
-
- getAccumulatorCoder(CoderRegistry, Coder<Double>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
-
- getAccumulatorCoder(CoderRegistry, Coder<V>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
-
- getAccumulatorCoder(CoderRegistry, Coder<Integer>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
-
- getAccumulatorCoder(CoderRegistry, Coder<Long>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
-
- getAccumulatorCoder(CoderRegistry, Coder<InputT>) - Method in interface org.apache.beam.sdk.transforms.CombineFnBase.GlobalCombineFn
-
Returns the Coder
to use for accumulator AccumT
values, or null if it is not able to be inferred.
- getAccumulatorCoder(CoderRegistry, Coder<DataT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
-
- getAccumulatorCoder(CoderRegistry, Coder<DataT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
-
- getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
-
- getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
-
- getAdaptedSource() - Method in class org.apache.beam.sdk.io.BoundedReadFromUnboundedSource
-
- getAdditionalInputs() - Method in class org.apache.beam.sdk.io.WriteFiles
-
- getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
-
- getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
-
- getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-
- getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
-
- getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.PTransform
-
- getAdditionalOutputTags() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-
- getAddresses() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-
- getAlgorithm() - Method in enum org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType
-
Returns the string representation of this type.
- getAll(TupleTag<V>) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
-
Returns the values from the table represented by the given
TupleTag<V>
as an Iterable<V>
(which may be empty if there
are no results).
- getAll() - Method in class org.apache.beam.sdk.values.PCollectionList
-
- getAll() - Method in class org.apache.beam.sdk.values.PCollectionTuple
-
- getAll() - Method in class org.apache.beam.sdk.values.TupleTagList
-
- getAllowedLateness() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
- getAllowedTimestampSkew() - Method in class org.apache.beam.sdk.transforms.DoFn
-
- getAllowedTimestampSkew() - Method in class org.apache.beam.sdk.transforms.WithTimestamps
-
- getApexDAG() - Method in class org.apache.beam.runners.apex.ApexRunnerResult
-
Return the DAG executed by the pipeline.
- getApexLauncher() - Method in class org.apache.beam.runners.apex.ApexYarnLauncher.LaunchParams
-
- getApiRootUrl() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
The root URL for the Dataflow API.
- getApplicationName() - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
-
- getApplicationName() - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
-
- getAppliedFn(CoderRegistry, Coder<? extends KV<K, ? extends Iterable<InputT>>>, WindowingStrategy<?, ?>) - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
-
- getAppName() - Method in interface org.apache.beam.sdk.options.ApplicationNameOptions
-
Name of application, for display purposes.
- getApproximateArrivalTimestamp() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
-
- getArgumentTypes(Method) - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
Returns a list of argument types for the given method, which must
be a part of the class.
- getAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
-
Returns the given attribute value.
- getAttributeMap() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
-
Returns the full map of attributes.
- getAutoscalingAlgorithm() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
[Experimental] The autoscaling algorithm to use for the workerpool.
- getBatchDuration() - Method in class org.apache.beam.runners.spark.io.CreateStream
-
- getBatches() - Method in class org.apache.beam.runners.spark.io.CreateStream
-
Get the underlying queue representing the mock stream of micro-batches.
- getBatchIntervalMillis() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- getBigDecimal(String) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a BigDecimal
value by field name, ClassCastException
is thrown
if type doesn't match.
- getBigDecimal(int) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a BigDecimal
value by field index, ClassCastException
is thrown
if type doesn't match.
- getBigtableOptions() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-
Returns the Google Cloud Bigtable instance being read from, and other parameters.
- getBigtableOptions() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
-
Returns the Google Cloud Bigtable instance being written to, and other parameters.
- getBoolean(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getBoolean(Map<String, Object>, String, Boolean) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getBoolean() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlPrimitive
-
- getBoolean(String) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Boolean
value by field name, ClassCastException
is thrown
if type doesn't match.
- getBoolean(int) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Boolean
value by field index, ClassCastException
is thrown
if type doesn't match.
- getByte() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlPrimitive
-
- getByte(String) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Byte
value by field name, ClassCastException
is thrown
if type doesn't match.
- getByte(int) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Byte
value by field index, ClassCastException
is thrown
if type doesn't match.
- getBytes(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getBytes(Map<String, Object>, String, byte[]) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getBytes() - Method in class org.apache.beam.sdk.io.range.ByteKey
-
Returns a newly-allocated
byte[]
representing this
ByteKey
.
- getBytesPerOffset() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-
Returns approximately how many bytes of data correspond to a single offset in this source.
- getCause() - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
-
Returns the reason that this
WindowFn
is invalid.
- getChannelFactory() - Method in class org.apache.beam.sdk.io.CompressedSource
-
- getCheckpointDir() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- getCheckpointDurationMillis() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- getCheckpointingInterval() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- getCheckpointingMode() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- getCheckpointMark() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-
- getCheckpointMark() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
-
- getCheckpointMarkCoder() - Method in class org.apache.beam.runners.gearpump.translators.io.ValuesSource
-
- getCheckpointMarkCoder() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
-
- getCheckpointMarkCoder() - Method in class org.apache.beam.sdk.io.jms.JmsIO.UnboundedJmsSource
-
- getCheckpointMarkCoder() - Method in class org.apache.beam.sdk.io.UnboundedSource
-
Returns a
Coder
for encoding and decoding the checkpoints for this source.
- getCheckpointTimeoutMillis() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- getClasses() - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
Returns a set of
TypeDescriptor
s, one for each
superclass (including this class).
- getClassName() - Method in class org.apache.beam.runners.dataflow.util.CloudObject
-
Gets the name of the Java class that this CloudObject represents.
- getClientContext() - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
-
- getCloningBehavior() - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
- getClosingBehavior() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
- getCloudWatchClient() - Method in interface org.apache.beam.sdk.io.kinesis.AWSClientsProvider
-
- getCmd() - Method in class org.apache.beam.runners.apex.ApexYarnLauncher.LaunchParams
-
- getCodec(DestinationT) - Method in class org.apache.beam.sdk.io.DynamicAvroDestinations
-
Return an AVRO codec for a given destination.
- getCoder(Class<T>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
-
Returns the
Coder
to use for values of the given class.
- getCoder(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
-
Returns the
Coder
to use for values of the given type.
- getCoder(TypeDescriptor<OutputT>, TypeDescriptor<InputT>, Coder<InputT>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
-
- getCoder(Class<? extends T>, Class<T>, Map<Type, ? extends Coder<?>>, TypeVariable<?>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
-
- getCoder() - Method in class org.apache.beam.sdk.coders.DelegateCoder
-
Returns the coder used to encode/decode the intermediate values produced/consumed by the
coding functions of this DelegateCoder
.
- getCoder() - Static method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
-
- getCoder() - Method in class org.apache.beam.sdk.values.PCollection
-
Returns the
Coder
used by this
PCollection
to encode and decode
the values stored in it.
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.AtomicCoder
-
If this is a
Coder
for a parameterized type, returns the
list of
Coder
s being used for each of the parameters in the same order they appear
within the parameterized type's type signature.
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.Coder
-
If this is a
Coder
for a parameterized type, returns the
list of
Coder
s being used for each of the parameters in the same order they appear
within the parameterized type's type signature.
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.CustomCoder
-
If this is a
Coder
for a parameterized type, returns the
list of
Coder
s being used for each of the parameters in the same order they appear
within the parameterized type's type signature.
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.KvCoder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.MapCoder
-
If this is a
Coder
for a parameterized type, returns the
list of
Coder
s being used for each of the parameters in the same order they appear
within the parameterized type's type signature.
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.NullableCoder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.ShardedKeyCoder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.VarLongCoder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.io.hadoop.WritableCoder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
-
- getCoderArguments() - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
-
- getCoderInternal() - Method in interface org.apache.beam.sdk.values.PCollectionView
-
- getCoderInternal() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
-
- getCoderProvider() - Static method in class org.apache.beam.sdk.coders.AvroCoder
-
- getCoderProvider() - Static method in class org.apache.beam.sdk.coders.SerializableCoder
-
- getCoderProvider() - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
-
- getCoderProvider() - Static method in class org.apache.beam.sdk.io.hadoop.WritableCoder
-
- getCoderProviders() - Method in interface org.apache.beam.sdk.coders.CoderProviderRegistrar
-
- getCoderProviders() - Method in class org.apache.beam.sdk.coders.DefaultCoder.DefaultCoderProviderRegistrar
-
- getCoderProviders() - Method in class org.apache.beam.sdk.coders.SerializableCoder.SerializableCoderProviderRegistrar
-
- getCoderProviders() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtobufCoderProviderRegistrar
-
- getCoderProviders() - Method in class org.apache.beam.sdk.io.amqp.AmqpMessageCoderProviderRegistrar
-
- getCoderProviders() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryCoderProviderRegistrar
-
- getCoderProviders() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubCoderProviderRegistrar
-
- getCoderProviders() - Method in class org.apache.beam.sdk.io.hadoop.WritableCoder.WritableCoderProviderRegistrar
-
- getCoderProviders() - Method in class org.apache.beam.sdk.io.hbase.HBaseCoderProviderRegistrar
-
- getCoderRegistry() - Method in class org.apache.beam.sdk.Pipeline
-
- getCoders() - Method in class org.apache.beam.sdk.coders.BeamRecordCoder
-
- getCoGbkResultSchema() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
-
- getCollection() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple.TaggedKeyedPCollection
-
Returns the underlying PCollection of this TaggedKeyedPCollection.
- getColumns() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-
- getCombineFn() - Method in interface org.apache.beam.runners.spark.aggregators.NamedAggregators.State
-
- getCombineFn() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.UdafImpl
-
- getCombineFn() - Method in class org.apache.beam.sdk.transforms.Combine.GloballyAsSingletonView
-
- getComponents() - Method in class org.apache.beam.sdk.coders.AtomicCoder
-
Returns the list of
Coders
that are components of this
Coder
.
- getComponents() - Method in class org.apache.beam.sdk.coders.StructuredCoder
-
Returns the list of
Coders
that are components of this
Coder
.
- getComponents() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Path
-
Hierarchy list of component paths making up the full path, starting with the top-level child
component path.
- getComponents() - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
-
- getComponents() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
-
- getComponents() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
-
- getComponentType() - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
Returns the component type if this type is an array type,
otherwise returns null
.
- getCompression() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
-
- getConfigFile() - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
-
- getConfiguration() - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.HadoopInputFormatBoundedSource
-
- getConfiguration() - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.Read
-
- getConfiguration() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
-
- getConfiguration() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
-
- getContainerVersion() - Method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
-
Provides the container version that will be used for constructing harness image paths.
- getContentTypeHint() - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterAll
-
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterEach
-
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterFirst
-
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
-
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
-
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterSynchronizedProcessingTime
-
- getContinuationTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
-
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.DefaultTrigger
-
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.Never.NeverTrigger
-
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
-
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
-
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.ReshuffleTrigger
-
Deprecated.
- getContinuationTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
-
Return a trigger to use after a
GroupByKey
to preserve the intention of this trigger.
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
-
- getContinuationTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.Trigger.OnceTrigger
-
- getCorrectlyTypedResult(BigDecimal) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic.BeamSqlArithmeticExpression
-
- getCount() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
-
- getCounter(MetricName) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
-
Return the
Counter
that should be used for implementing the given
metricName
in this container.
- getCounters(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
-
- getCredential() - Method in interface org.apache.beam.sdk.extensions.gcp.auth.CredentialFactory
-
- getCredential() - Method in class org.apache.beam.sdk.extensions.gcp.auth.GcpCredentialFactory
-
Returns a default GCP Credentials
or null when it fails.
- getCredential() - Method in class org.apache.beam.sdk.extensions.gcp.auth.NoopCredentialFactory
-
- getCredentialFactoryClass() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-
The class of the credential factory that should be created and used to create
credentials.
- getCurrent() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-
- getCurrent() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
-
- getCurrent() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-
Gets the current record from the delegate reader.
- getCurrent() - Method in class org.apache.beam.sdk.io.Source.Reader
-
- getCurrentBlock() - Method in class org.apache.beam.sdk.io.AvroSource.AvroReader
-
- getCurrentBlock() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
-
- getCurrentBlockOffset() - Method in class org.apache.beam.sdk.io.AvroSource.AvroReader
-
- getCurrentBlockOffset() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
-
Returns the largest offset such that starting to read from that offset includes the current
block.
- getCurrentBlockSize() - Method in class org.apache.beam.sdk.io.AvroSource.AvroReader
-
- getCurrentBlockSize() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
-
Returns the size of the current block in bytes as it is represented in the underlying file,
if possible.
- getCurrentContainer() - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
-
- getCurrentDirectory() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
-
- getCurrentDirectory() - Method in interface org.apache.beam.sdk.io.fs.ResourceId
-
Returns the ResourceId
that represents the current directory of
this ResourceId
.
- getCurrentOffset() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
-
- getCurrentOffset() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-
- getCurrentOffset() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
- getCurrentRecord() - Method in class org.apache.beam.sdk.io.BlockBasedSource.Block
-
Returns the current record.
- getCurrentRecordId() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
-
Returns a unique identifier for the current record.
- getCurrentSource() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-
- getCurrentSource() - Method in class org.apache.beam.sdk.io.AvroSource.AvroReader
-
- getCurrentSource() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
-
Returns a Source
describing the same input that this Reader
currently reads
(including items already read).
- getCurrentSource() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
-
- getCurrentSource() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
- getCurrentSource() - Method in class org.apache.beam.sdk.io.Source.Reader
-
Returns a Source
describing the same input that this Reader
currently reads
(including items already read).
- getCurrentSource() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
-
- getCurrentTimestamp() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-
- getCurrentTimestamp() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
-
By default, returns the minimum possible timestamp.
- getCurrentTimestamp() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-
- getCurrentTimestamp() - Method in class org.apache.beam.sdk.io.Source.Reader
-
Returns the timestamp associated with the current data item.
- getData() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
-
- getDataAsBytes() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
-
- getDatabaseClient() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerAccessor
-
- getDataCoder() - Method in class org.apache.beam.runners.dataflow.DataflowRunner.StreamingPCollectionViewWriterFn
-
- getDataflowClient() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
An instance of the Dataflow client.
- getDataflowEndpoint() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
Dataflow endpoint to use.
- getDataflowJobFile() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
The path to write the translated Dataflow job specification out to
at job submission time.
- getDataflowOptions() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-
- getDataflowRunnerInfo() - Static method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
-
- getDataType() - Method in class org.apache.beam.sdk.values.BeamRecord
-
- getDataValues() - Method in class org.apache.beam.sdk.values.BeamRecord
-
Return the list of data values.
- getDate() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlPrimitive
-
- getDate(String) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Date
value by field name, ClassCastException
is thrown
if type doesn't match.
- getDate(int) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Date
value by field index, ClassCastException
is thrown
if type doesn't match.
- getDayOfMonth() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-
- getDayOfMonth() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-
- getDebuggee() - Method in interface org.apache.beam.runners.dataflow.options.CloudDebuggerOptions
-
The Cloud Debugger debuggee to associate with.
- getDecimal() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlPrimitive
-
- getDefaultCoder(TypeDescriptor<?>, CoderRegistry) - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.Read
-
Returns the default coder for a given type descriptor.
- getDefaultDestination() - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
-
Returns the default destination.
- getDefaultOutputCoder() - Method in class org.apache.beam.runners.gearpump.translators.io.ValuesSource
-
- getDefaultOutputCoder() - Method in class org.apache.beam.sdk.io.Source
-
- getDefaultOutputCoder() - Method in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
- getDefaultOutputCoder(CoderRegistry, Coder<Double>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
-
- getDefaultOutputCoder(CoderRegistry, Coder<V>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
-
- getDefaultOutputCoder(CoderRegistry, Coder<Integer>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
-
- getDefaultOutputCoder(CoderRegistry, Coder<Long>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
-
- getDefaultOutputCoder(CoderRegistry, Coder<InputT>) - Method in interface org.apache.beam.sdk.transforms.CombineFnBase.GlobalCombineFn
-
Returns the Coder
to use by default for output
OutputT
values, or null if it is not able to be inferred.
- getDefaultOutputCoder() - Method in class org.apache.beam.sdk.transforms.PTransform
-
- getDefaultOutputCoder(InputT) - Method in class org.apache.beam.sdk.transforms.PTransform
-
- getDefaultOutputCoder(InputT, PCollection<T>) - Method in class org.apache.beam.sdk.transforms.PTransform
-
- getDefaultOutputCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
-
- getDefaultOverrides(boolean) - Static method in class org.apache.beam.runners.flink.FlinkTransformOverrides
-
- getDefaultValue() - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn
-
Deprecated.
Returns the default value that was specified.
- getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-
- getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
-
- getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
-
- getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-
- getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-
Return a
WindowMappingFn
that returns the earliest window that contains the end of the
main-input window.
- getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
-
Returns the default
WindowMappingFn
to use to map main input windows to side input
windows.
- getDefaultWorkerLogLevel() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
-
This option controls the default log level of all loggers without a log level override.
- getDelay() - Method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform.Delay
-
- getDestination(UserT) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
-
Returns an object that represents at a high level the destination being written to.
- getDestination() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-
- getDestination() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
-
Return the user destination object for this writer.
- getDestination(ValueInSingleWindow<T>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
-
Returns an object that represents at a high level which table is being written to.
- getDestinationCoder() - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
-
Returns the coder for DestinationT
.
- getDestinationCoder() - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
-
Returns the coder for DestinationT
.
- getDestinationFile(FileBasedSink.DynamicDestinations<?, DestinationT, ?>, int, FileBasedSink.OutputFileHints) - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-
- getDictionary(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getDictionary(Map<String, Object>, String, Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getDiskSizeGb() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
Remote worker disk size, in gigabytes, or 0 to use the default size.
- getDistribution(MetricName) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
-
Return the
Distribution
that should be used for implementing the given
metricName
in this container.
- getDouble() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlPrimitive
-
- getDouble(String) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Double
value by field name, ClassCastException
is thrown
if type doesn't match.
- getDouble(int) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Double
value by field index, ClassCastException
is thrown
if type doesn't match.
- getDumpHeapOnOOM() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
If true, save a heap dump before killing a thread or process which is GC
thrashing or out of memory.
- getDynamicDestinations() - Method in class org.apache.beam.sdk.io.FileBasedSink
-
- getEarlyTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-
- getElemCoder() - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
-
- getElementCoders() - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
-
- getElementCount() - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
-
The number of elements after which this trigger may fire.
- getElements() - Method in class org.apache.beam.sdk.testing.TestStream.ElementEvent
-
- getElements() - Method in class org.apache.beam.sdk.transforms.Create.Values
-
- getEmbeddedCluster() - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
-
- getEmptyMatchTreatment() - Method in class org.apache.beam.sdk.io.FileBasedSource
-
- getEnableCloudDebugger() - Method in interface org.apache.beam.runners.dataflow.options.CloudDebuggerOptions
-
Whether to enable the Cloud Debugger snapshot agent for the current job.
- getEnableMetrics() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- getEnableSparkMetricSinks() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- getEncodedElementByteSize(RandomAccessData) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
-
- getEncodedElementByteSize(BigDecimal) - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
-
Returns the size in bytes of the encoded value using this coder.
- getEncodedElementByteSize(Integer) - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
-
Returns the size in bytes of the encoded value using this coder.
- getEncodedElementByteSize(Long) - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
-
Returns the size in bytes of the encoded value using this coder.
- getEncodedElementByteSize(BigInteger) - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
-
Returns the size in bytes of the encoded value using this coder.
- getEncodedElementByteSize(Boolean) - Method in class org.apache.beam.sdk.coders.BooleanCoder
-
- getEncodedElementByteSize(byte[]) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
-
- getEncodedElementByteSize(Byte) - Method in class org.apache.beam.sdk.coders.ByteCoder
-
Returns the size in bytes of the encoded value using this coder.
- getEncodedElementByteSize(T) - Method in class org.apache.beam.sdk.coders.Coder
-
Returns the size in bytes of the encoded value using this coder.
- getEncodedElementByteSize(Double) - Method in class org.apache.beam.sdk.coders.DoubleCoder
-
Returns the size in bytes of the encoded value using this coder.
- getEncodedElementByteSize(Instant) - Method in class org.apache.beam.sdk.coders.InstantCoder
-
- getEncodedElementByteSize(T) - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
-
Overridden to short-circuit the default StructuredCoder
behavior of encoding and
counting the bytes.
- getEncodedElementByteSize(T) - Method in class org.apache.beam.sdk.coders.NullableCoder
-
Overridden to short-circuit the default StructuredCoder
behavior of encoding and
counting the bytes.
- getEncodedElementByteSize(String) - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
-
Returns the size in bytes of the encoded value using this coder.
- getEncodedElementByteSize(Integer) - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
-
- getEncodedElementByteSize(Integer) - Method in class org.apache.beam.sdk.coders.VarIntCoder
-
- getEncodedElementByteSize(Long) - Method in class org.apache.beam.sdk.coders.VarLongCoder
-
- getEncodedElementByteSize(Void) - Method in class org.apache.beam.sdk.coders.VoidCoder
-
- getEncodedElementByteSize(ByteString) - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
-
- getEncodedElementByteSize(TableRow) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.AvroCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.ByteCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.Coder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.CollectionCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.DelegateCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.DoubleCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.DurationCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.InstantCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.IterableCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.KvCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.ListCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.MapCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.NullableCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.SerializableCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.SetCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.VarIntCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.VarLongCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.VoidCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.io.xml.JAXBCoder
-
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
-
- getEndKey() - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
-
- getEndOffset() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-
Returns the specified ending offset of the source.
- getEnv() - Method in class org.apache.beam.runners.apex.ApexYarnLauncher.LaunchParams
-
- getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.BoundedSource
-
An estimate of the total size (in bytes) of the data that would be read from this source.
- getEstimatedSizeBytes(CassandraIO.Read<T>) - Method in interface org.apache.beam.sdk.io.cassandra.CassandraService
-
Returns an estimation of the size that could be read.
- getEstimatedSizeBytes(CassandraIO.Read<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraServiceImpl
-
- getEstimatedSizeBytes(List<CassandraServiceImpl.TokenRange>) - Static method in class org.apache.beam.sdk.io.cassandra.CassandraServiceImpl
-
Actually estimate the size of the data to read on the cluster, based on the given token
ranges to address.
- getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
-
- getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
-
- getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.HadoopInputFormatBoundedSource
-
- getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
-
- getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-
- getEvents() - Method in class org.apache.beam.sdk.testing.TestStream
-
- getExecutionRetryDelay() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- getExecutorService() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-
The ExecutorService instance to use to create threads, can be overridden to specify an
ExecutorService that is compatible with the users environment.
- getExpectedAssertions() - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
-
- getExperiments() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
The list of backend experiments to enable.
- getExtendedSequenceNumber() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
-
- getExtensionHosts() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
-
- getExtensionRegistry() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
-
Returns the
ExtensionRegistry
listing all known Protocol Buffers extension messages
to
T
registered with this
ProtoCoder
.
- getFailedInserts() - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
-
- getFanout() - Method in class org.apache.beam.sdk.transforms.Combine.GloballyAsSingletonView
-
- getFieldCount() - Method in class org.apache.beam.sdk.values.BeamRecord
-
Return the size of data fields.
- getFieldCount() - Method in class org.apache.beam.sdk.values.BeamRecordType
-
Return the count of fields.
- getFieldNameByIndex(int) - Method in class org.apache.beam.sdk.values.BeamRecordType
-
Return the name of field by index.
- getFieldNames() - Method in class org.apache.beam.sdk.values.BeamRecordType
-
Returns an immutable list of field names.
- getFieldType(BeamRecordSqlType, int) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-
Get the SqlTypeName
for the specified column of a table.
- getFieldTypeByIndex(int) - Method in class org.apache.beam.sdk.extensions.sql.BeamRecordSqlType
-
- getFieldTypes() - Method in class org.apache.beam.sdk.extensions.sql.BeamRecordSqlType
-
- getFieldValue(String) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get value by field name.
- getFieldValue(int) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get value by field index.
- getFilename() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
-
- getFilename() - Method in interface org.apache.beam.sdk.io.fs.ResourceId
-
Returns the name of the file or directory denoted by this ResourceId
.
- getFilenamePolicy(DestinationT) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
-
- getFileOrPatternSpec() - Method in class org.apache.beam.sdk.io.FileBasedSource
-
- getFileOrPatternSpecProvider() - Method in class org.apache.beam.sdk.io.FileBasedSource
-
- getFilesToStage() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
List of local files to make available to workers.
- getFilesToStage() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
List of local files to make available to workers.
- getFlinkMaster() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
The url of the Flink JobManager on which to execute pipelines.
- getFloat() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlPrimitive
-
- getFloat(String) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Float
value by field name, ClassCastException
is thrown
if type doesn't match.
- getFloat(int) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Float
value by field index, ClassCastException
is thrown
if type doesn't match.
- getFn() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.ParDoSingle
-
- getFn() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
-
- getFn() - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
-
- getFn() - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
-
- getFn() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-
- getFn() - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
-
- getFnApiEnvironmentMajorVersion() - Method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
-
Provides the FnAPI environment's major version number.
- getFractionConsumed() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
-
- getFractionConsumed() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
-
Returns a value in [0, 1] representing approximately what fraction of the
current source
this reader has read so far, or
null
if such
an estimate is not available.
- getFractionConsumed() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
- getFractionConsumed() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-
- getFractionConsumed() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-
- getFractionConsumed() - Method in interface org.apache.beam.sdk.io.range.RangeTracker
-
- getFractionOfBlockConsumed() - Method in class org.apache.beam.sdk.io.BlockBasedSource.Block
-
Returns the fraction of the block already consumed, if possible, as a value in
[0, 1]
.
- getFrom() - Method in class org.apache.beam.sdk.io.range.OffsetRange
-
- getFullName(PTransform<?, ?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
-
Returns the full name of the currently being translated transform.
- getGapDuration() - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-
- getGauge(MetricName) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
-
Return the
Gauge
that should be used for implementing the given
metricName
in this container.
- getGauges(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
-
- getGcloudCancelCommand(DataflowPipelineOptions, String) - Static method in class org.apache.beam.runners.dataflow.util.MonitoringUtil
-
- getGcpCredential() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-
The credential instance that should be used to authenticate against GCP services.
- getGcpTempLocation() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-
A GCS path for storing temporary files in GCP.
- getGcsEndpoint() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-
GCS endpoint to use.
- getGcsUploadBufferSizeBytes() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-
The buffer size (in bytes) to use when uploading files to GCS.
- getGcsUtil() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-
The GcsUtil instance that should be used to communicate with Google Cloud Storage.
- getGoogleApiTrace() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions
-
This option enables tracing of API calls to Google services used within the Apache
Beam SDK.
- getGregorianCalendar(String) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a GregorianCalendar
value by field name, ClassCastException
is thrown
if type doesn't match.
- getGregorianCalendar(int) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a GregorianCalendar
value by field index, ClassCastException
is thrown
if type doesn't match.
- getHdfsConfiguration() - Method in interface org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptions
-
- getHighWatermark() - Method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.SparkWatermarks
-
- getHistograms(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
-
- getId() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
-
- getId() - Method in class org.apache.beam.sdk.values.TupleTag
-
Returns the id of this TupleTag
.
- getId() - Method in class org.apache.beam.sdk.values.ValueWithRecordId
-
- getIdAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
-
Get the id attribute.
- getIdAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
-
Get the id attribute.
- getImplementor(boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.UdafImpl
-
- getIncompatibleGlobalWindowErrorMessage() - Method in interface org.apache.beam.sdk.transforms.CombineFnBase.GlobalCombineFn
-
Returns the error message for not supported default values in Combine.globally().
- getIncompatibleGlobalWindowErrorMessage() - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
-
- getIndex() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-
- getIndex() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-
- getIndex(TupleTag<?>) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
-
Returns the index for the given tuple tag, if the tag is present in this
schema, -1 if it isn't.
- getIndex() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
-
The zero-based index of this trigger firing that produced this pane.
- getInput(PTransform<InputT, ?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
-
- getInput() - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
-
- getInput() - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- getinputFormatClass() - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.Read
-
- getinputFormatKeyClass() - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.Read
-
- getinputFormatValueClass() - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.Read
-
- getInputRef() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlInputRefExpression
-
- getInputs(PTransform<InputT, ?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
-
Returns the input of the currently being translated transform.
- getInputs() - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
-
- getInputStream(PValue) - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
-
- getInputTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.DoFn
-
Returns a
TypeDescriptor
capturing what is known statically
about the input type of this
DoFn
instance's most-derived
class.
- getInputTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.SimpleFunction
-
- getInputValueCoder(Coder<KV<K, V>>) - Static method in class org.apache.beam.sdk.transforms.GroupByKey
-
Returns the Coder
of the values of the input to this transform.
- getInsertDefault() - Method in class org.apache.beam.sdk.transforms.Combine.GloballyAsSingletonView
-
- getInstance() - Static method in class org.apache.beam.runners.spark.aggregators.AggregatorsAccumulator
-
- getInstance() - Static method in class org.apache.beam.runners.spark.metrics.MetricsAccumulator
-
- getInt(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getInt(Map<String, Object>, String, Integer) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getInteger() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlPrimitive
-
- getInteger(String) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Integer
value by field name, ClassCastException
is thrown
if type doesn't match.
- getInteger(int) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Integer
value by field index, ClassCastException
is thrown
if type doesn't match.
- getInterface() - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-
- getInterfaces() - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
Returns a set of
TypeDescriptor
s, one for each
interface implemented by this class.
- getJAXBClass() - Method in class org.apache.beam.sdk.io.xml.JAXBCoder
-
- getJmsCorrelationID() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-
- getJmsDeliveryMode() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-
- getJmsDestination() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-
- getJmsExpiration() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-
- getJmsMessageID() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-
- getJmsPriority() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-
- getJmsRedelivered() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-
- getJmsReplyTo() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-
- getJmsTimestamp() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-
- getJmsType() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-
- getJob(String) - Method in class org.apache.beam.runners.dataflow.DataflowClient
-
Gets the Dataflow Job
with the given jobId
.
- getJob() - Method in exception org.apache.beam.runners.dataflow.DataflowJobException
-
Returns the failed job.
- getJob() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator.JobSpecification
-
- getJobId() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-
Get the id of this job.
- getJobId() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
-
The identity of the Dataflow job.
- getJobId() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
-
- getJobMessages(String, long) - Method in class org.apache.beam.runners.dataflow.util.MonitoringUtil
-
Return job messages sorted in ascending order by timestamp.
- getJobMetrics(String) - Method in class org.apache.beam.runners.dataflow.DataflowClient
-
Gets the JobMetrics
with the given jobId
.
- getJobMonitoringPageURL(String, String) - Static method in class org.apache.beam.runners.dataflow.util.MonitoringUtil
-
- getJobMonitoringPageURL(String, String, String) - Static method in class org.apache.beam.runners.dataflow.util.MonitoringUtil
-
- getJobName() - Method in interface org.apache.beam.sdk.options.PipelineOptions
-
- getJsonTimePartitioning() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-
- getKey() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
-
- getKey() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
-
The key for the display item.
- getKey() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
-
The key for the display item.
- getKey() - Method in class org.apache.beam.sdk.values.KV
-
Returns the key of this
KV
.
- getKey() - Method in class org.apache.beam.sdk.values.ShardedKey
-
- getKeyCoder() - Method in class org.apache.beam.sdk.coders.KvCoder
-
- getKeyCoder() - Method in class org.apache.beam.sdk.coders.MapCoder
-
- getKeyCoder(Coder<KV<K, V>>) - Static method in class org.apache.beam.sdk.transforms.GroupByKey
-
Returns the Coder
of the keys of the input to this
transform, which is also used as the Coder
of the keys of
the output of this transform.
- getKeyCoder() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
-
- getKeyedCollections() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
-
- getKeyRange() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-
Returns the range of keys that will be read from the table.
- getKeyRange() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
-
Returns the range of keys that will be read from the table.
- getKeySet() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-
- getKeystorePassword() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-
- getKeystorePath() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-
- getKeyTranslationFunction() - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.Read
-
- getKeyTypeDescriptor() - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.Read
-
- getKindString() - Method in class org.apache.beam.sdk.io.BoundedReadFromUnboundedSource
-
- getKindString() - Method in class org.apache.beam.sdk.io.Read.Bounded
-
- getKindString() - Method in class org.apache.beam.sdk.io.Read.Unbounded
-
- getKindString() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
-
- getKindString() - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
-
- getKindString() - Method in class org.apache.beam.sdk.transforms.Combine.PerKeyWithHotKeyFanout
-
- getKindString() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-
- getKindString() - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
-
- getKindString() - Method in class org.apache.beam.sdk.transforms.PTransform
-
Returns the name to use by default for this PTransform
(not including the names of any enclosing PTransform
s).
- getKindString() - Method in class org.apache.beam.sdk.transforms.windowing.Window
-
- getKindString() - Method in class org.apache.beam.sdk.values.PValueBase
-
Returns a
String
capturing the kind of this
PValueBase
.
- getKinesisClient() - Method in interface org.apache.beam.sdk.io.kinesis.AWSClientsProvider
-
- getKV() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
-
- getLabel() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
-
Retrieve the optional label for an item.
- getLabel() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
-
The optional label for an item.
- getLabels() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-
Labels that will be applied to the billing records for this job.
- getLastEmitted() - Method in class org.apache.beam.sdk.io.CountingSource.CounterMark
-
Returns the last value emitted by the reader.
- getLastWatermarkedBatchTime() - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
-
- getLateTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-
- getLegacyEnvironmentMajorVersion() - Method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
-
Provides the legacy environment's major version number.
- getLinkUrl() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
-
Retrieve the optional link URL for an item.
- getLinkUrl() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
-
The optional link URL for an item.
- getListeners() - Method in interface org.apache.beam.runners.spark.SparkContextOptions
-
- getListOfMaps(Map<String, Object>, String, List<Map<String, Object>>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getLiteralGqlQuery() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
- getLocalhost() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
- getLocalValue() - Method in class org.apache.beam.runners.flink.metrics.MetricsAccumulator
-
- getLong(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getLong(Map<String, Object>, String, Long) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getLong() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlPrimitive
-
- getLong(String) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Long
value by field name, ClassCastException
is thrown
if type doesn't match.
- getLong(int) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Long
value by field index, ClassCastException
is thrown
if type doesn't match.
- getLowWatermark() - Method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.SparkWatermarks
-
- getMainOutputTag() - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
- getMainOutputTag() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-
- getMainTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
-
The main trigger, which will continue firing until the "until" trigger fires.
- getMaterialization() - Method in class org.apache.beam.sdk.transforms.ViewFn
-
Gets the materialization of this
ViewFn
.
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn
-
Deprecated.
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.ListViewFn
-
Deprecated.
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.MapViewFn
-
Deprecated.
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.MultimapViewFn
-
Deprecated.
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn
-
Deprecated.
- getMaxBundleSize() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- getMaxBundleTimeMills() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- getMaxConditionCost() - Method in interface org.apache.beam.runners.dataflow.options.CloudDebuggerOptions
-
The maximum cost (as a ratio of CPU time) allowed for evaluating conditional snapshots.
- getMaxEndOffset(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
-
- getMaxEndOffset(PipelineOptions) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-
Returns the actual ending offset of the current source.
- getMaxNumericPrecision() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelDataTypeSystem
-
- getMaxNumericScale() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelDataTypeSystem
-
- getMaxNumWorkers() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
The maximum number of workers to use for the workerpool.
- getMaxRecordsPerBatch() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- getMean() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
-
- getMemoryMB() - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
-
Returns the configured size of the memory buffer.
- getMessage() - Method in exception org.apache.beam.sdk.coders.Coder.NonDeterministicException
-
- getMessage() - Method in exception org.apache.beam.sdk.transforms.windowing.IncompatibleWindowException
-
- getMessages() - Method in class org.apache.beam.sdk.io.jms.JmsCheckpointMark
-
- getMessageType() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
-
Returns the Protocol Buffers
Message
type this
ProtoCoder
supports.
- getMetadata(DestinationT) - Method in class org.apache.beam.sdk.io.DynamicAvroDestinations
-
Return AVRO file metadata for a given destination.
- getMetadata() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
-
Returns the MatchResult.Metadata
of the file.
- getMeters(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
-
- getMimeType() - Method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
-
Deprecated.
- getMimeType() - Method in interface org.apache.beam.sdk.io.FileBasedSink.OutputFileHints
-
Returns the MIME type that should be used for the files that will hold the output data.
- getMinBundleSize() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-
Returns the minimum bundle size that should be used when splitting the source into sub-sources.
- getMinimumTextLength() - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- getMinReadTimeMillis() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- getMode() - Method in class org.apache.beam.sdk.io.FileBasedSource
-
- getMode() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
- getMonthOfYear() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-
- getMutableOutput(TupleTag<T>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
- getName() - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-
- getName() - Method in interface org.apache.beam.sdk.metrics.Metric
-
- getName() - Method in class org.apache.beam.sdk.metrics.MetricNameFilter
-
- getName() - Method in class org.apache.beam.sdk.transforms.PTransform
-
Returns the transform name.
- getName() - Method in class org.apache.beam.sdk.values.PCollection
-
- getName() - Method in interface org.apache.beam.sdk.values.PValue
-
Returns the name of this
PValue
.
- getName() - Method in class org.apache.beam.sdk.values.PValueBase
-
- getNameOverride() - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
-
- getNameOverride() - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
-
- getNamespace() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
- getNamespace() - Method in class org.apache.beam.sdk.metrics.MetricNameFilter
-
- getNamespace() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
-
- getNamespace() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
-
The namespace for the display item.
- getNamespace() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
-
The namespace for the display item.
- getNeedsAttributes() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
-
- getNetwork() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- getNextOffset() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
-
- getNonSpeculativeIndex() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
-
The zero-based index of this trigger firing among non-speculative panes.
- getNum() - Method in class org.apache.beam.runners.spark.io.ConsoleIO.Write.Unbound
-
- getNumber() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-
- getNumber() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-
- getNumber() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-
- getNumberOfExecutionRetries() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- getNumberOfWorkerHarnessThreads() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
Number of threads to use on the Dataflow worker harness.
- getNumQuerySplits() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
- getNumShards() - Method in class org.apache.beam.sdk.io.WriteFiles
-
- getNumWorkers() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
Number of workers to use when executing the Dataflow job.
- getObject(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getObject(Map<String, Object>, String, Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getObjectReuse() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- getOffset() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
-
- getOffset() - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-
- getOffset() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-
- getOffset() - Method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform.AlignTo
-
- getOldestPendingTimestamp() - Method in class org.apache.beam.sdk.io.jms.JmsCheckpointMark
-
- getOnCreateMatcher() - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-
- getOnly(TupleTag<V>) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
-
If there is a singleton value for the given tag, returns it.
- getOnly(TupleTag<V>, V) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
-
If there is a singleton value for the given tag, returns it.
- getOnSuccessMatcher() - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-
- getOnTimeBehavior() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
- getOperands() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlExpression
-
- getOptions() - Method in class org.apache.beam.sdk.testing.TestPipeline
-
- getOptionsId() - Method in interface org.apache.beam.sdk.options.PipelineOptions
-
Provides a unique ID for this
PipelineOptions
object, assigned at graph
construction time.
- getOrCreateReader(PipelineOptions, CheckpointMarkT) - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
-
- getOriginalWindowFn() - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
-
Returns the original windowFn that this InvalidWindows replaced.
- getOutName(int) - Method in class org.apache.beam.sdk.values.TupleTag
-
If this TupleTag
is tagging output outputIndex
of
a PTransform
, returns the name that should be used by
default for the output.
- getOutput(PTransform<?, OutputT>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
-
- getOutput() - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
-
- getOutput() - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- getOutputCoder() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
-
- getOutputCoder(SerializableFunction<InputT, OutputT>, Coder<InputT>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
-
- getOutputCoder() - Method in class org.apache.beam.sdk.io.AvroSource
-
- getOutputCoder() - Method in class org.apache.beam.sdk.io.CompressedSource
-
Returns the delegate source's output coder.
- getOutputCoder() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
-
- getOutputCoder() - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.HadoopInputFormatBoundedSource
-
- getOutputCoder() - Method in class org.apache.beam.sdk.io.jms.JmsIO.UnboundedJmsSource
-
- getOutputCoder() - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
-
- getOutputCoder() - Method in class org.apache.beam.sdk.io.Source
-
Returns the Coder
to use for the data read from this source.
- getOutputCoder() - Method in class org.apache.beam.sdk.io.xml.XmlSource
-
- getOutputKvCoder(Coder<KV<K, V>>) - Static method in class org.apache.beam.sdk.transforms.GroupByKey
-
Returns the Coder
of the output of this transform.
- getOutputs(PTransform<?, OutputT>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
-
Returns the output of the currently being translated transform.
- getOutputs() - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
-
- getOutputStrategyInternal(WindowingStrategy<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.Window
-
- getOutputTime(Instant, GlobalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-
- getOutputTime(Instant, W) - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
-
- getOutputTime(Instant, W) - Method in class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
-
- getOutputTime(Instant, IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-
Ensures that later sliding windows have an output time that is past the end of earlier windows.
- getOutputTime(Instant, W) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
-
Returns the output timestamp to use for data depending on the given
inputTimestamp
in the specified window
.
- getOutputType() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlExpression
-
- getOutputType() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlPrimitive
-
- getOutputType() - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
-
Returns a
TypeDescriptor
capturing what is known statically
about the output type of this
CombineFn
instance's
most-derived class.
- getOutputTypeDescriptor() - Method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides.BatchStatefulDoFn
-
- getOutputTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.DoFn
-
Returns a
TypeDescriptor
capturing what is known statically
about the output type of this
DoFn
instance's
most-derived class.
- getOutputTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.SimpleFunction
-
- getOverrideWindmillBinary() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
Custom windmill_main binary to use with the streaming runner.
- getPane() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow
-
Returns the pane of this ValueInSingleWindow
in its window.
- getPaneInfo() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-
- getParallelism() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- getParallelism() - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
-
- getParameters() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.UdafImpl
-
- getParseSynchronously() - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- getPartition() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
-
- getPartition() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
-
- getPartitionKey() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
-
- getPartitions() - Method in class org.apache.beam.runners.spark.io.SourceRDD.Bounded
-
- getPartitions() - Method in class org.apache.beam.runners.spark.io.SourceRDD.Unbounded
-
- getPartitions() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark
-
- getPassword() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-
- getPath() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
-
- getPath() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
-
The path for the display item within a component hierarchy.
- getPathValidator() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-
The path validator instance that should be used to validate paths.
- getPathValidatorClass() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-
The class of the validator that should be created and used to validate paths.
- getPayload() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
-
Returns the main PubSub message.
- getPayload() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-
- getPCollection() - Method in interface org.apache.beam.sdk.values.PCollectionView
-
For internal use only.
- getPCollection() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
-
- getPerDestinationOutputFilenames() - Method in class org.apache.beam.sdk.io.WriteFilesResult
-
Returns a
PCollection
of all output filenames generated by this
WriteFiles
organized by user destination type.
- getPeriod() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-
- getPeriod() - Method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform.AlignTo
-
- getPipeline() - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
-
- getPipeline() - Method in class org.apache.beam.sdk.io.WriteFilesResult
-
- getPipeline() - Method in class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
-
- getPipeline() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
-
- getPipeline() - Method in class org.apache.beam.sdk.values.PBegin
-
- getPipeline() - Method in class org.apache.beam.sdk.values.PCollectionList
-
- getPipeline() - Method in class org.apache.beam.sdk.values.PCollectionTuple
-
- getPipeline() - Method in class org.apache.beam.sdk.values.PDone
-
- getPipeline() - Method in interface org.apache.beam.sdk.values.PInput
-
- getPipeline() - Method in interface org.apache.beam.sdk.values.POutput
-
- getPipeline() - Method in class org.apache.beam.sdk.values.PValueBase
-
- getPipelineName() - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- getPipelineOptions() - Method in class org.apache.beam.runners.apex.ApexRunnerRegistrar.Options
-
- getPipelineOptions() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineRegistrar.Options
-
- getPipelineOptions() - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
-
Returns the configured pipeline options.
- getPipelineOptions() - Method in class org.apache.beam.runners.direct.DirectRegistrar.Options
-
- getPipelineOptions() - Method in class org.apache.beam.runners.direct.DirectRunner
-
- getPipelineOptions() - Method in class org.apache.beam.runners.flink.FlinkRunner
-
For testing.
- getPipelineOptions() - Method in class org.apache.beam.runners.flink.FlinkRunnerRegistrar.Options
-
- getPipelineOptions() - Method in class org.apache.beam.runners.flink.TestFlinkRunner
-
- getPipelineOptions() - Method in class org.apache.beam.runners.gearpump.GearpumpRunnerRegistrar.Options
-
- getPipelineOptions() - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
-
- getPipelineOptions() - Method in class org.apache.beam.runners.spark.SparkRunnerRegistrar.Options
-
- getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpPipelineOptionsRegistrar
-
- getPipelineOptions() - Method in class org.apache.beam.sdk.io.gcp.common.GcpIoPipelineOptionsRegistrar
-
- getPipelineOptions() - Method in class org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptionsRegistrar
-
- getPipelineOptions() - Method in class org.apache.beam.sdk.options.DefaultPipelineOptionsRegistrar
-
- getPipelineOptions() - Method in interface org.apache.beam.sdk.options.PipelineOptionsRegistrar
-
- getPipelineOptions() - Method in interface org.apache.beam.sdk.state.StateContext
-
- getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.CombineWithContext.Context
-
Returns the
PipelineOptions
specified with the
PipelineRunner
invoking this
KeyedCombineFn
.
- getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.DoFn.FinishBundleContext
-
Returns the
PipelineOptions
specified with the
PipelineRunner
invoking this
DoFn
.
- getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.DoFn.StartBundleContext
-
Returns the
PipelineOptions
specified with the
PipelineRunner
invoking this
DoFn
.
- getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
-
Returns the
PipelineOptions
specified with the
PipelineRunner
invoking this
DoFn
.
- getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
- getPipelineRunners() - Method in class org.apache.beam.runners.apex.ApexRunnerRegistrar.Runner
-
- getPipelineRunners() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineRegistrar.Runner
-
- getPipelineRunners() - Method in class org.apache.beam.runners.direct.DirectRegistrar.Runner
-
- getPipelineRunners() - Method in class org.apache.beam.runners.flink.FlinkRunnerRegistrar.Runner
-
- getPipelineRunners() - Method in class org.apache.beam.runners.gearpump.GearpumpRunnerRegistrar.Runner
-
- getPipelineRunners() - Method in class org.apache.beam.runners.spark.SparkRunnerRegistrar.Runner
-
- getPlanner() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
-
- getPlanner() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamQueryPlanner
-
- getPositionForFractionConsumed(double) - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-
Returns a position P
such that the range [start, P)
represents approximately
the given fraction of the range [start, end)
.
- getProcessingTimeAdvance() - Method in class org.apache.beam.sdk.testing.TestStream.ProcessingTimeEvent
-
- getProducer(PValue) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
-
Get the
AppliedPTransform
that produced the provided
PValue
.
- getProfilingAgentConfiguration() - Method in interface org.apache.beam.runners.dataflow.options.DataflowProfilingOptions
-
- getProject() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-
- getProject() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-
Project id to use when launching jobs.
- getProject() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
-
Get the project path.
- getProjectId() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-
Get the project this job exists in.
- getProjectId() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
-
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
- getProperties() - Method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
-
- getProperties() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-
- getProvidedSparkContext() - Method in interface org.apache.beam.runners.spark.SparkContextOptions
-
- getProviderRuntimeValues() - Method in interface org.apache.beam.sdk.testing.TestPipeline.TestValueProviderOptions
-
- getPTransformForInput() - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.kafka.BeamKafkaCSVTable
-
- getPTransformForInput() - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.kafka.BeamKafkaTable
-
- getPTransformForOutput() - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.kafka.BeamKafkaCSVTable
-
- getPTransformForOutput() - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.kafka.BeamKafkaTable
-
- getPubsubRootUrl() - Method in interface org.apache.beam.sdk.io.gcp.pubsub.PubsubOptions
-
Root URL for use with the Google Cloud Pub/Sub API.
- getQuery() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
- getQuery() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-
- getQueueMaxPollTime() - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- getQueuePollTime() - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- getRange() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-
Returns the current range.
- getRawType() - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
Returns the
Class
underlying the
Type
represented by
this
TypeDescriptor
.
- getReadDurationMillis() - Method in class org.apache.beam.runners.spark.io.SparkUnboundedSource.Metadata
-
- getReadOutputMetadata() - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- getReadTime() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
-
- getReadTimePercentage() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- getReason() - Method in exception org.apache.beam.sdk.coders.CannotProvideCoderException
-
- getReasons() - Method in exception org.apache.beam.sdk.coders.Coder.NonDeterministicException
-
- getRecord() - Method in class org.apache.beam.sdk.io.gcp.bigquery.SchemaAndRecord
-
- getRecordCoder() - Method in class org.apache.beam.sdk.values.BeamRecordType
-
- getRecordType() - Method in class org.apache.beam.sdk.coders.BeamRecordCoder
-
- getRecordType() - Method in class org.apache.beam.sdk.coders.SerializableCoder
-
- getRegion() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-
Get the region this job exists in.
- getRegion() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-
The Google Compute Engine
region
for creating Dataflow jobs.
- getRegisteredOptions() - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
-
- getRepeatedTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
-
- getReplacedByJob() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-
- getReplacedByJob() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
-
- getReplacementTransform(AppliedPTransform<PCollection<? extends InputT>, PCollection<OutputT>, ParDo.SingleOutput<InputT, OutputT>>) - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory
-
- getRetainExternalizedCheckpointsOnCancellation() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- getReturnType(RelDataTypeFactory) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.UdafImpl
-
- getRingFraction(List<CassandraServiceImpl.TokenRange>) - Static method in class org.apache.beam.sdk.io.cassandra.CassandraServiceImpl
-
Compute the percentage of token addressed compared with the whole tokens in the cluster.
- getRootCause() - Method in exception org.apache.beam.sdk.coders.CannotProvideCoderException
-
- getRowType() - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.BaseBeamTable
-
- getRowType() - Method in interface org.apache.beam.sdk.extensions.sql.impl.schema.BeamSqlTable
-
Get the schema info of the table.
- getRuleSets() - Static method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRuleSets
-
- getRunMillis() - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
-
- getRunner() - Method in interface org.apache.beam.sdk.options.PipelineOptions
-
The pipeline runner that will be used to execute the pipeline.
- getSaveProfilesToGcs() - Method in interface org.apache.beam.runners.dataflow.options.DataflowProfilingOptions
-
- getSchema() - Method in class org.apache.beam.sdk.coders.AvroCoder
-
Returns the schema used by this coder.
- getSchema(DestinationT) - Method in class org.apache.beam.sdk.io.DynamicAvroDestinations
-
Return an AVRO schema for a given destination.
- getSchema(DestinationT) - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
-
Returns the table schema for the destination.
- getSchema() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
-
- getSchema() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
-
- getScheme() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
-
- getScheme() - Method in class org.apache.beam.sdk.io.FileSystem
-
Get the URI scheme which defines the namespace of the
FileSystem
.
- getScheme() - Method in interface org.apache.beam.sdk.io.fs.ResourceId
-
Get the scheme which defines the namespace of the
ResourceId
.
- getSequenceNumber() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
-
- getSerializers() - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
-
- getServiceAccount() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-
Run the job as a specific service account, instead of the default GCE robot.
- getShard() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-
- getShardId() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
-
- getSharding() - Method in class org.apache.beam.sdk.io.WriteFiles
-
Gets the
PTransform
that will be used to determine sharding.
- getShardNumber() - Method in class org.apache.beam.sdk.values.ShardedKey
-
- getShort() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlPrimitive
-
- getShort(String) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Short
value by field name, ClassCastException
is thrown
if type doesn't match.
- getShort(int) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a Short
value by field index, ClassCastException
is thrown
if type doesn't match.
- getShortValue() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
-
Return the optional short value for an item, or null if none is provided.
- getShortValue() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
-
The optional short value for an item, or null
if none is provided.
- getSideInputs() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.ParDoSingle
-
- getSideInputs() - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
-
Override to specify that this object needs access to one or more side inputs.
- getSideInputs() - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
-
Specifies that this object needs access to one or more side inputs.
- getSideInputs() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
-
Returns the side inputs used by this Combine operation.
- getSideInputs() - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
-
- getSideInputs() - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
-
Returns the side inputs used by this Combine operation.
- getSideInputs() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-
- getSideInputs() - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
-
- getSideInputWindow(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.WindowMappingFn
-
Returns the window of the side input corresponding to the given window of the main input.
- getSingleFileMetadata() - Method in class org.apache.beam.sdk.io.FileBasedSource
-
Returns the information about the single file that this source is reading from.
- getSink() - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-
Returns the FileBasedSink for this write operation.
- getSink() - Method in class org.apache.beam.sdk.io.WriteFiles
-
- getSize() - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-
- getSize() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-
- getSource() - Method in class org.apache.beam.sdk.io.Read.Bounded
-
Returns the BoundedSource
used to create this Read
PTransform
.
- getSource() - Method in class org.apache.beam.sdk.io.Read.Unbounded
-
Returns the UnboundedSource
used to create this Read
PTransform
.
- getSource() - Method in class org.apache.beam.sdk.io.TextIO.Read
-
- getSource() - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
-
- getSourceStream(DataSource) - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
-
- getSourceTables() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamQueryPlanner
-
- getSourceType() - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamPCollectionTable
-
- getSourceType() - Method in interface org.apache.beam.sdk.extensions.sql.impl.schema.BeamSqlTable
-
In Beam SQL, there's no difference between a batch query and a streaming
query.
- getSourceType() - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.kafka.BeamKafkaTable
-
- getSourceType() - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.text.BeamTextTable
-
- getSparkMaster() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- getSplit() - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.SerializableSplit
-
- getSplitBacklogBytes() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
-
Returns the size of the backlog of unread data in the underlying data source represented by
this split of this source.
- getSplitPointsConsumed() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
-
- getSplitPointsConsumed() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-
- getSplitPointsConsumed() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
- getSplitPointsConsumed() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-
- getSplitPointsProcessed() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-
Returns the total number of split points that have been processed.
- getSplitPointsRemaining() - Method in class org.apache.beam.sdk.io.AvroSource.AvroReader
-
- getSplitPointsRemaining() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
-
- getSplitPointsRemaining() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-
- getSplitPointsRemaining() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
- getSqlRecordType(BeamRecord) - Static method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper
-
- getStableUniqueNames() - Method in interface org.apache.beam.sdk.options.PipelineOptions
-
Whether to check for stable unique names on each transform.
- getStager() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
The resource stager instance that should be used to stage resources.
- getStagerClass() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
The class responsible for staging resources to be accessible by workers
during job execution.
- getStagingLocation() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-
GCS path for staging local files, e.g.
- getStartDate() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-
- getStartDate() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-
- getStartDate() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-
- getStartKey() - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
-
- getStartOffset() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-
Returns the starting offset of the source.
- getStartPosition() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-
- getStartPosition() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-
- getStartPosition() - Method in interface org.apache.beam.sdk.io.range.RangeTracker
-
Returns the starting position of the current range, inclusive.
- getStartTime() - Method in class org.apache.beam.sdk.io.CountingSource.CounterMark
-
Returns the time the reader was started.
- getState() - Method in class org.apache.beam.runners.apex.ApexRunnerResult
-
- getState() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-
- getState() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
-
- getState() - Method in class org.apache.beam.runners.direct.DirectRunner.DirectPipelineResult
-
- getState() - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
-
- getState() - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
-
- getState() - Method in class org.apache.beam.runners.gearpump.GearpumpPipelineResult
-
- getState(JobApi.GetJobStateRequest, StreamObserver<JobApi.GetJobStateResponse>) - Method in class org.apache.beam.runners.reference.job.ReferenceRunnerJobService
-
- getState() - Method in class org.apache.beam.runners.spark.SparkPipelineResult
-
- getState() - Method in interface org.apache.beam.sdk.PipelineResult
-
Retrieves the current state of the pipeline execution.
- getStateBackend() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
State backend to store Beam's state during computation.
- getStateCoder() - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
-
- getStepNames() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator.JobSpecification
-
Returns the mapping of AppliedPTransforms
to the internal step
name for that AppliedPTransform
.
- getStopPipelineWatermark() - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
-
- getStopPosition() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-
- getStopPosition() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-
- getStopPosition() - Method in interface org.apache.beam.sdk.io.range.RangeTracker
-
Returns the ending position of the current range, exclusive.
- getStorageLevel() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- getStreamName() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
-
- getString(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getString(Map<String, Object>, String, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getString() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlPrimitive
-
- getString(String) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a String
value by field name, ClassCastException
is thrown
if type doesn't match.
- getString(int) - Method in class org.apache.beam.sdk.values.BeamRecord
-
Get a String
value by field index, ClassCastException
is thrown
if type doesn't match.
- getStrings(Map<String, Object>, String, List<String>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-
- getSubnetwork() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- getSubscription() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
-
Get the subscription being read from.
- getSubscriptionProvider() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
-
- getSubSequenceNumber() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
-
- getSuggestedFilenameSuffix() - Method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
-
Deprecated.
- getSuggestedFilenameSuffix() - Method in interface org.apache.beam.sdk.io.FileBasedSink.OutputFileHints
-
- getSuggestedSuffix() - Method in enum org.apache.beam.sdk.io.Compression
-
- getSum() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
-
- getSumAndReset() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
-
- getSupertype(Class<? super T>) - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
Returns the generic form of a supertype.
- getSupportedClass() - Method in interface org.apache.beam.runners.dataflow.util.CloudObjectTranslator
-
- getSynchronizedProcessingTime() - Method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.SparkWatermarks
-
- getTable() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
-
Returns the table to read, or null
if reading from a query instead.
- getTable() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-
- getTable() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
Returns the table reference, or null
.
- getTable(DestinationT) - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
-
- getTable() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-
- getTableDescription() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-
- getTableId() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-
Returns the table being read from.
- getTableId() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
-
- getTableId() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
-
- getTableProvider() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
-
Returns the table to read, or null
if reading from a query instead.
- getTableProvider() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-
- getTableReference() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-
- getTableSchema() - Method in class org.apache.beam.sdk.io.gcp.bigquery.SchemaAndRecord
-
- getTableSpec() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-
- getTag(int) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
-
Returns the tuple tag at the given index.
- getTag() - Method in class org.apache.beam.sdk.values.TaggedPValue
-
Returns the local tag associated with the
PValue
.
- getTagInternal() - Method in interface org.apache.beam.sdk.values.PCollectionView
-
- getTagInternal() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
-
- getTagsToSideInputs(Collection<PCollectionView<?>>) - Static method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils
-
- getTargetParallelism() - Method in interface org.apache.beam.runners.direct.DirectOptions
-
- getTempDatasetId() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-
- getTempDirectoryProvider() - Method in class org.apache.beam.sdk.io.FileBasedSink
-
- getTempFilename() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-
- getTemplateLocation() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-
Where the runner should generate a template file.
- getTempLocation() - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
-
Returns the configured temporary location.
- getTempLocation() - Method in interface org.apache.beam.sdk.options.PipelineOptions
-
A pipeline level default location for storing temporary files.
- getTempRoot() - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-
- getTestTimeoutSeconds() - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-
- getTikaConfigPath() - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- getTimeDomain() - Method in interface org.apache.beam.sdk.state.TimerSpec
-
- getTimePartitioning() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-
- getTimers(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
-
- getTimes() - Method in class org.apache.beam.runners.spark.io.CreateStream
-
- getTimestamp() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
-
- getTimestamp() - Method in class org.apache.beam.sdk.values.TimestampedValue
-
- getTimestamp() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow
-
Returns the timestamp of this ValueInSingleWindow
.
- getTimestampAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
-
Get the timestamp attribute.
- getTimestampAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
-
Get the timestamp attribute.
- getTimestampCombiner() - Method in interface org.apache.beam.sdk.state.WatermarkHoldState
-
Return the
TimestampCombiner
which will be used to determine a watermark hold time
given an element timestamp, and to combine watermarks from windows which are about to be
merged.
- getTimestampCombiner() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
- getTimestampTransforms() - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
-
The transforms applied to the arrival time of an element to determine when this trigger allows
output.
- getTimeZone() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-
- getTimeZone() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-
- getTimeZone() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-
- getTiming() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
-
Return the timing of this pane.
- getTo() - Method in class org.apache.beam.sdk.io.range.OffsetRange
-
- getTopic() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
-
Get the topic being written to.
- getTopic() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
-
Get the topic being read from.
- getTopic() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
-
- getTopic() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
-
- getTopicProvider() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
-
- getTopicProvider() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
-
- getTotalBacklogBytes() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
-
Returns the size of the backlog of unread data in the underlying data source represented by
all splits of this source.
- getTraitDef() - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-
- getTransformNameMapping() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
Mapping of old PTranform names to new ones, specified as JSON
{"oldName":"newName",...}
.
- getTransformTranslator(Class<TransformT>) - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator
-
Returns the
TransformTranslator
to use for instances of the
specified PTransform class, or null if none registered.
- getTranslator() - Method in class org.apache.beam.runners.dataflow.DataflowRunner
-
Returns the DataflowPipelineTranslator associated with this object.
- getTrigger() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
- getTupleTag() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple.TaggedKeyedPCollection
-
Returns the TupleTag of this TaggedKeyedPCollection.
- getTupleTagList() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
-
Returns the TupleTagList tuple associated with this schema.
- getType() - Method in class org.apache.beam.sdk.coders.AvroCoder
-
Returns the type this coder encodes/decodes.
- getType() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-
- getType() - Method in interface org.apache.beam.sdk.testing.TestStream.Event
-
- getType() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
-
- getType() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
-
- getType() - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollection
-
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.TupleTag
-
Returns a TypeDescriptor
capturing what is known statically
about the type of this TupleTag
instance's most-derived
class.
- getTypeParameter(String) - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
Returns a TypeVariable
for the named type parameter.
- getTypes() - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
Returns a set of
TypeDescriptor
, one for each
superclass as well as each interface implemented by this class.
- getUnderlyingDoFn() - Method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides.BatchStatefulDoFn
-
- getUnionCoder() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
-
- getUnionTag() - Method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils.RawUnionValue
-
- getUnionTag() - Method in class org.apache.beam.sdk.transforms.join.RawUnionValue
-
- getUniqueId() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
-
- getUntilTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
-
The trigger that signals termination of this trigger.
- getUrn() - Method in interface org.apache.beam.sdk.transforms.Materialization
-
- getUsePublicIps() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
Specifies whether worker pools should be started with public IP addresses.
- getUserAgent() - Method in interface org.apache.beam.sdk.options.PipelineOptions
-
A user agent string as per RFC2616, describing the pipeline to external services.
- getUsername() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-
- getUsesProvidedSparkContext() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- getValidate() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-
- getValue() - Method in class org.apache.beam.runners.flink.metrics.FlinkMetricContainer.FlinkDistributionGauge
-
- getValue() - Method in class org.apache.beam.runners.flink.metrics.FlinkMetricContainer.FlinkGauge
-
- getValue() - Method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils.RawUnionValue
-
- getValue(String, Class<T>) - Method in class org.apache.beam.runners.spark.aggregators.NamedAggregators
-
- getValue() - Method in class org.apache.beam.runners.spark.util.ByteArray
-
- getValue() - Method in class org.apache.beam.runners.spark.util.SideInputBroadcast
-
- getValue() - Method in class org.apache.beam.sdk.coders.StructuralByteArray
-
- getValue() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlPrimitive
-
- getValue() - Method in class org.apache.beam.sdk.io.range.ByteKey
-
Returns a read-only
ByteBuffer
representing this
ByteKey
.
- getValue() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
-
Retrieve the value of the display item.
- getValue() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
-
The value of the display item.
- getValue() - Method in class org.apache.beam.sdk.transforms.join.RawUnionValue
-
- getValue() - Method in class org.apache.beam.sdk.values.KV
-
Returns the value of this
KV
.
- getValue() - Method in class org.apache.beam.sdk.values.TaggedPValue
-
- getValue() - Method in class org.apache.beam.sdk.values.TimestampedValue
-
- getValue() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow
-
Returns the value of this ValueInSingleWindow
.
- getValue() - Method in class org.apache.beam.sdk.values.ValueWithRecordId
-
- getValueCoder() - Method in class org.apache.beam.sdk.coders.KvCoder
-
- getValueCoder() - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
-
Gets the value coder that will be prefixed by the length.
- getValueCoder() - Method in class org.apache.beam.sdk.coders.MapCoder
-
- getValueCoder() - Method in class org.apache.beam.sdk.coders.NullableCoder
-
- getValueCoder() - Method in class org.apache.beam.sdk.testing.TestStream
-
- getValueCoder() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
-
- getValueCoder() - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
-
- getValueTranslationFunction() - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.Read
-
- getValueTypeDescriptor() - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.Read
-
- getView() - Method in class org.apache.beam.runners.apex.ApexRunner.CreateApexPCollectionView
-
- getView() - Method in class org.apache.beam.runners.dataflow.CreateDataflowView
-
- getView() - Method in class org.apache.beam.runners.dataflow.DataflowRunner.StreamingPCollectionViewWriterFn
-
- getView() - Method in class org.apache.beam.sdk.transforms.View.CreatePCollectionView
-
- getViewFn() - Method in interface org.apache.beam.sdk.values.PCollectionView
-
- getViewFn() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
-
- getWatermark() - Method in class org.apache.beam.runners.gearpump.translators.io.GearpumpSource
-
- getWatermark() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-
- getWatermark() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
-
Returns a timestamp before or at the timestamps of all future elements read by this reader.
- getWatermark() - Method in class org.apache.beam.sdk.testing.TestStream.WatermarkEvent
-
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterAll
-
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterEach
-
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterFirst
-
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
-
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
-
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterSynchronizedProcessingTime
-
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
-
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.DefaultTrigger
-
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.Never.NeverTrigger
-
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
-
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
-
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.ReshuffleTrigger
-
Deprecated.
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
-
For internal use only; no backwards-compatibility guarantees.
- getWindmillServiceEndpoint() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
Custom windmill service endpoint.
- getWindmillServicePort() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
- getWindow() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-
- getWindow() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow
-
Returns the window of this ValueInSingleWindow
.
- getWindowFn() - Method in class org.apache.beam.sdk.transforms.windowing.Window.Assign
-
- getWindowFn() - Method in class org.apache.beam.sdk.transforms.windowing.Window
-
- getWindowFn() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
- getWindowingStrategy() - Method in class org.apache.beam.sdk.values.PCollection
-
- getWindowingStrategyInternal() - Method in interface org.apache.beam.sdk.values.PCollectionView
-
- getWindowingStrategyInternal() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
-
- getWindowMappingFn() - Method in interface org.apache.beam.sdk.values.PCollectionView
-
For internal use only.
- getWindowMappingFn() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
-
- getWindowTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
-
Returns a
TypeDescriptor
capturing what is known statically about the window type of
this
WindowFn
instance's most-derived class.
- getWorkerCacheMb() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
-
The size of the worker's in-memory cache, in megabytes.
- getWorkerDiskType() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
Specifies what type of persistent disk is used.
- getWorkerHarnessContainerImage() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
Docker container image that executes Dataflow worker harness, residing in Google Container
Registry.
- getWorkerId() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
-
The identity of the worker running this pipeline.
- getWorkerLogLevelOverrides() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
-
This option controls the log levels for specifically named loggers.
- getWorkerMachineType() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
Machine type to create Dataflow worker VMs as.
- getWorkerSystemErrMessageLevel() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
-
Controls the log level given to messages printed to System.err
.
- getWorkerSystemOutMessageLevel() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
-
Controls the log level given to messages printed to System.out
.
- getWritableByteChannelFactory() - Method in class org.apache.beam.sdk.io.FileBasedSink
-
- getWriteOperation() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
-
Return the WriteOperation that this Writer belongs to.
- getYarnDeployDependencies() - Static method in class org.apache.beam.runners.apex.ApexYarnLauncher
-
From the current classpath, find the jar files that need to be deployed
with the application to run on YARN.
- getZone() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- getZone() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-
- global(Map<Integer, GlobalWatermarkHolder.SparkWatermarks>) - Static method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-
Build a global TimerInternals
for all feeding streams.
- globalDefault() - Static method in class org.apache.beam.sdk.values.WindowingStrategy
-
Return a fully specified, default windowing strategy.
- globally(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles
-
Returns a PTransform
that takes a PCollection<T>
and returns a PCollection<List<T>>
whose single value is a
List
of the approximate N
-tiles of the elements
of the input PCollection
.
- globally(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles
-
- globally(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateUnique
-
Returns a PTransform
that takes a PCollection<T>
and returns a PCollection<Long>
containing a single value
that is an estimate of the number of distinct elements in the
input PCollection
.
- globally(double) - Static method in class org.apache.beam.sdk.transforms.ApproximateUnique
-
- globally(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
-
Returns a
Combine.Globally
PTransform
that uses the given
SerializableFunction
to combine all
the elements in each window of the input
PCollection
into a
single value in the output
PCollection
.
- globally(CombineFnBase.GlobalCombineFn<? super InputT, ?, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Combine
-
Returns a
Combine.Globally
PTransform
that uses the given
GloballyCombineFn
to combine all
the elements in each window of the input
PCollection
into a
single value in the output
PCollection
.
- globally() - Static method in class org.apache.beam.sdk.transforms.Count
-
- globally() - Static method in class org.apache.beam.sdk.transforms.Latest
-
Returns a
PTransform
that takes as input a
PCollection<T>
and returns a
PCollection<T>
whose contents is the latest element according to its event time, or
null if there are no elements.
- globally() - Static method in class org.apache.beam.sdk.transforms.Max
-
Returns a PTransform
that takes an input PCollection<T>
and returns a PCollection<T>
whose contents is the maximum according to the natural ordering of T
of the input PCollection
's elements, or null
if there are no elements.
- globally(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Max
-
Returns a PTransform
that takes an input PCollection<T>
and returns a PCollection<T>
whose contents is the maximum of the input PCollection
's elements, or
null
if there are no elements.
- globally() - Static method in class org.apache.beam.sdk.transforms.Mean
-
Returns a PTransform
that takes an input
PCollection<NumT>
and returns a
PCollection<Double>
whose contents is the mean of the
input PCollection
's elements, or
0
if there are no elements.
- globally() - Static method in class org.apache.beam.sdk.transforms.Min
-
Returns a PTransform
that takes an input PCollection<T>
and returns a PCollection<T>
whose contents is the minimum according to the natural ordering of T
of the input PCollection
's elements, or null
if there are no elements.
- globally(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Min
-
Returns a PTransform
that takes an input PCollection<T>
and returns a PCollection<T>
whose contents is the minimum of the input PCollection
's elements, or
null
if there are no elements.
- GlobalWatermarkHolder - Class in org.apache.beam.runners.spark.util
-
A BlockStore
variable to hold the global watermarks for a micro-batch.
- GlobalWatermarkHolder() - Constructor for class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
-
- GlobalWatermarkHolder.SparkWatermarks - Class in org.apache.beam.runners.spark.util
-
- GlobalWatermarkHolder.WatermarkAdvancingStreamingListener - Class in org.apache.beam.runners.spark.util
-
Advance the WMs onBatchCompleted event.
- GlobalWindow - Class in org.apache.beam.sdk.transforms.windowing
-
The default window into which all data is placed (via
GlobalWindows
).
- GlobalWindow.Coder - Class in org.apache.beam.sdk.transforms.windowing
-
- GlobalWindows - Class in org.apache.beam.sdk.transforms.windowing
-
A
WindowFn
that assigns all data to the same window.
- GlobalWindows() - Constructor for class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-
- GoogleApiDebugOptions - Interface in org.apache.beam.sdk.extensions.gcp.options
-
These options configure debug settings for Google API clients created within the Apache Beam SDK.
- GoogleApiDebugOptions.GoogleApiTracer - Class in org.apache.beam.sdk.extensions.gcp.options
-
- GoogleApiTracer() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions.GoogleApiTracer
-
- GraphiteSink - Class in org.apache.beam.runners.spark.metrics.sink
-
- GraphiteSink(Properties, MetricRegistry, SecurityManager) - Constructor for class org.apache.beam.runners.spark.metrics.sink.GraphiteSink
-
- greaterThan(T) - Static method in class org.apache.beam.sdk.transforms.Filter
-
Returns a PTransform
that takes an input
PCollection<T>
and returns a PCollection<T>
with
elements that are greater than a given value, based on the
elements' natural ordering.
- greaterThanEq(T) - Static method in class org.apache.beam.sdk.transforms.Filter
-
Returns a PTransform
that takes an input
PCollection<T>
and returns a PCollection<T>
with
elements that are greater than or equal to a given value, based on
the elements' natural ordering.
- groupAlsoByWindow(JavaDStream<WindowedValue<KV<K, Iterable<WindowedValue<InputT>>>>>, Coder<K>, Coder<WindowedValue<InputT>>, WindowingStrategy<?, W>, SerializablePipelineOptions, List<Integer>, String) - Static method in class org.apache.beam.runners.spark.stateful.SparkGroupAlsoByWindowViaWindowSet
-
- groupBy(WindowedValue<KV<K, V>>) - Method in class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator.GroupByFn
-
- GroupByKey<K,V> - Class in org.apache.beam.sdk.transforms
-
GroupByKey<K, V>
takes a PCollection<KV<K, V>>
,
groups the values by key and windows, and returns a
PCollection<KV<K, Iterable<V>>>
representing a map from
each distinct key and window of the input PCollection
to an
Iterable
over all the values associated with that key in
the input per window.
- GroupByKeyTranslator<K,V> - Class in org.apache.beam.runners.gearpump.translators
-
GroupByKey
is translated to Gearpump groupBy function.
- GroupByKeyTranslator() - Constructor for class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator
-
- GroupByKeyTranslator.GearpumpWindowFn<T,W extends BoundedWindow> - Class in org.apache.beam.runners.gearpump.translators
-
A transform used internally to translate Beam's Window to Gearpump's Window.
- GroupByKeyTranslator.GroupByFn<K,V> - Class in org.apache.beam.runners.gearpump.translators
-
A transform used internally to group KV message by its key.
- GroupByKeyTranslator.KeyedByTimestamp<K,V> - Class in org.apache.beam.runners.gearpump.translators
-
A transform used internally to transform WindowedValue to KV.
- GroupByKeyTranslator.Merge<K,V> - Class in org.apache.beam.runners.gearpump.translators
-
A transform used internally by Gearpump which encapsulates the merge logic.
- grouped() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-
- groupedValues(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
-
Returns a
Combine.GroupedValues
PTransform
that takes a
PCollection
of
KV
s where a key maps to an
Iterable
of values, e.g.,
the result of a
GroupByKey
, then uses the given
SerializableFunction
to combine all the values associated
with a key, ignoring the key.
- groupedValues(CombineFnBase.GlobalCombineFn<? super InputT, ?, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Combine
-
Returns a
Combine.GroupedValues
PTransform
that takes a
PCollection
of
KV
s where a key maps to an
Iterable
of values, e.g.,
the result of a
GroupByKey
, then uses the given
CombineFn
to combine all the values associated with a
key, ignoring the key.
- GroupingState<InputT,OutputT> - Interface in org.apache.beam.sdk.state
-
A
ReadableState
cell that combines multiple input values and outputs a single value of a
different type.
- GroupIntoBatches<K,InputT> - Class in org.apache.beam.sdk.transforms
-
A
PTransform
that batches inputs to a desired batch size.
- Growth() - Constructor for class org.apache.beam.sdk.transforms.Watch.Growth
-
- growthOf(Watch.Growth.PollFn<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Watch
-
Watches the growth of the given poll function.
- Identifier() - Constructor for class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
-
- identity() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
-
Returns the identity element of this operation, i.e.
- identity() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
-
Returns the value that should be used for the combine of the empty set.
- identity() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
-
Returns the identity element of this operation, i.e.
- identity() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
-
Returns the identity element of this operation, i.e.
- identity() - Static method in class org.apache.beam.sdk.transforms.SerializableFunctions
-
- ignoreInput(Watch.Growth.TerminationCondition<?, StateT>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
-
Wraps a given input-independent
Watch.Growth.TerminationCondition
as an equivalent condition
with a given input type, passing
null
to the original condition as input.
- immediate(T) - Static method in class org.apache.beam.sdk.state.ReadableStates
-
A
ReadableState
constructed from a constant value, hence immediately available.
- immutableNames() - Method in class org.apache.beam.sdk.metrics.MetricsFilter
-
- immutableNamesBuilder() - Method in class org.apache.beam.sdk.metrics.MetricsFilter.Builder
-
- immutableSteps() - Method in class org.apache.beam.sdk.metrics.MetricsFilter
-
- immutableStepsBuilder() - Method in class org.apache.beam.sdk.metrics.MetricsFilter.Builder
-
- in(Pipeline) - Static method in class org.apache.beam.sdk.values.PBegin
-
- in(Pipeline) - Static method in class org.apache.beam.sdk.values.PDone
-
- inc() - Method in interface org.apache.beam.sdk.metrics.Counter
-
Increment the counter.
- inc(long) - Method in interface org.apache.beam.sdk.metrics.Counter
-
Increment the counter by the given amount.
- include(String, HasDisplayData) - Method in interface org.apache.beam.sdk.transforms.display.DisplayData.Builder
-
Register display data from the specified subcomponent at the given path.
- inCombinedNonLatePanes(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
-
Creates a new
PAssert.IterableAssert
like this one, but with the assertion restricted to only
run on the provided window across all panes that were not produced by the arrival of late
data.
- inCombinedNonLatePanes(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-
- IncompatibleWindowException - Exception in org.apache.beam.sdk.transforms.windowing
-
- IncompatibleWindowException(WindowFn<?, ?>, String) - Constructor for exception org.apache.beam.sdk.transforms.windowing.IncompatibleWindowException
-
- incomplete(List<TimestampedValue<OutputT>>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
-
Constructs a
Watch.Growth.PollResult
with the given outputs and declares that new outputs might
appear for the current input.
- incomplete(Instant, List<OutputT>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
-
- increment() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
-
Returns a RandomAccessData that is the smallest value of same length which
is strictly greater than this.
- INDEX_OF_MAX - Static variable in class org.apache.beam.sdk.io.ShardNameTemplate
-
Shard name containing the index and max.
- inEarlyGlobalWindowPanes() - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
-
- inEarlyGlobalWindowPanes() - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-
- inferType(Object) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
-
- inFinalPane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
-
Creates a new
PAssert.IterableAssert
like this one, but with the assertion restricted to only
run on the provided window, running the checker only on the final pane for each key.
- inFinalPane(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-
- inFinalPane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
-
Creates a new
PAssert.SingletonAssert
like this one, but with the assertion restricted to
only run on the provided window, running the checker only on the final pane for each key.
- init() - Method in class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator.Merge
-
- init(SparkPipelineOptions, JavaSparkContext) - Static method in class org.apache.beam.runners.spark.aggregators.AggregatorsAccumulator
-
Init aggregators accumulator if it has not been initiated.
- init(SparkPipelineOptions, JavaSparkContext) - Static method in class org.apache.beam.runners.spark.metrics.MetricsAccumulator
-
Init metrics accumulator if it has not been initiated.
- initAccumulators(SparkPipelineOptions, JavaSparkContext) - Static method in class org.apache.beam.runners.spark.SparkRunner
-
Init Metrics/Aggregators accumulators.
- initialize(HttpRequest) - Method in class org.apache.beam.sdk.extensions.gcp.auth.NullCredentialInitializer
-
- initialize(AbstractGoogleClientRequest<?>) - Method in class org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions.GoogleApiTracer
-
- initialSystemTimeAt(Instant) - Method in class org.apache.beam.runners.spark.io.CreateStream
-
Set the initial synchronized processing time.
- inNamespace(String) - Static method in class org.apache.beam.sdk.metrics.MetricNameFilter
-
- inNamespace(Class<?>) - Static method in class org.apache.beam.sdk.metrics.MetricNameFilter
-
- innerJoin(PCollection<KV<K, V1>>, PCollection<KV<K, V2>>) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
-
Inner join of two collections of KV elements.
- inOnlyPane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
-
Creates a new
PAssert.SingletonAssert
like this one, but with the assertion restricted to
only run on the provided window.
- inOnTimePane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
-
Creates a new
PAssert.IterableAssert
like this one, but with the assertion restricted to only
run on the provided window.
- inOnTimePane(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-
- inOnTimePane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
-
Creates a new
PAssert.SingletonAssert
like this one, but with the assertion restricted to
only run on the provided window, running the checker only on the on-time pane for each key.
- inOrder(Trigger...) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterEach
-
Returns an AfterEach
Trigger
with the given subtriggers.
- inOrder(List<Trigger>) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterEach
-
Returns an AfterEach
Trigger
with the given subtriggers.
- inputOf(SerializableFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-
Returns a type descriptor for the input of the given
SerializableFunction
, subject to
Java type erasure: returns
null
if the type was erased.
- INPUTS - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- InsertRetryPolicy - Class in org.apache.beam.sdk.io.gcp.bigquery
-
A retry policy for streaming BigQuery inserts.
- InsertRetryPolicy() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy
-
- InsertRetryPolicy.Context - Class in org.apache.beam.sdk.io.gcp.bigquery
-
Contains information about a failed insert.
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamAggregationRule
-
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamFilterRule
-
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIntersectRule
-
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIOSinkRule
-
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIOSourceRule
-
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamJoinRule
-
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamMinusRule
-
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamProjectRule
-
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSortRule
-
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamUnionRule
-
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamValuesRule
-
- INSTANCE - Static variable in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
-
- INSTANCE - Static variable in class org.apache.beam.sdk.transforms.windowing.GlobalWindow
-
- INSTANCE - Static variable in class org.apache.beam.sdk.transforms.windowing.PaneInfo.PaneInfoCoder
-
- InstantCoder - Class in org.apache.beam.sdk.coders
-
A
Coder
for joda
Instant
that encodes it as a big endian
Long
shifted such that lexicographic ordering of the bytes corresponds to chronological order.
- InstantDeserializer - Class in org.apache.beam.sdk.io.kafka.serialization
-
- InstantDeserializer() - Constructor for class org.apache.beam.sdk.io.kafka.serialization.InstantDeserializer
-
- InstantSerializer - Class in org.apache.beam.sdk.io.kafka.serialization
-
- InstantSerializer() - Constructor for class org.apache.beam.sdk.io.kafka.serialization.InstantSerializer
-
- integers() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-
- integersGlobally() - Static method in class org.apache.beam.sdk.transforms.Max
-
Returns a PTransform
that takes an input PCollection<Integer>
and returns a
PCollection<Integer>
whose contents is the maximum of the input PCollection
's
elements, or Integer.MIN_VALUE
if there are no elements.
- integersGlobally() - Static method in class org.apache.beam.sdk.transforms.Min
-
Returns a PTransform
that takes an input PCollection<Integer>
and returns a
PCollection<Integer>
whose contents is a single value that is the minimum of the input
PCollection
's elements, or Integer.MAX_VALUE
if there are no elements.
- integersGlobally() - Static method in class org.apache.beam.sdk.transforms.Sum
-
Returns a PTransform
that takes an input
PCollection<Integer>
and returns a
PCollection<Integer>
whose contents is the sum of the
input PCollection
's elements, or
0
if there are no elements.
- integersPerKey() - Static method in class org.apache.beam.sdk.transforms.Max
-
Returns a PTransform
that takes an input PCollection<KV<K, Integer>>
and
returns a PCollection<KV<K, Integer>>
that contains an output element mapping each
distinct key in the input PCollection
to the maximum of the values associated with that
key in the input PCollection
.
- integersPerKey() - Static method in class org.apache.beam.sdk.transforms.Min
-
Returns a PTransform
that takes an input PCollection<KV<K, Integer>>
and
returns a PCollection<KV<K, Integer>>
that contains an output element mapping each
distinct key in the input PCollection
to the minimum of the values associated with that
key in the input PCollection
.
- integersPerKey() - Static method in class org.apache.beam.sdk.transforms.Sum
-
Returns a PTransform
that takes an input
PCollection<KV<K, Integer>>
and returns a
PCollection<KV<K, Integer>>
that contains an output
element mapping each distinct key in the input
PCollection
to the sum of the values associated with
that key in the input PCollection
.
- Internal - Annotation Type in org.apache.beam.sdk.annotations
-
Signifies that a publicly accessible API (public class, method or field) is intended for internal
use only and not for public consumption.
- interpolateKey(double) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
-
Returns a
ByteKey
key
such that
[startKey, key)
represents
approximately the specified fraction of the range
[startKey, endKey)
.
- intersects(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
-
Returns whether this window intersects the given window.
- IntervalWindow - Class in org.apache.beam.sdk.transforms.windowing
-
- IntervalWindow(Instant, Instant) - Constructor for class org.apache.beam.sdk.transforms.windowing.IntervalWindow
-
Creates a new IntervalWindow that represents the half-open time
interval [start, end).
- IntervalWindow(Instant, ReadableDuration) - Constructor for class org.apache.beam.sdk.transforms.windowing.IntervalWindow
-
- IntervalWindow.IntervalWindowCoder - Class in org.apache.beam.sdk.transforms.windowing
-
- IntervalWindowCoder() - Constructor for class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
-
- into(TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.transforms.FlatMapElements
-
- into(TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.transforms.MapElements
-
- into(WindowFn<? super T, ?>) - Static method in class org.apache.beam.sdk.transforms.windowing.Window
-
Creates a
Window
PTransform
that uses the given
WindowFn
to window the data.
- InvalidWindows<W extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
-
A
WindowFn
that represents an invalid pipeline state.
- InvalidWindows(String, WindowFn<?, W>) - Constructor for class org.apache.beam.sdk.transforms.windowing.InvalidWindows
-
- invokeAdvance(ReaderT) - Method in class org.apache.beam.runners.flink.metrics.ReaderInvocationUtil
-
- invokeStart(ReaderT) - Method in class org.apache.beam.runners.flink.metrics.ReaderInvocationUtil
-
- inWindow(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
-
Creates a new
PAssert.IterableAssert
like this one, but with the assertion restricted to only
run on the provided window.
- inWindow(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-
- IS_MERGING_WINDOW_FN - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- IS_PAIR_LIKE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- IS_STREAM_LIKE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- IS_WRAPPER - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- isAccessible() - Method in interface org.apache.beam.sdk.options.ValueProvider
-
- isAccessible() - Method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
-
- isAccessible() - Method in class org.apache.beam.sdk.options.ValueProvider.RuntimeValueProvider
-
- isAccessible() - Method in class org.apache.beam.sdk.options.ValueProvider.StaticValueProvider
-
- isAllowedLatenessSpecified() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
- isArray() - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
Returns true if this type is known to be an array type.
- isAtSplitPoint() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
-
Returns true if the reader is at a split point.
- isAtSplitPoint() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-
Returns true only for the first record; compressed sources cannot be split.
- isAtSplitPoint() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
- isBlockOnRun() - Method in interface org.apache.beam.runners.direct.DirectOptions
-
- isBounded() - Method in class org.apache.beam.sdk.values.PCollection
-
- isBoundedCollection(Collection<PValue>) - Method in class org.apache.beam.runners.spark.SparkRunner.Evaluator
-
- isCompatible(Trigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
-
- isCompatible(Trigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
-
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-
- isCompatible(Trigger) - Method in class org.apache.beam.sdk.transforms.windowing.DefaultTrigger
-
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
-
InvalidWindows
objects with the same originalWindowFn
are compatible.
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-
- isCompatible(Trigger) - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
-
For internal use only; no backwards-compatibility guarantees.
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
-
- isCompressed(String) - Static method in enum org.apache.beam.sdk.io.CompressedSource.CompressionMode
-
Deprecated.
Returns whether the file's extension matches of one of the known compression formats.
- isCompressed(String) - Method in enum org.apache.beam.sdk.io.Compression
-
- isDirectory() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
-
- isDirectory() - Method in interface org.apache.beam.sdk.io.fs.ResourceId
-
Returns
true
if this
ResourceId
represents a directory, false otherwise.
- isDisjoint(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
-
Returns whether this window is disjoint from the given window.
- isDone() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
- isDone() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-
- isDone() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-
- isEmbeddedExecution() - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
-
- isEmbeddedExecutionDebugMode() - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
-
- isEmpty() - Method in class org.apache.beam.runners.spark.util.SparkSideInputReader
-
- isEmpty() - Method in class org.apache.beam.sdk.io.range.ByteKey
-
Returns
true
if the
byte[]
backing this
ByteKey
is of length 0.
- isEmpty() - Method in interface org.apache.beam.sdk.state.GroupingState
-
- isEmpty() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
-
- isEmpty() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
-
- isEnforceEncodability() - Method in interface org.apache.beam.runners.direct.DirectOptions
-
- isEnforceImmutability() - Method in interface org.apache.beam.runners.direct.DirectOptions
-
- isEqualTo(T) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
-
Asserts that the value in question is equal to the provided value, according to
Object.equals(java.lang.Object)
.
- isExternalizedCheckpointsEnabled() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
Enables or disables externalized checkpoints.
- isFinished() - Method in class org.apache.beam.runners.apex.ApexYarnLauncher.ProcessWatcher
-
- isFirst() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
-
Return true if this is the first pane produced for the associated window.
- isForceStreaming() - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
-
- isForceWatermarkSync() - Method in class org.apache.beam.runners.spark.io.CreateStream
-
- isInsertDefault() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
-
Returns whether or not this transformation applies a default value.
- isLast() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
-
Return true if this is the last pane that will be produced in the associated window.
- isMetricsSupported() - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
-
Indicates whether metrics reporting is supported.
- isModeSpecified() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
- isMurmur3Partitioner(Cluster) - Static method in class org.apache.beam.sdk.io.cassandra.CassandraServiceImpl
-
Check if the current partitioner is the Murmur3 (default in Cassandra version newer than 2).
- isNonMerging() - Method in class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator.GearpumpWindowFn
-
- isNonMerging() - Method in class org.apache.beam.sdk.transforms.windowing.NonMergingWindowFn
-
- isNonMerging() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
-
Returns true if this WindowFn
never needs to merge any windows.
- isOperandNumeric(SqlTypeName) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math.BeamSqlMathBinaryExpression
-
The method to check whether operands are numeric or not.
- isParDoFusionEnabled() - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
-
- isProtoTranslation() - Method in interface org.apache.beam.runners.direct.DirectOptions
-
- isReadSeekEfficient() - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata
-
- isRegisterByteSizeObserverCheap(RandomAccessData) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
-
- isRegisterByteSizeObserverCheap(BigDecimal) - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
-
- isRegisterByteSizeObserverCheap(Integer) - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
-
- isRegisterByteSizeObserverCheap(Long) - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
-
- isRegisterByteSizeObserverCheap(BigInteger) - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
-
- isRegisterByteSizeObserverCheap(Boolean) - Method in class org.apache.beam.sdk.coders.BooleanCoder
-
- isRegisterByteSizeObserverCheap(byte[]) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
-
- isRegisterByteSizeObserverCheap(Byte) - Method in class org.apache.beam.sdk.coders.ByteCoder
-
- isRegisterByteSizeObserverCheap(T) - Method in class org.apache.beam.sdk.coders.Coder
-
- isRegisterByteSizeObserverCheap(Double) - Method in class org.apache.beam.sdk.coders.DoubleCoder
-
- isRegisterByteSizeObserverCheap(ReadableDuration) - Method in class org.apache.beam.sdk.coders.DurationCoder
-
- isRegisterByteSizeObserverCheap(Instant) - Method in class org.apache.beam.sdk.coders.InstantCoder
-
- isRegisterByteSizeObserverCheap(IterableT) - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
-
- isRegisterByteSizeObserverCheap(KV<K, V>) - Method in class org.apache.beam.sdk.coders.KvCoder
-
Returns whether both keyCoder and valueCoder are considered not expensive.
- isRegisterByteSizeObserverCheap(T) - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
-
LengthPrefixCoder
is cheap if valueCoder
is cheap.
- isRegisterByteSizeObserverCheap(T) - Method in class org.apache.beam.sdk.coders.NullableCoder
-
NullableCoder
is cheap if valueCoder
is cheap.
- isRegisterByteSizeObserverCheap(Integer) - Method in class org.apache.beam.sdk.coders.VarIntCoder
-
- isRegisterByteSizeObserverCheap(Long) - Method in class org.apache.beam.sdk.coders.VarLongCoder
-
- isRegisterByteSizeObserverCheap(Void) - Method in class org.apache.beam.sdk.coders.VoidCoder
-
- isRegisterByteSizeObserverCheap(ByteString) - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
-
- isRegisterByteSizeObserverCheap(KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
-
- isRegisterByteSizeObserverCheap(RawUnionValue) - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
-
Since this coder uses elementCoders.get(index) and coders that are known to run in constant
time, we defer the return value to that coder.
- isRunnerDeterminedSharding() - Method in interface org.apache.beam.runners.direct.DirectTestOptions
-
- isSplittable() - Method in class org.apache.beam.sdk.io.CompressedSource
-
Determines whether a single file represented by this source is splittable.
- isSplittable() - Method in class org.apache.beam.sdk.io.FileBasedSource
-
Determines whether a file represented by this source is can be split into bundles.
- isStarted() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
- isStarted() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-
- isStreaming() - Method in interface org.apache.beam.sdk.options.StreamingOptions
-
Set to true if running a streaming pipeline.
- isSubtypeOf(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
Return true if this type is a subtype of the given type.
- isSuccess() - Method in class org.apache.beam.sdk.testing.SuccessOrFailure
-
- isSupertypeOf(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
Returns true if this type is assignable from the given type.
- isTerminal() - Method in enum org.apache.beam.sdk.PipelineResult.State
-
- isTimestampCombinerSpecified() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
- isTriggerSpecified() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
- isTupleTracingEnabled() - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
-
- isUnknown() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
-
Return true if there is no timing information for the current
PaneInfo
.
- isUpdate() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-
Whether to update the currently running pipeline with the same name as this one.
- isWholeStream - Variable in class org.apache.beam.sdk.coders.Coder.Context
-
Deprecated.
Whether the encoded or decoded value fills the remainder of the
output or input (resp.) record/stream contents.
- isWindowedWrites() - Method in class org.apache.beam.sdk.io.WriteFiles
-
Returns whether or not to perform windowed writes.
- item(String, String) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
-
Create a display item for the specified key and string value.
- item(String, ValueProvider<?>) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
-
- item(String, Integer) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
-
Create a display item for the specified key and integer value.
- item(String, Long) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
-
Create a display item for the specified key and integer value.
- item(String, Float) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
-
Create a display item for the specified key and floating point value.
- item(String, Double) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
-
Create a display item for the specified key and floating point value.
- item(String, Boolean) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
-
Create a display item for the specified key and boolean value.
- item(String, Instant) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
-
Create a display item for the specified key and timestamp value.
- item(String, Duration) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
-
Create a display item for the specified key and duration value.
- item(String, Class<T>) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
-
Create a display item for the specified key and class value.
- item(String, DisplayData.Type, T) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
-
Create a display item for the specified key, type, and value.
- Item() - Constructor for class org.apache.beam.sdk.transforms.display.DisplayData.Item
-
- items() - Method in class org.apache.beam.sdk.transforms.display.DisplayData
-
- ItemSpec() - Constructor for class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
-
- iterable() - Static method in class org.apache.beam.sdk.transforms.Materializations
-
For internal use only; no backwards-compatibility guarantees.
- ITERABLE_MATERIALIZATION_URN - Static variable in class org.apache.beam.sdk.transforms.Materializations
-
The URN for a
Materialization
where the primitive view type is an iterable of fully
specified windowed values.
- IterableCoder<T> - Class in org.apache.beam.sdk.coders
-
- IterableCoder(Coder<T>) - Constructor for class org.apache.beam.sdk.coders.IterableCoder
-
- IterableLikeCoder<T,IterableT extends java.lang.Iterable<T>> - Class in org.apache.beam.sdk.coders
-
An abstract base class with functionality for assembling a
Coder
for a class that implements
Iterable
.
- IterableLikeCoder(Coder<T>, String) - Constructor for class org.apache.beam.sdk.coders.IterableLikeCoder
-
- iterables() - Static method in class org.apache.beam.sdk.transforms.Flatten
-
Returns a PTransform
that takes a PCollection<Iterable<T>>
and returns a PCollection<T>
containing all the elements from
all the Iterable
s.
- iterables() - Static method in class org.apache.beam.sdk.transforms.ToString
-
Transforms each item in the iterable of the input
PCollection
to a
String
using the
Object.toString()
method followed by a "," until
the last element in the iterable.
- iterables(String) - Static method in class org.apache.beam.sdk.transforms.ToString
-
Transforms each item in the iterable of the input
PCollection
to a
String
using
the
Object.toString()
method followed by the specified delimiter until the last element
in the iterable.
- iterables(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-
- iterableView(PCollection<T>, WindowingStrategy<?, W>, Coder<T>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
-
Returns a
PCollectionView<Iterable<T>>
capable of processing elements encoded using the
provided
Coder
and windowed using the provided
WindowingStrategy
.
- IterableViewFn() - Constructor for class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn
-
Deprecated.
- iterator() - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
-
- OBJECT_TYPE_NAME - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- of(PCollectionView<ViewT>) - Static method in class org.apache.beam.runners.apex.ApexRunner.CreateApexPCollectionView
-
- of(PCollectionView<ViewT>) - Static method in class org.apache.beam.runners.dataflow.CreateDataflowView
-
- of() - Static method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
-
- of(Coder<T>, Duration, boolean) - Static method in class org.apache.beam.runners.spark.io.CreateStream
-
Creates a new Spark based stream intended for test purposes.
- of(Coder<T>, Duration) - Static method in class org.apache.beam.runners.spark.io.CreateStream
-
Creates a new Spark based stream without forced watermark sync, intended for test purposes.
- of(NamedAggregators) - Static method in class org.apache.beam.runners.spark.metrics.AggregatorMetric
-
- of(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.coders.AvroCoder
-
Returns an AvroCoder
instance for the provided element type.
- of(Class<T>) - Static method in class org.apache.beam.sdk.coders.AvroCoder
-
Returns an AvroCoder
instance for the provided element class.
- of(Schema) - Static method in class org.apache.beam.sdk.coders.AvroCoder
-
Returns an AvroCoder
instance for the Avro schema.
- of(Class<T>, Schema) - Static method in class org.apache.beam.sdk.coders.AvroCoder
-
Returns an AvroCoder
instance for the provided element type
using the provided Avro schema.
- of(BeamRecordType, List<Coder>) - Static method in class org.apache.beam.sdk.coders.BeamRecordCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.BigDecimalCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.BigEndianLongCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.BigIntegerCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.BitSetCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.BooleanCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.ByteArrayCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.ByteCoder
-
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.CollectionCoder
-
- of(Coder<IntermediateT>, DelegateCoder.CodingFunction<T, IntermediateT>, DelegateCoder.CodingFunction<IntermediateT, T>) - Static method in class org.apache.beam.sdk.coders.DelegateCoder
-
- of(Coder<IntermediateT>, DelegateCoder.CodingFunction<T, IntermediateT>, DelegateCoder.CodingFunction<IntermediateT, T>, TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.coders.DelegateCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.DoubleCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.DurationCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.InstantCoder
-
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.IterableCoder
-
- of(Coder<K>, Coder<V>) - Static method in class org.apache.beam.sdk.coders.KvCoder
-
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.LengthPrefixCoder
-
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.ListCoder
-
- of(Coder<K>, Coder<V>) - Static method in class org.apache.beam.sdk.coders.MapCoder
-
Produces a MapCoder with the given keyCoder and valueCoder.
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.NullableCoder
-
- of(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.coders.SerializableCoder
-
- of(Class<T>) - Static method in class org.apache.beam.sdk.coders.SerializableCoder
-
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.SetCoder
-
Produces a
SetCoder
with the given
elementCoder
.
- of(Coder<KeyT>) - Static method in class org.apache.beam.sdk.coders.ShardedKeyCoder
-
- of(Class<T>) - Static method in class org.apache.beam.sdk.coders.StringDelegateCoder
-
- of(Class<T>, TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.coders.StringDelegateCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.StringUtf8Coder
-
- of() - Static method in class org.apache.beam.sdk.coders.TextualIntegerCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.VarIntCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.VarLongCoder
-
- of() - Static method in class org.apache.beam.sdk.coders.VoidCoder
-
- of(Class<? extends OutputT>) - Static method in class org.apache.beam.sdk.extensions.jackson.AsJsons
-
- of(Class<? extends OutputT>) - Static method in class org.apache.beam.sdk.extensions.jackson.ParseJsons
-
- of() - Static method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
-
- of(Class<T>) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
-
Returns a
ProtoCoder
for the given Protocol Buffers
Message
.
- of(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
-
- of() - Static method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.BooleanCoder
-
- of() - Static method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.DateCoder
-
- of() - Static method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.DoubleCoder
-
- of() - Static method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.FloatCoder
-
- of() - Static method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.ShortCoder
-
- of() - Static method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.TimeCoder
-
- of(SqlTypeName, T) - Static method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlPrimitive
-
A builder function to create from Type and value directly.
- of() - Static method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.ParamsCoder
-
- of(Coder<BoundedWindow>, Coder<DestinationT>) - Static method in class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
-
- of() - Static method in class org.apache.beam.sdk.io.fs.MetadataCoder
-
- of() - Static method in class org.apache.beam.sdk.io.fs.ResourceIdCoder
-
- of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoder
-
- of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV2
-
- of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
-
- of() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessagePayloadOnlyCoder
-
- of(TypeDescriptor<PubsubMessage>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
-
- of() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
-
- of(Class<T>) - Static method in class org.apache.beam.sdk.io.hadoop.WritableCoder
-
Returns a WritableCoder
instance for the provided element class.
- of(Coder<K>, Coder<V>) - Static method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
-
- of(int...) - Static method in class org.apache.beam.sdk.io.range.ByteKey
-
Creates a new
ByteKey
backed by a copy of the specified
int[]
.
- of(ByteKey, ByteKey) - Static method in class org.apache.beam.sdk.io.range.ByteKeyRange
-
Creates a new
ByteKeyRange
with the given start and end keys.
- of(ByteKeyRange) - Static method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-
- of() - Static method in class org.apache.beam.sdk.io.ReadableFileCoder
-
- of(Class<T>) - Static method in class org.apache.beam.sdk.io.xml.JAXBCoder
-
Create a coder for a given type of JAXB annotated objects.
- of(ValueProvider<X>, SerializableFunction<X, T>) - Static method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
-
- of(T) - Static method in class org.apache.beam.sdk.options.ValueProvider.StaticValueProvider
-
- of(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
-
Returns a CombineFn
that uses the given
SerializableFunction
to combine values.
- of(SerializableFunction<Iterable<V>, V>, int) - Static method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
-
Returns a CombineFn
that uses the given
SerializableFunction
to combine values,
attempting to buffer at least bufferSize
values between invocations.
- of(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine.SimpleCombineFn
-
Deprecated.
Returns a CombineFn
that uses the given
SerializableFunction
to combine values.
- of(Iterable<T>) - Static method in class org.apache.beam.sdk.transforms.Create
-
Returns a new
Create.Values
transform that produces a
PCollection
containing elements of the provided
Iterable
.
- of(T, T...) - Static method in class org.apache.beam.sdk.transforms.Create
-
Returns a new
Create.Values
transform that produces a
PCollection
containing the specified elements.
- of(Map<K, V>) - Static method in class org.apache.beam.sdk.transforms.Create
-
Returns a new
Create.Values
transform that produces a
PCollection
of
KV
s corresponding to the keys and
values of the specified
Map
.
- of(DisplayData.Path, Class<?>, String) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
-
- of(DoFn<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.DoFnTester
-
Returns a
DoFnTester
supporting unit-testing of the given
DoFn
.
- of(CoGbkResultSchema, UnionCoder) - Static method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
-
- of(TupleTag<V>, List<V>) - Static method in class org.apache.beam.sdk.transforms.join.CoGbkResult
-
Returns a new CoGbkResult that contains just the given tag and given data.
- of(List<TupleTag<?>>) - Static method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
-
- of(TupleTag<InputT>, PCollection<KV<K, InputT>>) - Static method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
-
Returns a new KeyedPCollectionTuple<K>
with the given tag and initial
PCollection.
- of(List<Coder<?>>) - Static method in class org.apache.beam.sdk.transforms.join.UnionCoder
-
Builds a union coder with the given list of element coders.
- of(T, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Max
-
- of(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Max
-
- of() - Static method in class org.apache.beam.sdk.transforms.Mean
-
A Combine.CombineFn
that computes the arithmetic mean
(a.k.a.
- of(T, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Min
-
- of(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Min
-
- of(DoFn<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.ParDo
-
- of(int, Partition.PartitionFn<? super T>) - Static method in class org.apache.beam.sdk.transforms.Partition
-
Returns a new Partition
PTransform
that divides
its input PCollection
into the given number of partitions,
using the given partitioning function.
- of() - Static method in class org.apache.beam.sdk.transforms.Reshuffle
-
Deprecated.
- of(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Top
-
Returns a PTransform
that takes an input
PCollection<T>
and returns a PCollection<List<T>>
with a
single element containing the largest count
elements of the input
PCollection<T>
, in decreasing order, sorted using the
given Comparator<T>
.
- of(PCollectionView<ViewT>) - Static method in class org.apache.beam.sdk.transforms.View.CreatePCollectionView
-
- of(Trigger.OnceTrigger...) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterAll
-
Returns an AfterAll
Trigger
with the given subtriggers.
- of(List<Trigger>) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterAll
-
Returns an AfterAll
Trigger
with the given subtriggers.
- of(Trigger.OnceTrigger...) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterFirst
-
Returns an AfterFirst
Trigger
with the given subtriggers.
- of(List<Trigger>) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterFirst
-
Returns an AfterFirst
Trigger
with the given subtriggers.
- of() - Static method in class org.apache.beam.sdk.transforms.windowing.DefaultTrigger
-
Returns the default trigger.
- of(Duration) - Static method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-
Partitions the timestamp space into half-open intervals of the form
[N * size, (N + 1) * size), where 0 is the epoch.
- of() - Static method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
-
- of() - Static method in class org.apache.beam.sdk.transforms.windowing.PaneInfo.PaneInfoCoder
-
- of(Duration) - Static method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-
Assigns timestamps into half-open intervals of the form
[N * period, N * period + size), where 0 is the epoch.
- of(SerializableFunction<V, K>) - Static method in class org.apache.beam.sdk.transforms.WithKeys
-
Returns a PTransform
that takes a PCollection<V>
and returns a PCollection<KV<K, V>>
, where each of the
values in the input PCollection
has been paired with a
key computed from the value by invoking the given
SerializableFunction
.
- of(K) - Static method in class org.apache.beam.sdk.transforms.WithKeys
-
Returns a PTransform
that takes a PCollection<V>
and returns a PCollection<KV<K, V>>
, where each of the
values in the input PCollection
has been paired with the
given key.
- of(SerializableFunction<T, Instant>) - Static method in class org.apache.beam.sdk.transforms.WithTimestamps
-
- of(K, V) - Static method in class org.apache.beam.sdk.values.KV
-
Returns a
KV
with the given key and value.
- of(PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionList
-
- of(Iterable<PCollection<T>>) - Static method in class org.apache.beam.sdk.values.PCollectionList
-
- of(TupleTag<T>, PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
-
- of(K, int) - Static method in class org.apache.beam.sdk.values.ShardedKey
-
- of(TupleTag<?>, PValue) - Static method in class org.apache.beam.sdk.values.TaggedPValue
-
- of(V, Instant) - Static method in class org.apache.beam.sdk.values.TimestampedValue
-
Returns a new TimestampedValue
with the given value and timestamp.
- of(Coder<T>) - Static method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
-
- of(TupleTag<?>) - Static method in class org.apache.beam.sdk.values.TupleTagList
-
- of(List<TupleTag<?>>) - Static method in class org.apache.beam.sdk.values.TupleTagList
-
- of(Class<T>) - Static method in class org.apache.beam.sdk.values.TypeDescriptor
-
- of(Type) - Static method in class org.apache.beam.sdk.values.TypeDescriptor
-
- of(Coder<T>, Coder<? extends BoundedWindow>) - Static method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
-
- of(T, Instant, BoundedWindow, PaneInfo) - Static method in class org.apache.beam.sdk.values.ValueInSingleWindow
-
- of(Coder<ValueT>) - Static method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
-
- of(WindowFn<T, W>) - Static method in class org.apache.beam.sdk.values.WindowingStrategy
-
- ofDoubles() - Static method in class org.apache.beam.sdk.transforms.Max
-
- ofDoubles() - Static method in class org.apache.beam.sdk.transforms.Min
-
- ofDoubles() - Static method in class org.apache.beam.sdk.transforms.Sum
-
- ofExpandedValue(PValue) - Static method in class org.apache.beam.sdk.values.TaggedPValue
-
- offerCoders(Coder[]) - Method in interface org.apache.beam.sdk.state.StateSpec
-
For internal use only; no backwards-compatibility guarantees.
- ofFirstElement() - Static method in class org.apache.beam.sdk.transforms.windowing.AfterSynchronizedProcessingTime
-
- offset(Duration) - Method in interface org.apache.beam.sdk.state.Timer
-
- OFFSET_INFINITY - Static variable in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-
Offset corresponding to infinity.
- OffsetBasedReader(OffsetBasedSource<T>) - Constructor for class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
- OffsetBasedSource<T> - Class in org.apache.beam.sdk.io
-
A
BoundedSource
that uses offsets to define starting and ending positions.
- OffsetBasedSource(long, long, long) - Constructor for class org.apache.beam.sdk.io.OffsetBasedSource
-
- OffsetBasedSource.OffsetBasedReader<T> - Class in org.apache.beam.sdk.io
-
- OffsetRange - Class in org.apache.beam.sdk.io.range
-
A restriction represented by a range of integers [from, to).
- OffsetRange(long, long) - Constructor for class org.apache.beam.sdk.io.range.OffsetRange
-
- OffsetRangeTracker - Class in org.apache.beam.sdk.io.range
-
- OffsetRangeTracker(long, long) - Constructor for class org.apache.beam.sdk.io.range.OffsetRangeTracker
-
Creates an OffsetRangeTracker
for the specified range.
- OffsetRangeTracker - Class in org.apache.beam.sdk.transforms.splittabledofn
-
- OffsetRangeTracker(OffsetRange) - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
-
- ofIntegers() - Static method in class org.apache.beam.sdk.transforms.Max
-
- ofIntegers() - Static method in class org.apache.beam.sdk.transforms.Min
-
- ofIntegers() - Static method in class org.apache.beam.sdk.transforms.Sum
-
- ofLongs() - Static method in class org.apache.beam.sdk.transforms.Max
-
- ofLongs() - Static method in class org.apache.beam.sdk.transforms.Min
-
- ofLongs() - Static method in class org.apache.beam.sdk.transforms.Sum
-
- ofPrimitiveOutputsInternal(Pipeline, TupleTagList, Map<TupleTag<?>, Coder<?>>, WindowingStrategy<?, ?>, PCollection.IsBounded) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
-
For internal use only; no backwards-compatibility guarantees.
- ofProvider(ValueProvider<T>, Coder<T>) - Static method in class org.apache.beam.sdk.transforms.Create
-
- ofSize(long) - Static method in class org.apache.beam.sdk.transforms.GroupIntoBatches
-
- ON_TIME_AND_ONLY_FIRING - Static variable in class org.apache.beam.sdk.transforms.windowing.PaneInfo
-
PaneInfo
to use when there will be exactly one firing and it is on time.
- onBatchCompleted(JavaStreamingListenerBatchCompleted) - Method in class org.apache.beam.runners.spark.aggregators.AggregatorsAccumulator.AccumulatorCheckpointingSparkListener
-
- onBatchCompleted(JavaStreamingListenerBatchCompleted) - Method in class org.apache.beam.runners.spark.metrics.MetricsAccumulator.AccumulatorCheckpointingSparkListener
-
- onBatchCompleted(JavaStreamingListenerBatchCompleted) - Method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.WatermarkAdvancingStreamingListener
-
- OnceTrigger(List<Trigger>) - Constructor for class org.apache.beam.sdk.transforms.windowing.Trigger.OnceTrigger
-
- onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamAggregationRule
-
- onSeenNewOutput(Instant, StateT) - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
-
- onTimer(String, BoundedWindow, Instant, TimeDomain) - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
-
- OnTimerContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.OnTimerContext
-
- op(int) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlExpression
-
- open(TaskContext, Instant) - Method in class org.apache.beam.runners.gearpump.translators.io.GearpumpSource
-
- open() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
-
- open(ResourceIdT) - Method in class org.apache.beam.sdk.io.FileSystem
-
Returns a read channel for the given ResourceIdT
.
- open(ResourceId) - Static method in class org.apache.beam.sdk.io.FileSystems
-
- openSeekable() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
-
- openUnwindowed(String, int, DestinationT) - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
-
- openWindowed(String, BoundedWindow, PaneInfo, int, DestinationT) - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
-
Performs bundle initialization.
- operands - Variable in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlExpression
-
- Options() - Constructor for class org.apache.beam.runners.apex.ApexRunnerRegistrar.Options
-
- Options() - Constructor for class org.apache.beam.runners.dataflow.DataflowPipelineRegistrar.Options
-
- Options() - Constructor for class org.apache.beam.runners.direct.DirectRegistrar.Options
-
- Options() - Constructor for class org.apache.beam.runners.flink.FlinkRunnerRegistrar.Options
-
- Options() - Constructor for class org.apache.beam.runners.gearpump.GearpumpRunnerRegistrar.Options
-
- Options() - Constructor for class org.apache.beam.runners.spark.SparkRunnerRegistrar.Options
-
- options() - Static method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter
-
- opType(int) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlExpression
-
- opValueEvaluated(int, BeamRecord, BoundedWindow) - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlExpression
-
- OrderByKey() - Constructor for class org.apache.beam.sdk.values.KV.OrderByKey
-
- OrderByValue() - Constructor for class org.apache.beam.sdk.values.KV.OrderByValue
-
- orFinally(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
-
Specify an ending condition for this trigger.
- OrFinallyTrigger - Class in org.apache.beam.sdk.transforms.windowing
-
A
Trigger
that executes according to its main trigger until its "finally" trigger fires.
- org.apache.beam.artifact.local - package org.apache.beam.artifact.local
-
Provides local implementations of the Artifact API services.
- org.apache.beam.runners.apex - package org.apache.beam.runners.apex
-
Implementation of the Beam runner for Apache Apex.
- org.apache.beam.runners.dataflow - package org.apache.beam.runners.dataflow
-
Provides a Beam runner that executes pipelines on the Google Cloud Dataflow service.
- org.apache.beam.runners.dataflow.options - package org.apache.beam.runners.dataflow.options
-
- org.apache.beam.runners.dataflow.util - package org.apache.beam.runners.dataflow.util
-
Provides miscellaneous internal utilities used by the Google Cloud Dataflow runner.
- org.apache.beam.runners.direct - package org.apache.beam.runners.direct
-
- org.apache.beam.runners.flink - package org.apache.beam.runners.flink
-
Internal implementation of the Beam runner for Apache Flink.
- org.apache.beam.runners.flink.metrics - package org.apache.beam.runners.flink.metrics
-
Internal metrics implementation of the Beam runner for Apache Flink.
- org.apache.beam.runners.gearpump - package org.apache.beam.runners.gearpump
-
Internal implementation of the Beam runner for Apache Gearpump.
- org.apache.beam.runners.gearpump.translators - package org.apache.beam.runners.gearpump.translators
-
Gearpump specific translators.
- org.apache.beam.runners.gearpump.translators.functions - package org.apache.beam.runners.gearpump.translators.functions
-
Gearpump specific wrappers for Beam DoFn.
- org.apache.beam.runners.gearpump.translators.io - package org.apache.beam.runners.gearpump.translators.io
-
Gearpump specific wrappers for Beam I/O.
- org.apache.beam.runners.gearpump.translators.utils - package org.apache.beam.runners.gearpump.translators.utils
-
Utilities for translators.
- org.apache.beam.runners.reference.job - package org.apache.beam.runners.reference.job
-
An execution engine for Beam Pipelines that uses the Java Runner harness and the Fn API to
execute.
- org.apache.beam.runners.spark - package org.apache.beam.runners.spark
-
Internal implementation of the Beam runner for Apache Spark.
- org.apache.beam.runners.spark.aggregators - package org.apache.beam.runners.spark.aggregators
-
Provides internal utilities for implementing Beam aggregators using Spark accumulators.
- org.apache.beam.runners.spark.aggregators.metrics - package org.apache.beam.runners.spark.aggregators.metrics
-
Defines classes for integrating with Spark's metrics mechanism (Sinks, Sources, etc.).
- org.apache.beam.runners.spark.coders - package org.apache.beam.runners.spark.coders
-
Beam coders and coder-related utilities for running on Apache Spark.
- org.apache.beam.runners.spark.io - package org.apache.beam.runners.spark.io
-
Spark-specific transforms for I/O.
- org.apache.beam.runners.spark.metrics - package org.apache.beam.runners.spark.metrics
-
Provides internal utilities for implementing Beam metrics using Spark accumulators.
- org.apache.beam.runners.spark.metrics.sink - package org.apache.beam.runners.spark.metrics.sink
-
Spark sinks that supports beam metrics and aggregators.
- org.apache.beam.runners.spark.stateful - package org.apache.beam.runners.spark.stateful
-
Spark-specific stateful operators.
- org.apache.beam.runners.spark.util - package org.apache.beam.runners.spark.util
-
Internal utilities to translate Beam pipelines to Spark.
- org.apache.beam.sdk - package org.apache.beam.sdk
-
Provides a simple, powerful model for building both batch and
streaming parallel data processing
Pipeline
s.
- org.apache.beam.sdk.annotations - package org.apache.beam.sdk.annotations
-
Defines annotations used across the SDK.
- org.apache.beam.sdk.coders - package org.apache.beam.sdk.coders
-
Defines
Coders
to specify how data is encoded to and decoded from byte strings.
- org.apache.beam.sdk.extensions.gcp.auth - package org.apache.beam.sdk.extensions.gcp.auth
-
Defines classes related to interacting with Credentials
for
pipeline creation and execution containing Google Cloud Platform components.
- org.apache.beam.sdk.extensions.gcp.options - package org.apache.beam.sdk.extensions.gcp.options
-
Defines
PipelineOptions
for
configuring pipeline execution for Google Cloud Platform components.
- org.apache.beam.sdk.extensions.gcp.storage - package org.apache.beam.sdk.extensions.gcp.storage
-
Defines IO connectors for Google Cloud Storage.
- org.apache.beam.sdk.extensions.jackson - package org.apache.beam.sdk.extensions.jackson
-
Utilities for parsing and creating JSON serialized objects.
- org.apache.beam.sdk.extensions.joinlibrary - package org.apache.beam.sdk.extensions.joinlibrary
-
Utilities for performing SQL-style joins of keyed
PCollections
.
- org.apache.beam.sdk.extensions.protobuf - package org.apache.beam.sdk.extensions.protobuf
-
Defines a
Coder
for Protocol Buffers messages,
ProtoCoder
.
- org.apache.beam.sdk.extensions.sorter - package org.apache.beam.sdk.extensions.sorter
-
Utility for performing local sort of potentially large sets of values.
- org.apache.beam.sdk.extensions.sql - package org.apache.beam.sdk.extensions.sql
-
BeamSQL provides a new interface to run a SQL statement with Beam.
- org.apache.beam.sdk.extensions.sql.example - package org.apache.beam.sdk.extensions.sql.example
-
examples on how to use BeamSQL.
- org.apache.beam.sdk.extensions.sql.impl - package org.apache.beam.sdk.extensions.sql.impl
-
Implementation classes of BeamSql.
- org.apache.beam.sdk.extensions.sql.impl.interpreter - package org.apache.beam.sdk.extensions.sql.impl.interpreter
-
interpreter generate runnable 'code' to execute SQL relational expressions.
- org.apache.beam.sdk.extensions.sql.impl.interpreter.operator - package org.apache.beam.sdk.extensions.sql.impl.interpreter.operator
-
Implementation for operators in org.apache.calcite.sql.fun.SqlStdOperatorTable
.
- org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic - package org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.arithmetic
-
Arithmetic operators.
- org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison - package org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.comparison
-
Comparison operators.
- org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date - package org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.date
-
date functions.
- org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.logical - package org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.logical
-
Logical operators.
- org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math - package org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.math
-
MATH functions/operators.
- org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string - package org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.string
-
String operators.
- org.apache.beam.sdk.extensions.sql.impl.planner - package org.apache.beam.sdk.extensions.sql.impl.planner
-
- org.apache.beam.sdk.extensions.sql.impl.rel - package org.apache.beam.sdk.extensions.sql.impl.rel
-
BeamSQL specified nodes, to replace org.apache.calcite.rel.RelNode
.
- org.apache.beam.sdk.extensions.sql.impl.rule - package org.apache.beam.sdk.extensions.sql.impl.rule
-
org.apache.calcite.plan.RelOptRule
to generate
BeamRelNode
.
- org.apache.beam.sdk.extensions.sql.impl.schema - package org.apache.beam.sdk.extensions.sql.impl.schema
-
define table schema, to map with Beam IO components.
- org.apache.beam.sdk.extensions.sql.impl.schema.kafka - package org.apache.beam.sdk.extensions.sql.impl.schema.kafka
-
table schema for KafkaIO.
- org.apache.beam.sdk.extensions.sql.impl.schema.text - package org.apache.beam.sdk.extensions.sql.impl.schema.text
-
Table schema for text files.
- org.apache.beam.sdk.extensions.sql.impl.transform - package org.apache.beam.sdk.extensions.sql.impl.transform
-
- org.apache.beam.sdk.extensions.sql.impl.utils - package org.apache.beam.sdk.extensions.sql.impl.utils
-
Utility classes.
- org.apache.beam.sdk.io - package org.apache.beam.sdk.io
-
Defines transforms for reading and writing common storage formats, including
AvroIO
, and
TextIO
.
- org.apache.beam.sdk.io.amqp - package org.apache.beam.sdk.io.amqp
-
Transforms for reading and writing using AMQP 1.0 protocol.
- org.apache.beam.sdk.io.cassandra - package org.apache.beam.sdk.io.cassandra
-
Transforms for reading and writing from Apache Cassandra database.
- org.apache.beam.sdk.io.elasticsearch - package org.apache.beam.sdk.io.elasticsearch
-
Transforms for reading and writing from Elasticsearch.
- org.apache.beam.sdk.io.fs - package org.apache.beam.sdk.io.fs
-
Apache Beam FileSystem interfaces and their default implementations.
- org.apache.beam.sdk.io.gcp.bigquery - package org.apache.beam.sdk.io.gcp.bigquery
-
Defines transforms for reading and writing from Google BigQuery.
- org.apache.beam.sdk.io.gcp.bigtable - package org.apache.beam.sdk.io.gcp.bigtable
-
Defines transforms for reading and writing from Google Cloud Bigtable.
- org.apache.beam.sdk.io.gcp.common - package org.apache.beam.sdk.io.gcp.common
-
Defines common Google Cloud Platform IO support classes.
- org.apache.beam.sdk.io.gcp.datastore - package org.apache.beam.sdk.io.gcp.datastore
-
Provides an API for reading from and writing to
Google Cloud Datastore over different
versions of the Cloud Datastore Client libraries.
- org.apache.beam.sdk.io.gcp.pubsub - package org.apache.beam.sdk.io.gcp.pubsub
-
- org.apache.beam.sdk.io.gcp.spanner - package org.apache.beam.sdk.io.gcp.spanner
-
- org.apache.beam.sdk.io.gcp.testing - package org.apache.beam.sdk.io.gcp.testing
-
Defines utilities for unit testing Google Cloud Platform components of Apache Beam pipelines.
- org.apache.beam.sdk.io.hadoop - package org.apache.beam.sdk.io.hadoop
-
Classes shared by Hadoop based IOs.
- org.apache.beam.sdk.io.hadoop.inputformat - package org.apache.beam.sdk.io.hadoop.inputformat
-
Defines transforms for reading from Data sources which implement Hadoop Input Format.
- org.apache.beam.sdk.io.hbase - package org.apache.beam.sdk.io.hbase
-
Defines transforms for reading and writing from HBase.
- org.apache.beam.sdk.io.hcatalog - package org.apache.beam.sdk.io.hcatalog
-
Transforms for reading and writing using HCatalog.
- org.apache.beam.sdk.io.hdfs - package org.apache.beam.sdk.io.hdfs
-
FileSystem
implementation for any Hadoop
FileSystem
.
- org.apache.beam.sdk.io.jdbc - package org.apache.beam.sdk.io.jdbc
-
Transforms for reading and writing from JDBC.
- org.apache.beam.sdk.io.jms - package org.apache.beam.sdk.io.jms
-
Transforms for reading and writing from JMS (Java Messaging Service).
- org.apache.beam.sdk.io.kafka - package org.apache.beam.sdk.io.kafka
-
Transforms for reading and writing from Apache Kafka.
- org.apache.beam.sdk.io.kafka.serialization - package org.apache.beam.sdk.io.kafka.serialization
-
Kafka serializers and deserializers.
- org.apache.beam.sdk.io.kinesis - package org.apache.beam.sdk.io.kinesis
-
Transforms for reading and writing from Amazon Kinesis.
- org.apache.beam.sdk.io.mongodb - package org.apache.beam.sdk.io.mongodb
-
Transforms for reading and writing from MongoDB.
- org.apache.beam.sdk.io.mqtt - package org.apache.beam.sdk.io.mqtt
-
Transforms for reading and writing from MQTT.
- org.apache.beam.sdk.io.range - package org.apache.beam.sdk.io.range
-
Provides thread-safe helpers for implementing dynamic work rebalancing in position-based
bounded sources.
- org.apache.beam.sdk.io.redis - package org.apache.beam.sdk.io.redis
-
Transforms for reading and writing from Redis.
- org.apache.beam.sdk.io.solr - package org.apache.beam.sdk.io.solr
-
Transforms for reading and writing from/to Solr.
- org.apache.beam.sdk.io.tika - package org.apache.beam.sdk.io.tika
-
Transform for reading and parsing files with Apache Tika.
- org.apache.beam.sdk.io.xml - package org.apache.beam.sdk.io.xml
-
Transforms for reading and writing Xml files.
- org.apache.beam.sdk.metrics - package org.apache.beam.sdk.metrics
-
Metrics allow exporting information about the execution of a pipeline.
- org.apache.beam.sdk.options - package org.apache.beam.sdk.options
-
- org.apache.beam.sdk.state - package org.apache.beam.sdk.state
-
Classes and interfaces for interacting with state.
- org.apache.beam.sdk.testing - package org.apache.beam.sdk.testing
-
Defines utilities for unit testing Apache Beam pipelines.
- org.apache.beam.sdk.transforms - package org.apache.beam.sdk.transforms
-
Defines
PTransform
s for transforming
data in a pipeline.
- org.apache.beam.sdk.transforms.display - package org.apache.beam.sdk.transforms.display
-
- org.apache.beam.sdk.transforms.join - package org.apache.beam.sdk.transforms.join
-
Defines the
CoGroupByKey
transform
for joining multiple PCollections.
- org.apache.beam.sdk.transforms.splittabledofn - package org.apache.beam.sdk.transforms.splittabledofn
-
- org.apache.beam.sdk.transforms.windowing - package org.apache.beam.sdk.transforms.windowing
-
Defines the
Window
transform
for dividing the elements in a PCollection into windows, and the
Trigger
for controlling when those
elements are output.
- org.apache.beam.sdk.values - package org.apache.beam.sdk.values
-
- out() - Static method in class org.apache.beam.runners.spark.io.ConsoleIO.Write
-
- out(int) - Static method in class org.apache.beam.runners.spark.io.ConsoleIO.Write
-
- OUTER - Static variable in class org.apache.beam.sdk.coders.Coder.Context
-
Deprecated.
The outer context: the value being encoded or decoded takes
up the remainder of the record/stream contents.
- OUTPUT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- output(T) - Method in interface org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.ParserCallback
-
Output the object.
- output(T, Instant) - Method in interface org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.ParserCallback
-
Output the object using the specified timestamp.
- output(OutputT, Instant, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFn.FinishBundleContext
-
Adds the given element to the main output PCollection
at the given
timestamp in the given window.
- output(TupleTag<T>, T, Instant, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFn.FinishBundleContext
-
Adds the given element to the output PCollection
with the given tag at the given
timestamp in the given window.
- output(T) - Method in interface org.apache.beam.sdk.transforms.DoFn.OutputReceiver
-
- output(OutputT) - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
-
Adds the given element to the main output PCollection
.
- output(TupleTag<T>, T) - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
-
Adds the given element to the output PCollection
with the
given tag.
- OUTPUT_INFO - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- OUTPUT_NAME - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- outputOf(SerializableFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-
Returns a type descriptor for the output of the given
SerializableFunction
, subject to
Java type erasure: returns
null
if the type was erased.
- OutputReference - Class in org.apache.beam.runners.dataflow.util
-
A representation used by Step
s
to reference the output of other Step
s.
- OutputReference(String, String) - Constructor for class org.apache.beam.runners.dataflow.util.OutputReference
-
- outputRuntimeOptions() - Method in interface org.apache.beam.sdk.options.PipelineOptions
-
- outputType - Variable in class org.apache.beam.sdk.extensions.sql.impl.interpreter.operator.BeamSqlExpression
-
- outputWithTimestamp(OutputT, Instant) - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
-
Adds the given element to the main output PCollection
,
with the given timestamp.
- outputWithTimestamp(TupleTag<T>, T, Instant) - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
-
Adds the given element to the specified output PCollection
,
with the given timestamp.
- overlaps(ByteKeyRange) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
-
Returns
true
if the specified
ByteKeyRange
overlaps this range.
- pane() - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContext
-
Returns information about the pane within this window into which the
input element has been assigned.
- PaneInfo - Class in org.apache.beam.sdk.transforms.windowing
-
Provides information about the pane an element belongs to.
- PaneInfo.PaneInfoCoder - Class in org.apache.beam.sdk.transforms.windowing
-
A Coder for encoding PaneInfo instances.
- PaneInfo.Timing - Enum in org.apache.beam.sdk.transforms.windowing
-
Enumerates the possibilities for the timing of this pane firing related to the
input and output watermarks for its computation.
- PARALLEL_INPUT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- Params() - Constructor for class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
-
Construct a default Params object.
- ParamsCoder() - Constructor for class org.apache.beam.sdk.io.DefaultFilenamePolicy.ParamsCoder
-
- ParDo - Class in org.apache.beam.sdk.transforms
-
ParDo
is the core element-wise transform in Apache Beam, invoking a user-specified
function on each of the elements of the input
PCollection
to produce zero or more output
elements, all of which are collected into the output
PCollection
.
- ParDo() - Constructor for class org.apache.beam.sdk.transforms.ParDo
-
- ParDo.MultiOutput<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
-
A
PTransform
that, when applied to a
PCollection<InputT>
, invokes a
user-specified
DoFn<InputT, OutputT>
on all its elements, which can emit elements to
any of the
PTransform
's output
PCollection
s, which are bundled into a result
PCollectionTuple
.
- ParDo.SingleOutput<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
-
A
PTransform
that, when applied to a
PCollection<InputT>
,
invokes a user-specified
DoFn<InputT, OutputT>
on all its elements,
with all its outputs collected into an output
PCollection<OutputT>
.
- ParDoMultiOutputTranslator<InputT,OutputT> - Class in org.apache.beam.runners.gearpump.translators
-
ParDo.MultiOutput
is translated to Gearpump flatMap function
with
DoFn
wrapped in
DoFnFunction
.
- ParDoMultiOutputTranslator() - Constructor for class org.apache.beam.runners.gearpump.translators.ParDoMultiOutputTranslator
-
- Parse() - Constructor for class org.apache.beam.sdk.io.AvroIO.Parse
-
- parse(GridFSDBFile, MongoDbGridFSIO.ParserCallback<T>) - Method in interface org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Parser
-
- ParseAll() - Constructor for class org.apache.beam.sdk.io.AvroIO.ParseAll
-
- parseAllGenericRecords(SerializableFunction<GenericRecord, T>) - Static method in class org.apache.beam.sdk.io.AvroIO
-
- parseGenericRecords(SerializableFunction<GenericRecord, T>) - Static method in class org.apache.beam.sdk.io.AvroIO
-
Reads Avro file(s) containing records of an unspecified schema and converting each record to a
custom type.
- ParseJsons<OutputT> - Class in org.apache.beam.sdk.extensions.jackson
-
- parseQuery(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamQueryPlanner
-
Parse input SQL query, and return a SqlNode
as grammar tree.
- parseTableSpec(String) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
-
Parse a table specification in the form "[project_id]:[dataset_id].[table_id]"
or
"[dataset_id].[table_id]"
.
- Partition<T> - Class in org.apache.beam.sdk.transforms
-
Partition
takes a PCollection<T>
and a
PartitionFn
, uses the PartitionFn
to split the
elements of the input PCollection
into N
partitions, and
returns a PCollectionList<T>
that bundles N
PCollection<T>
s containing the split elements.
- Partition.PartitionFn<T> - Interface in org.apache.beam.sdk.transforms
-
A function object that chooses an output partition for an element.
- partitioner() - Method in class org.apache.beam.runners.spark.io.SourceRDD.Unbounded
-
- partitionFor(T, int) - Method in interface org.apache.beam.sdk.transforms.Partition.PartitionFn
-
Chooses the partition into which to put the given element.
- PartitioningWindowFn<T,W extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
-
A
WindowFn
that places each value into exactly one window based on its timestamp and
never merges windows.
- PartitioningWindowFn() - Constructor for class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
-
- PartitionMark(String, int, long) - Constructor for class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
-
- PAssert - Class in org.apache.beam.sdk.testing
-
An assertion on the contents of a
PCollection
incorporated into the pipeline.
- PAssert.DefaultConcludeTransform - Class in org.apache.beam.sdk.testing
-
Default transform to check that a PAssert was successful.
- PAssert.GroupThenAssert<T> - Class in org.apache.beam.sdk.testing
-
A transform that applies an assertion-checking function over iterables of ActualT
to
the entirety of the contents of its input.
- PAssert.GroupThenAssertForSingleton<T> - Class in org.apache.beam.sdk.testing
-
A transform that applies an assertion-checking function to a single iterable contained as the
sole element of a
PCollection
.
- PAssert.IterableAssert<T> - Interface in org.apache.beam.sdk.testing
-
Builder interface for assertions applicable to iterables and PCollection contents.
- PAssert.OneSideInputAssert<ActualT> - Class in org.apache.beam.sdk.testing
-
An assertion checker that takes a single
PCollectionView<ActualT>
and an assertion over
ActualT
, and checks it within a
Beam pipeline.
- PAssert.PAssertionSite - Class in org.apache.beam.sdk.testing
-
Track the place where an assertion is defined.
- PAssert.PCollectionContentsAssert<T> - Class in org.apache.beam.sdk.testing
-
- PAssert.PCollectionContentsAssert.MatcherCheckerFn<T> - Class in org.apache.beam.sdk.testing
-
Check that the passed-in matchers match the existing data.
- PAssert.SingletonAssert<T> - Interface in org.apache.beam.sdk.testing
-
Builder interface for assertions applicable to a single value.
- pastEndOfWindow() - Static method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark
-
Creates a trigger that fires when the watermark passes the end of the window.
- pastFirstElementInPane() - Static method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
-
Creates a trigger that fires when the current processing time passes the processing time at
which this trigger saw the first element in a pane.
- PathValidator - Interface in org.apache.beam.sdk.extensions.gcp.storage
-
For internal use only; no backwards compatibility guarantees.
- PathValidatorFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcsOptions.PathValidatorFactory
-
- PBegin - Class in org.apache.beam.sdk.values
-
- PBegin(Pipeline) - Constructor for class org.apache.beam.sdk.values.PBegin
-
- PCollection<T> - Class in org.apache.beam.sdk.values
-
- PCollection.IsBounded - Enum in org.apache.beam.sdk.values
-
The enumeration of cases for whether a
PCollection
is bounded.
- PCollectionContentsAssert(PCollection<T>, PAssert.PAssertionSite) - Constructor for class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-
- PCollectionContentsAssert(PCollection<T>, PAssert.AssertionWindows, SimpleFunction<Iterable<ValueInSingleWindow<T>>, Iterable<T>>, PAssert.PAssertionSite) - Constructor for class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-
- PCollectionList<T> - Class in org.apache.beam.sdk.values
-
- pCollections() - Static method in class org.apache.beam.sdk.transforms.Flatten
-
- PCollectionTuple - Class in org.apache.beam.sdk.values
-
- PCollectionView<T> - Interface in org.apache.beam.sdk.values
-
- PCollectionViews - Class in org.apache.beam.sdk.values
-
For internal use only; no backwards compatibility guarantees.
- PCollectionViews() - Constructor for class org.apache.beam.sdk.values.PCollectionViews
-
- PCollectionViews.IterableViewFn<T> - Class in org.apache.beam.sdk.values
-
- PCollectionViews.ListViewFn<T> - Class in org.apache.beam.sdk.values
-
- PCollectionViews.MapViewFn<K,V> - Class in org.apache.beam.sdk.values
-
- PCollectionViews.MultimapViewFn<K,V> - Class in org.apache.beam.sdk.values
-
- PCollectionViews.SimplePCollectionView<ElemT,ViewT,W extends BoundedWindow> - Class in org.apache.beam.sdk.values
-
A class for
PCollectionView
implementations, with additional type parameters
that are not visible at pipeline assembly time when the view is used as a side input.
- PCollectionViews.SingletonViewFn<T> - Class in org.apache.beam.sdk.values
-
- PDone - Class in org.apache.beam.sdk.values
-
- peekOutputElements() - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
Returns the elements output so far to the main output.
- peekOutputElements(TupleTag<T>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
Returns the elements output so far to the output with the
given tag.
- peekOutputElementsInWindow(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
Returns the elements output so far to the main output in the provided window with associated
timestamps.
- peekOutputElementsInWindow(TupleTag<OutputT>, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
Returns the elements output so far to the specified output in the provided window with
associated timestamps.
- peekOutputElementsWithTimestamp() - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
Returns the elements output so far to the main output with associated timestamps.
- perElement() - Static method in class org.apache.beam.sdk.transforms.Count
-
- perKey(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles
-
Returns a PTransform
that takes a
PCollection<KV<K, V>>
and returns a
PCollection<KV<K, List<V>>>
that contains an output
element mapping each distinct key in the input
PCollection
to a List
of the approximate
N
-tiles of the values associated with that key in the
input PCollection
.
- perKey(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles
-
- perKey(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateUnique
-
Returns a PTransform
that takes a
PCollection<KV<K, V>>
and returns a
PCollection<KV<K, Long>>
that contains an output element
mapping each distinct key in the input PCollection
to an
estimate of the number of distinct values associated with that
key in the input PCollection
.
- perKey(double) - Static method in class org.apache.beam.sdk.transforms.ApproximateUnique
-
- perKey(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
-
Returns a
Combine.PerKey
PTransform
that
first groups its input
PCollection
of
KV
s by keys and
windows, then invokes the given function on each of the values lists to
produce a combined value, and then returns a
PCollection
of
KV
s mapping each distinct key to its combined value for each
window.
- perKey(CombineFnBase.GlobalCombineFn<? super InputT, ?, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Combine
-
Returns a
Combine.PerKey
PTransform
that
first groups its input
PCollection
of
KV
s by keys and
windows, then invokes the given function on each of the values lists to
produce a combined value, and then returns a
PCollection
of
KV
s mapping each distinct key to its combined value for each
window.
- perKey() - Static method in class org.apache.beam.sdk.transforms.Count
-
Returns a
PTransform
that counts the number of elements
associated with each key of its input
PCollection
.
- perKey() - Static method in class org.apache.beam.sdk.transforms.Latest
-
Returns a
PTransform
that takes as input a
PCollection<KV<K, V>>
and returns a
PCollection<KV<K, V>>
whose contents is the latest element per-key according to its
event time.
- perKey() - Static method in class org.apache.beam.sdk.transforms.Max
-
Returns a PTransform
that takes an input PCollection<KV<K, T>>
and returns a
PCollection<KV<K, T>>
that contains an output element mapping each distinct key in the
input PCollection
to the maximum according to the natural ordering of T
of the
values associated with that key in the input PCollection
.
- perKey(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Max
-
Returns a PTransform
that takes an input PCollection<KV<K, T>>
and returns a
PCollection<KV<K, T>>
that contains one output element per key mapping each
to the maximum of the values associated with that key in the input PCollection
.
- perKey() - Static method in class org.apache.beam.sdk.transforms.Mean
-
Returns a PTransform
that takes an input
PCollection<KV<K, N>>
and returns a
PCollection<KV<K, Double>>
that contains an output
element mapping each distinct key in the input
PCollection
to the mean of the values associated with
that key in the input PCollection
.
- perKey() - Static method in class org.apache.beam.sdk.transforms.Min
-
Returns a PTransform
that takes an input PCollection<KV<K, T>>
and returns a
PCollection<KV<K, T>>
that contains an output element mapping each distinct key in the
input PCollection
to the minimum according to the natural ordering of T
of the
values associated with that key in the input PCollection
.
- perKey(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Min
-
Returns a PTransform
that takes an input PCollection<KV<K, T>>
and returns a
PCollection<KV<K, T>>
that contains one output element per key mapping each
to the minimum of the values associated with that key in the input PCollection
.
- perKey(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Top
-
Returns a PTransform
that takes an input
PCollection<KV<K, V>>
and returns a
PCollection<KV<K, List<V>>>
that contains an output
element mapping each distinct key in the input
PCollection
to the largest count
values
associated with that key in the input
PCollection<KV<K, V>>
, in decreasing order, sorted using
the given Comparator<V>
.
- PInput - Interface in org.apache.beam.sdk.values
-
The interface for things that might be input to a
PTransform
.
- Pipeline - Class in org.apache.beam.sdk
-
- Pipeline(PipelineOptions) - Constructor for class org.apache.beam.sdk.Pipeline
-
- Pipeline.PipelineExecutionException - Exception in org.apache.beam.sdk
-
Thrown during execution of a
Pipeline
, whenever user code within that
Pipeline
throws an exception.
- Pipeline.PipelineVisitor - Interface in org.apache.beam.sdk
-
For internal use only; no backwards-compatibility guarantees.
- Pipeline.PipelineVisitor.CompositeBehavior - Enum in org.apache.beam.sdk
-
Control enum for indicating whether or not a traversal should process the contents of
a composite transform or not.
- Pipeline.PipelineVisitor.Defaults - Class in org.apache.beam.sdk
-
- pipelineExecution - Variable in class org.apache.beam.runners.spark.SparkPipelineResult
-
- PipelineExecutionException(Throwable) - Constructor for exception org.apache.beam.sdk.Pipeline.PipelineExecutionException
-
- PipelineOptions - Interface in org.apache.beam.sdk.options
-
PipelineOptions are used to configure Pipelines.
- PipelineOptions.AtomicLongFactory - Class in org.apache.beam.sdk.options
-
DefaultValueFactory
which supplies an ID that is guaranteed to be unique
within the given process.
- PipelineOptions.CheckEnabled - Enum in org.apache.beam.sdk.options
-
Enumeration of the possible states for a given check.
- PipelineOptions.DirectRunner - Class in org.apache.beam.sdk.options
-
A
DefaultValueFactory
that obtains the class of the
DirectRunner
if it exists
on the classpath, and throws an exception otherwise.
- PipelineOptions.JobNameFactory - Class in org.apache.beam.sdk.options
-
- PipelineOptions.UserAgentFactory - Class in org.apache.beam.sdk.options
-
Returns a user agent string constructed from ReleaseInfo.getName()
and
ReleaseInfo.getVersion()
, in the format [name]/[version]
.
- PipelineOptionsFactory - Class in org.apache.beam.sdk.options
-
- PipelineOptionsFactory() - Constructor for class org.apache.beam.sdk.options.PipelineOptionsFactory
-
- PipelineOptionsFactory.Builder - Class in org.apache.beam.sdk.options
-
- PipelineOptionsRegistrar - Interface in org.apache.beam.sdk.options
-
PipelineOptions
creators have the ability to automatically have their
PipelineOptions
registered with this SDK by creating a
ServiceLoader
entry
and a concrete implementation of this interface.
- PipelineOptionsValidator - Class in org.apache.beam.sdk.options
-
- PipelineOptionsValidator() - Constructor for class org.apache.beam.sdk.options.PipelineOptionsValidator
-
- PipelineResult - Interface in org.apache.beam.sdk
-
- PipelineResult.State - Enum in org.apache.beam.sdk
-
Possible job states, for both completed and ongoing jobs.
- PipelineRunner<ResultT extends PipelineResult> - Class in org.apache.beam.sdk
-
- PipelineRunner() - Constructor for class org.apache.beam.sdk.PipelineRunner
-
- planner - Variable in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamQueryPlanner
-
- plusDelayOf(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
-
Adds some delay to the original target time.
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.amqp.AmqpIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.AvroIO.ParseAll
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.AvroIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.AvroIO.ReadAll
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.AvroIO.Write
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.BoundedReadFromUnboundedSource
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.CompressedSource
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
-
Populates the display data.
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileBasedSink.FilenamePolicy
-
Populates the display data.
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileBasedSink
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileBasedSource
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileIO.ReadMatches
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write.PubsubBoundedWriter
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.GenerateSequence
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.HadoopInputFormatBoundedSource
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.TypedWithoutMetadata
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Write
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.Read.Bounded
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.Read.Unbounded
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
-
Populate the display data with connectionConfiguration details.
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.Source
-
Register display data for the given transform or component.
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TextIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TextIO.Write
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.WriteFiles
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.Globally
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.GloballyAsSingletonView
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.PerKeyWithHotKeyFanout
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
-
- populateDisplayData(DisplayData.Builder) - Method in interface org.apache.beam.sdk.transforms.display.HasDisplayData
-
Register display data for the given transform or component.
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.DoFn
-
Register display data for the given transform or component.
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Filter
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.FlatMapElements
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.GroupByKey
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.MapElements
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
-
Register display data for the given transform or component.
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Partition
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.PTransform
-
Register display data for the given transform or component.
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.SimpleFunction
-
Register display data for the given transform or component.
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.Window.Assign
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.Window
-
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
-
Register display data for the given transform or component.
- POutput - Interface in org.apache.beam.sdk.values
-
The interface for things that might be output from a
PTransform
.
- prepare(JobApi.PrepareJobRequest, StreamObserver<JobApi.PrepareJobResponse>) - Method in class org.apache.beam.runners.reference.job.ReferenceRunnerJobService
-
- prepare() - Method in interface org.apache.beam.sdk.extensions.sql.impl.interpreter.BeamSqlExpressionExecutor
-
invoked before data processing.
- prepare() - Method in class org.apache.beam.sdk.extensions.sql.impl.interpreter.BeamSqlFnExecutor
-
- prepareForProcessing() - Method in class org.apache.beam.sdk.transforms.DoFn
-
- prepareWrite(WritableByteChannel) - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
-
Called with the channel that a subclass will write its header, footer, and values to.
- PrepareWrite<T,DestinationT> - Class in org.apache.beam.sdk.io.gcp.bigquery
-
- PrepareWrite(DynamicDestinations<T, DestinationT>, SerializableFunction<T, TableRow>) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.PrepareWrite
-
- primary() - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
-
- PrimitiveParDoSingleFactory<InputT,OutputT> - Class in org.apache.beam.runners.dataflow
-
- PrimitiveParDoSingleFactory() - Constructor for class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory
-
- PrimitiveParDoSingleFactory.ParDoSingle<InputT,OutputT> - Class in org.apache.beam.runners.dataflow
-
A single-output primitive
ParDo
.
- printHelp(PrintStream) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
-
Outputs the set of registered options with the PipelineOptionsFactory
with a description for each one if available to the output stream.
- printHelp(PrintStream, Class<? extends PipelineOptions>) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
-
Outputs the set of options available to be set for the passed in
PipelineOptions
interface.
- process(List<JobMessage>) - Method in interface org.apache.beam.runners.dataflow.util.MonitoringUtil.JobMessagesHandler
-
Process the rows.
- process(List<JobMessage>) - Method in class org.apache.beam.runners.dataflow.util.MonitoringUtil.LoggingHandler
-
- processBundle(Iterable<? extends InputT>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
- processBundle(InputT...) - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
A convenience method for testing
DoFns
with bundles of elements.
- ProcessContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.ProcessContext
-
- ProcessContinuation() - Constructor for class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
-
- processElement(DoFn<KV<K, Iterable<KV<Instant, WindowedValue<KV<K, V>>>>>, OutputT>.ProcessContext, BoundedWindow) - Method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides.BatchStatefulDoFn
-
- processElement(DoFn<Iterable<T>, T>.ProcessContext, BoundedWindow) - Method in class org.apache.beam.runners.dataflow.DataflowRunner.StreamingPCollectionViewWriterFn
-
- processElement(WindowedValue<InputT>) - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
-
- processElement(DoFn<KV<BeamRecord, BeamRecord>, BeamRecord>.ProcessContext, BoundedWindow) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamAggregationTransforms.MergeAggregationRecord
-
- processElement(DoFn<KV<BeamRecord, BeamRecord>, BeamRecord>.ProcessContext) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamJoinTransforms.SideInputJoinDoFn
-
- processElement(DoFn<KV<BeamRecord, CoGbkResult>, BeamRecord>.ProcessContext) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSetOperatorsTransforms.SetOperatorFilteringDoFn
-
- processElement(DoFn<BeamRecord, BeamRecord>.ProcessContext, BoundedWindow) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSqlFilterFn
-
- processElement(DoFn<BeamRecord, Void>.ProcessContext) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSqlOutputToConsoleFn
-
- processElement(DoFn<BeamRecord, BeamRecord>.ProcessContext, BoundedWindow) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSqlProjectFn
-
- processElement(DoFn<T, Void>.ProcessContext) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write.PubsubBoundedWriter
-
- processElement(InputT) - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
- processElement(DoFn<ValueWithRecordId<T>, T>.ProcessContext) - Method in class org.apache.beam.sdk.values.ValueWithRecordId.StripIdsDoFn
-
- ProcessingTimeEvent() - Constructor for class org.apache.beam.sdk.testing.TestStream.ProcessingTimeEvent
-
- processTimestampedElement(TimestampedValue<InputT>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
- ProcessWatcher(Process) - Constructor for class org.apache.beam.runners.apex.ApexYarnLauncher.ProcessWatcher
-
- processWindowedElement(InputT, Instant, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
- PROJECT_ID_REGEXP - Static variable in class org.apache.beam.runners.dataflow.DataflowRunner
-
Project IDs must contain lowercase letters, digits, or dashes.
- PROPERTY_BEAM_TEST_PIPELINE_OPTIONS - Static variable in class org.apache.beam.sdk.testing.TestPipeline
-
- propertyName() - Method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
-
Returns the property name associated with this provider.
- propertyName() - Method in class org.apache.beam.sdk.options.ValueProvider.RuntimeValueProvider
-
Returns the property name that corresponds to this provider.
- PropertyNames - Class in org.apache.beam.runners.dataflow.util
-
Constant property names used by the SDK in CloudWorkflow specifications.
- PropertyNames() - Constructor for class org.apache.beam.runners.dataflow.util.PropertyNames
-
- ProtobufCoderProviderRegistrar - Class in org.apache.beam.sdk.extensions.protobuf
-
- ProtobufCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.extensions.protobuf.ProtobufCoderProviderRegistrar
-
- ProtoCoder<T extends com.google.protobuf.Message> - Class in org.apache.beam.sdk.extensions.protobuf
-
A
Coder
using Google Protocol Buffers binary format.
- PTransform<InputT extends PInput,OutputT extends POutput> - Class in org.apache.beam.sdk.transforms
-
A
PTransform<InputT, OutputT>
is an operation that takes an
InputT
(some subtype of
PInput
) and produces an
OutputT
(some subtype of
POutput
).
- PTransform() - Constructor for class org.apache.beam.sdk.transforms.PTransform
-
- PTransform(String) - Constructor for class org.apache.beam.sdk.transforms.PTransform
-
- PUBSUB_ID_ATTRIBUTE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- PUBSUB_SERIALIZED_ATTRIBUTES_FN - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- PUBSUB_SUBSCRIPTION - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- PUBSUB_SUBSCRIPTION_OVERRIDE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- PUBSUB_TIMESTAMP_ATTRIBUTE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- PUBSUB_TOPIC - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- PUBSUB_TOPIC_OVERRIDE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- PubsubBoundedWriter() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write.PubsubBoundedWriter
-
- PubsubCoderProviderRegistrar - Class in org.apache.beam.sdk.io.gcp.pubsub
-
- PubsubCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubCoderProviderRegistrar
-
- PubsubIO - Class in org.apache.beam.sdk.io.gcp.pubsub
-
Read and Write
PTransform
s for Cloud Pub/Sub streams.
- PubsubIO.PubsubSubscription - Class in org.apache.beam.sdk.io.gcp.pubsub
-
Class representing a Cloud Pub/Sub Subscription.
- PubsubIO.PubsubTopic - Class in org.apache.beam.sdk.io.gcp.pubsub
-
Class representing a Cloud Pub/Sub Topic.
- PubsubIO.Read<T> - Class in org.apache.beam.sdk.io.gcp.pubsub
-
- PubsubIO.Write<T> - Class in org.apache.beam.sdk.io.gcp.pubsub
-
- PubsubIO.Write.PubsubBoundedWriter - Class in org.apache.beam.sdk.io.gcp.pubsub
-
Writer to Pubsub which batches messages from bounded collections.
- PubsubMessage - Class in org.apache.beam.sdk.io.gcp.pubsub
-
Class representing a Pub/Sub message.
- PubsubMessage(byte[], Map<String, String>) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
-
- PubsubMessagePayloadOnlyCoder - Class in org.apache.beam.sdk.io.gcp.pubsub
-
A coder for PubsubMessage treating the raw bytes being decoded as the message's payload.
- PubsubMessagePayloadOnlyCoder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessagePayloadOnlyCoder
-
- PubsubMessageWithAttributesCoder - Class in org.apache.beam.sdk.io.gcp.pubsub
-
A coder for PubsubMessage including attributes.
- PubsubMessageWithAttributesCoder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
-
- PubsubOptions - Interface in org.apache.beam.sdk.io.gcp.pubsub
-
Properties that can be set when using Google Cloud Pub/Sub with the Apache Beam SDK.
- PubsubUnboundedSink - Class in org.apache.beam.sdk.io.gcp.pubsub
-
A PTransform which streams messages to Pubsub.
- PubsubUnboundedSink(PubsubClient.PubsubClientFactory, ValueProvider<PubsubClient.TopicPath>, String, String, int) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
-
- PubsubUnboundedSource - Class in org.apache.beam.sdk.io.gcp.pubsub
-
- PubsubUnboundedSource(PubsubClient.PubsubClientFactory, ValueProvider<PubsubClient.ProjectPath>, ValueProvider<PubsubClient.TopicPath>, ValueProvider<PubsubClient.SubscriptionPath>, String, String, boolean) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
-
Construct an unbounded source to consume from the Pubsub subscription
.
- put(K, V) - Method in interface org.apache.beam.sdk.state.MapState
-
Associates the specified value with the specified key in this state.
- putArtifact(StreamObserver<ArtifactApi.PutArtifactResponse>) - Method in class org.apache.beam.artifact.local.LocalFileSystemArtifactStagerService
-
- putIfAbsent(K, V) - Method in interface org.apache.beam.sdk.state.MapState
-
A deferred read-followed-by-write.
- PValue - Interface in org.apache.beam.sdk.values
-
For internal use.
- PValueBase - Class in org.apache.beam.sdk.values
-
For internal use.
- PValueBase(Pipeline) - Constructor for class org.apache.beam.sdk.values.PValueBase
-
- PValueBase() - Constructor for class org.apache.beam.sdk.values.PValueBase
-
No-arg constructor for Java serialization only.
- RandomAccessData - Class in org.apache.beam.runners.dataflow.util
-
An elastic-sized byte array which allows you to manipulate it as a stream, or access
it directly.
- RandomAccessData() - Constructor for class org.apache.beam.runners.dataflow.util.RandomAccessData
-
Constructs a RandomAccessData with a default buffer size.
- RandomAccessData(byte[]) - Constructor for class org.apache.beam.runners.dataflow.util.RandomAccessData
-
Constructs a RandomAccessData with the initial buffer.
- RandomAccessData(int) - Constructor for class org.apache.beam.runners.dataflow.util.RandomAccessData
-
Constructs a RandomAccessData with the given buffer size.
- RandomAccessData.RandomAccessDataCoder - Class in org.apache.beam.runners.dataflow.util
-
A
Coder
which encodes the valid parts of this stream.
- RandomAccessData.UnsignedLexicographicalComparator - Class in org.apache.beam.runners.dataflow.util
-
A Comparator
that compares two byte arrays lexicographically.
- RandomAccessDataCoder() - Constructor for class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
-
- RangeTracker<PositionT> - Interface in org.apache.beam.sdk.io.range
-
A
RangeTracker
is a thread-safe helper object for implementing dynamic work rebalancing
in position-based
BoundedSource.BoundedReader
subclasses.
- RawUnionValue(String, Object) - Constructor for class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils.RawUnionValue
-
Constructs a partial union from the given union tag and value.
- RawUnionValue - Class in org.apache.beam.sdk.transforms.join
-
This corresponds to an integer union tag and value.
- RawUnionValue(int, Object) - Constructor for class org.apache.beam.sdk.transforms.join.RawUnionValue
-
Constructs a partial union from the given union tag and value.
- read() - Method in class org.apache.beam.runners.gearpump.translators.io.GearpumpSource
-
- read(JavaStreamingContext, SerializablePipelineOptions, UnboundedSource<T, CheckpointMarkT>, String) - Static method in class org.apache.beam.runners.spark.io.SparkUnboundedSource
-
- read() - Static method in class org.apache.beam.sdk.io.amqp.AmqpIO
-
- Read() - Constructor for class org.apache.beam.sdk.io.amqp.AmqpIO.Read
-
- read(Class<T>) - Static method in class org.apache.beam.sdk.io.AvroIO
-
Reads records of the given type from an Avro file (or multiple Avro files matching a pattern).
- Read() - Constructor for class org.apache.beam.sdk.io.AvroIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.cassandra.CassandraIO
-
- Read() - Constructor for class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO
-
- Read() - Constructor for class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
-
- read(SerializableFunction<SchemaAndRecord, T>) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
-
Reads from a BigQuery table or query and returns a
PCollection
with one element per
each row of the table or query result, parsed from the BigQuery AVRO format using the specified
function.
- read() - Static method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO
-
- Read() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-
- read() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1
-
- Read() - Constructor for class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
- Read() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO
-
- Read() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO
-
- Read() - Constructor for class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.hbase.HBaseIO
-
- read() - Static method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO
-
Read data from Hive.
- Read() - Constructor for class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
-
Read data from a JDBC datasource.
- Read() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.jms.JmsIO
-
- Read() - Constructor for class org.apache.beam.sdk.io.jms.JmsIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.kafka.KafkaIO
-
- Read() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.kinesis.KinesisIO
-
- Read() - Constructor for class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO
-
Read data from GridFS.
- Read() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbIO
-
Read data from MongoDB.
- Read() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.mqtt.MqttIO
-
- Read() - Constructor for class org.apache.beam.sdk.io.mqtt.MqttIO.Read
-
- Read - Class in org.apache.beam.sdk.io
-
- Read() - Constructor for class org.apache.beam.sdk.io.Read
-
- read() - Static method in class org.apache.beam.sdk.io.redis.RedisIO
-
Read data from a Redis server.
- Read() - Constructor for class org.apache.beam.sdk.io.redis.RedisIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.solr.SolrIO
-
- Read() - Constructor for class org.apache.beam.sdk.io.solr.SolrIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.TextIO
-
A
PTransform
that reads from one or more text files and returns a bounded
PCollection
containing one element for each line of the input files.
- Read() - Constructor for class org.apache.beam.sdk.io.TextIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.TFRecordIO
-
A
PTransform
that reads from a TFRecord file (or multiple TFRecord
files matching a pattern) and returns a
PCollection
containing
the decoding of each of the records of the TFRecord file(s) as a byte array.
- Read() - Constructor for class org.apache.beam.sdk.io.TFRecordIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.tika.TikaIO
-
A
PTransform
that parses one or more files and returns a bounded
PCollection
containing one element for each sequence of characters reported by Apache Tika SAX Parser.
- Read() - Constructor for class org.apache.beam.sdk.io.tika.TikaIO.Read
-
- read() - Static method in class org.apache.beam.sdk.io.xml.XmlIO
-
Reads XML files as a
PCollection
of a given type mapped via JAXB.
- Read() - Constructor for class org.apache.beam.sdk.io.xml.XmlIO.Read
-
- read() - Method in interface org.apache.beam.sdk.state.ReadableState
-
Read the current value, blocking until it is available.
- Read.Bounded<T> - Class in org.apache.beam.sdk.io
-
- Read.Builder - Class in org.apache.beam.sdk.io
-
Helper class for building
Read
transforms.
- Read.Unbounded<T> - Class in org.apache.beam.sdk.io
-
- ReadableFileCoder - Class in org.apache.beam.sdk.io
-
- ReadableFileCoder() - Constructor for class org.apache.beam.sdk.io.ReadableFileCoder
-
- ReadableState<T> - Interface in org.apache.beam.sdk.state
-
- ReadableStates - Class in org.apache.beam.sdk.state
-
For internal use only; no backwards-compatibility guarantees.
- ReadableStates() - Constructor for class org.apache.beam.sdk.state.ReadableStates
-
- readAll(Class<T>) - Static method in class org.apache.beam.sdk.io.AvroIO
-
- ReadAll() - Constructor for class org.apache.beam.sdk.io.AvroIO.ReadAll
-
- readAll() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO
-
- ReadAll() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-
- readAll() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
-
Like
JdbcIO.read()
, but executes multiple instances of the query substituting each element
of a
PCollection
as query parameters.
- ReadAll() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-
- readAll() - Static method in class org.apache.beam.sdk.io.redis.RedisIO
-
Like
RedisIO.read()
but executes multiple instances of the Redis query substituting each
element of a
PCollection
as key pattern.
- ReadAll() - Constructor for class org.apache.beam.sdk.io.redis.RedisIO.ReadAll
-
- readAll() - Static method in class org.apache.beam.sdk.io.TextIO
-
- ReadAll() - Constructor for class org.apache.beam.sdk.io.TextIO.ReadAll
-
- readAllGenericRecords(Schema) - Static method in class org.apache.beam.sdk.io.AvroIO
-
- readAllGenericRecords(String) - Static method in class org.apache.beam.sdk.io.AvroIO
-
- ReadAllViaFileBasedSource<T> - Class in org.apache.beam.sdk.io
-
- ReadAllViaFileBasedSource(long, SerializableFunction<String, ? extends FileBasedSource<T>>, Coder<T>) - Constructor for class org.apache.beam.sdk.io.ReadAllViaFileBasedSource
-
- readAvros(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
-
Returns A
PTransform
that continuously reads binary encoded Avro messages of the
given type from a Google Cloud Pub/Sub stream.
- ReadBoundedTranslator<T> - Class in org.apache.beam.runners.gearpump.translators
-
Read.Bounded
is translated to Gearpump source function
and
BoundedSource
is wrapped into Gearpump
DataSource
.
- ReadBoundedTranslator() - Constructor for class org.apache.beam.runners.gearpump.translators.ReadBoundedTranslator
-
- readBytes() - Static method in class org.apache.beam.sdk.io.kafka.KafkaIO
-
- readDecompressed(ReadableByteChannel) - Method in enum org.apache.beam.sdk.io.Compression
-
- Reader() - Constructor for class org.apache.beam.sdk.io.Source.Reader
-
- ReaderInvocationUtil<OutputT,ReaderT extends Source.Reader<OutputT>> - Class in org.apache.beam.runners.flink.metrics
-
Util for invoking Source.Reader
methods that might require a
MetricsContainerImpl
to be active.
- ReaderInvocationUtil(String, PipelineOptions, FlinkMetricContainer) - Constructor for class org.apache.beam.runners.flink.metrics.ReaderInvocationUtil
-
- readExternal(ObjectInput) - Method in class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
-
- readFiles() - Static method in class org.apache.beam.sdk.io.TextIO
-
- ReadFiles() - Constructor for class org.apache.beam.sdk.io.TextIO.ReadFiles
-
- readFiles() - Static method in class org.apache.beam.sdk.io.xml.XmlIO
-
- ReadFiles() - Constructor for class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
-
- readFrom(InputStream, int, int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
-
Reads length
bytes from the specified input stream writing them into the backing
data store starting at offset
.
- readFromSource(BoundedSource<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-
- readFromSplitsOfSource(BoundedSource<T>, long, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-
- readFromStartedReader(Source.Reader<T>) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-
Reads all elements from the given started Source.Reader
.
- readFromUnstartedReader(Source.Reader<T>) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-
Reads all elements from the given unstarted Source.Reader
.
- readFullyAsBytes() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
-
Returns the full contents of the file as bytes.
- readFullyAsUTF8String() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
-
Returns the full contents of the file as a String
decoded as UTF-8.
- readGenericRecords(Schema) - Static method in class org.apache.beam.sdk.io.AvroIO
-
Reads Avro file(s) containing records of the specified schema.
- readGenericRecords(String) - Static method in class org.apache.beam.sdk.io.AvroIO
-
Reads Avro file(s) containing records of the specified schema.
- readLater() - Method in interface org.apache.beam.sdk.state.BagState
-
- readLater() - Method in interface org.apache.beam.sdk.state.CombiningState
-
- readLater() - Method in interface org.apache.beam.sdk.state.GroupingState
-
- readLater() - Method in interface org.apache.beam.sdk.state.ReadableState
-
Indicate that the value will be read later.
- readLater() - Method in interface org.apache.beam.sdk.state.SetState
-
- readLater() - Method in interface org.apache.beam.sdk.state.ValueState
-
- readLater() - Method in interface org.apache.beam.sdk.state.WatermarkHoldState
-
- readMatches() - Static method in class org.apache.beam.sdk.io.FileIO
-
- ReadMatches() - Constructor for class org.apache.beam.sdk.io.FileIO.ReadMatches
-
- readMessages() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
-
Returns A
PTransform
that continuously reads from a Google Cloud Pub/Sub stream.
- readMessagesWithAttributes() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
-
Returns A
PTransform
that continuously reads from a Google Cloud Pub/Sub stream.
- readNextBlock() - Method in class org.apache.beam.sdk.io.AvroSource.AvroReader
-
- readNextBlock() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
-
Read the next block from the input.
- readNextRecord() - Method in class org.apache.beam.sdk.io.BlockBasedSource.Block
-
Reads the next record from the block and returns true iff one exists.
- readNextRecord() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
-
- readNextRecord() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-
Reads the next record via the delegate reader.
- readNextRecord() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
-
- readNItemsFromStartedReader(Source.Reader<T>, int) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-
Read elements from a Source.Reader
that has already had Source.Reader#start
called on it, until n elements are read.
- readNItemsFromUnstartedReader(Source.Reader<T>, int) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-
Read elements from a Source.Reader
until n elements are read.
- ReadOperation - Class in org.apache.beam.sdk.io.gcp.spanner
-
Encapsulates a spanner read operation.
- ReadOperation() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-
- readProtos(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
-
Returns A
PTransform
that continuously reads binary encoded protobuf messages of the
given type from a Google Cloud Pub/Sub stream.
- readRemainingFromReader(Source.Reader<T>, boolean) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-
Read all remaining elements from a Source.Reader
.
- readStrings() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
-
Returns A
PTransform
that continuously reads UTF-8 encoded strings from a Google Cloud
Pub/Sub stream.
- readTableRows() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
-
- ReadUnboundedTranslator<T> - Class in org.apache.beam.runners.gearpump.translators
-
Read.Unbounded
is translated to Gearpump source function
and
UnboundedSource
is wrapped into Gearpump
DataSource
.
- ReadUnboundedTranslator() - Constructor for class org.apache.beam.runners.gearpump.translators.ReadUnboundedTranslator
-
- RedisConnectionConfiguration - Class in org.apache.beam.sdk.io.redis
-
RedisConnectionConfiguration
describes and wraps a connectionConfiguration to Redis
server or cluster.
- RedisConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
-
- RedisIO - Class in org.apache.beam.sdk.io.redis
-
An IO to manipulate Redis key/value database.
- RedisIO.Read - Class in org.apache.beam.sdk.io.redis
-
- RedisIO.ReadAll - Class in org.apache.beam.sdk.io.redis
-
- RedisIO.Write - Class in org.apache.beam.sdk.io.redis
-
- ReferenceRunnerJobServer - Class in org.apache.beam.runners.reference.job
-
- ReferenceRunnerJobServer() - Constructor for class org.apache.beam.runners.reference.job.ReferenceRunnerJobServer
-
- ReferenceRunnerJobService - Class in org.apache.beam.runners.reference.job
-
The ReferenceRunner uses the portability framework to execute a Pipeline on a single machine.
- Regex - Class in org.apache.beam.sdk.transforms
-
PTransorm
s to use Regular Expressions to process elements in a
PCollection
.
- Regex.AllMatches - Class in org.apache.beam.sdk.transforms
-
Regex.MatchesName<String>
takes a PCollection<String>
and returns a PCollection<List<String>>
representing the value extracted from all the
Regex groups of the input
PCollection
to the number of times that element occurs in the input.
- Regex.Find - Class in org.apache.beam.sdk.transforms
-
Regex.Find<String>
takes a PCollection<String>
and returns a PCollection<String>
representing the value extracted from the Regex groups of the input PCollection
to the number of times that element occurs in the input.
- Regex.FindAll - Class in org.apache.beam.sdk.transforms
-
Regex.Find<String>
takes a PCollection<String>
and returns a PCollection<List<String>>
representing the value extracted from the
Regex groups of the input PCollection
to the number of times that element occurs in the input.
- Regex.FindKV - Class in org.apache.beam.sdk.transforms
-
Regex.MatchesKV<KV<String, String>>
takes a PCollection<String>
and returns a
PCollection<KV<String, String>>
representing the key and value extracted from the Regex
groups of the input PCollection
to the number of times that element occurs in the
input.
- Regex.FindName - Class in org.apache.beam.sdk.transforms
-
Regex.Find<String>
takes a PCollection<String>
and returns a PCollection<String>
representing the value extracted from the Regex groups of the input PCollection
to the number of times that element occurs in the input.
- Regex.FindNameKV - Class in org.apache.beam.sdk.transforms
-
Regex.MatchesKV<KV<String, String>>
takes a PCollection<String>
and returns a
PCollection<KV<String, String>>
representing the key and value extracted from the Regex
groups of the input PCollection
to the number of times that element occurs in the
input.
- Regex.Matches - Class in org.apache.beam.sdk.transforms
-
Regex.Matches<String>
takes a PCollection<String>
and returns a PCollection<String>
representing the value extracted from the Regex groups of the input PCollection
to the number of times that element occurs in the input.
- Regex.MatchesKV - Class in org.apache.beam.sdk.transforms
-
Regex.MatchesKV<KV<String, String>>
takes a PCollection<String>
and returns a
PCollection<KV<String, String>>
representing the key and value extracted from the Regex
groups of the input PCollection
to the number of times that element occurs in the
input.
- Regex.MatchesName - Class in org.apache.beam.sdk.transforms
-
Regex.MatchesName<String>
takes a PCollection<String>
and returns a PCollection<String>
representing the value extracted from the Regex groups of the input PCollection
to the number of times that element occurs in the input.
- Regex.MatchesNameKV - Class in org.apache.beam.sdk.transforms
-
Regex.MatchesNameKV<KV<String, String>>
takes a PCollection<String>
and returns
a PCollection<KV<String, String>>
representing the key and value extracted from the
Regex groups of the input PCollection
to the number of times that element occurs in the
input.
- Regex.ReplaceAll - Class in org.apache.beam.sdk.transforms
-
Regex.ReplaceAll<String>
takes a PCollection<String>
and returns a PCollection<String>
with all Strings that matched the Regex being replaced with the
replacement string.
- Regex.ReplaceFirst - Class in org.apache.beam.sdk.transforms
-
Regex.ReplaceFirst<String>
takes a PCollection<String>
and returns a PCollection<String>
with the first Strings that matched the Regex being replaced with the
replacement string.
- Regex.Split - Class in org.apache.beam.sdk.transforms
-
Regex.Split<String>
takes a PCollection<String>
and returns a PCollection<String>
with the input string split into individual items in a list.
- RegexMatcher - Class in org.apache.beam.sdk.testing
-
Hamcrest matcher to assert a string matches a pattern.
- RegexMatcher(String) - Constructor for class org.apache.beam.sdk.testing.RegexMatcher
-
- register(RelOptPlanner) - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-
- register(Class<? extends PipelineOptions>) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
-
This registers the interface with this factory.
- registerByteSizeObserver(T, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.Coder
-
Notifies the ElementByteSizeObserver
about the byte size
of the encoded value using this Coder
.
- registerByteSizeObserver(ReadableDuration, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.DurationCoder
-
- registerByteSizeObserver(IterableT, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
-
- registerByteSizeObserver(KV<K, V>, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.KvCoder
-
Notifies ElementByteSizeObserver about the byte size of the
encoded value using this coder.
- registerByteSizeObserver(Map<K, V>, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.MapCoder
-
- registerByteSizeObserver(T, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.NullableCoder
-
Overridden to short-circuit the default StructuredCoder
behavior of encoding and
counting the bytes.
- registerByteSizeObserver(RawUnionValue, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
-
Notifies ElementByteSizeObserver about the byte size of the encoded value using this coder.
- registerClasses(Kryo) - Method in class org.apache.beam.runners.spark.coders.BeamSparkRunnerRegistrator
-
- registerCoderForClass(Class<?>, Coder<?>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
-
Registers the provided
Coder
for the given class.
- registerCoderForType(TypeDescriptor<?>, Coder<?>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
-
Registers the provided
Coder
for the given type.
- registerCoderProvider(CoderProvider) - Method in class org.apache.beam.sdk.coders.CoderRegistry
-
Registers
coderProvider
as a potential
CoderProvider
which can produce
Coder
instances.
- registerTable(String, BeamSqlTable) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
-
Registers a
BaseBeamTable
which can be used for all subsequent queries.
- registerTransformTranslator(Class<TransformT>, TransformTranslator<? extends TransformT>) - Static method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator
-
Records that instances of the specified PTransform class
should be translated by default by the corresponding
TransformTranslator
.
- registerUdaf(String, Combine.CombineFn) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
-
Register a UDAF function which can be used in GROUP-BY expression.
- registerUdf(String, Class<? extends BeamSqlUdf>) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
-
Register a UDF function which can be used in SQL expression.
- registerUdf(String, SerializableFunction) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
-
- ReifyAsIterable<T> - Class in org.apache.beam.sdk.io.gcp.bigquery
-
This transforms turns a side input into a singleton PCollection that can be used as the main
input for another transform.
- ReifyAsIterable() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.ReifyAsIterable
-
- remerge() - Static method in class org.apache.beam.sdk.transforms.windowing.Window
-
Creates a
Window
PTransform
that does not change assigned
windows, but will cause windows to be merged again as part of the next
GroupByKey
.
- remove(K) - Method in interface org.apache.beam.sdk.state.MapState
-
Remove the mapping for a key from this map if it is present.
- remove(T) - Method in interface org.apache.beam.sdk.state.SetState
-
Removes the specified element from this set if it is present.
- removeTemporaryFiles(Set<ResourceId>) - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-
- rename(List<ResourceIdT>, List<ResourceIdT>) - Method in class org.apache.beam.sdk.io.FileSystem
-
Renames a List
of file-like resources from one location to another.
- rename(List<ResourceId>, List<ResourceId>, MoveOptions...) - Static method in class org.apache.beam.sdk.io.FileSystems
-
Renames a List
of file-like resources from one location to another.
- render() - Method in interface org.apache.beam.runners.spark.aggregators.NamedAggregators.State
-
- renderAll() - Method in class org.apache.beam.runners.spark.aggregators.NamedAggregators
-
- Repeatedly - Class in org.apache.beam.sdk.transforms.windowing
-
A
Trigger
that fires according to its subtrigger forever.
- replaceAll(List<PTransformOverride>) - Method in class org.apache.beam.sdk.Pipeline
-
For internal use only; no backwards-compatibility guarantees.
- replaceAll(String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
-
Returns a
Regex.ReplaceAll
PTransform
that checks if a portion of the line
matches the Regex and replaces all matches with the replacement String.
- replaceAll(Pattern, String) - Static method in class org.apache.beam.sdk.transforms.Regex
-
Returns a
Regex.ReplaceAll
PTransform
that checks if a portion of the line
matches the Regex and replaces all matches with the replacement String.
- ReplaceAll(Pattern, String) - Constructor for class org.apache.beam.sdk.transforms.Regex.ReplaceAll
-
- replaceFirst(String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
-
Returns a
Regex.ReplaceAll
PTransform
that checks if a portion of the line
matches the Regex and replaces the first match with the replacement String.
- replaceFirst(Pattern, String) - Static method in class org.apache.beam.sdk.transforms.Regex
-
Returns a
Regex.ReplaceAll
PTransform
that checks if a portion of the line
matches the Regex and replaces the first match with the replacement String.
- ReplaceFirst(Pattern, String) - Constructor for class org.apache.beam.sdk.transforms.Regex.ReplaceFirst
-
- reportElementSize(long) - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
-
- reportWorkItemStatus(String, ReportWorkItemStatusRequest) - Method in class org.apache.beam.runners.dataflow.DataflowClient
-
Reports the status of the work item for jobId
.
- requiresDeduping() - Method in class org.apache.beam.sdk.io.UnboundedSource
-
Returns whether this source requires explicit deduping.
- reset() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
-
- resetLocal() - Method in class org.apache.beam.runners.flink.metrics.MetricsAccumulator
-
- resetTo(int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
-
Resets the end of the stream to the specified position.
- Reshuffle<K,V> - Class in org.apache.beam.sdk.transforms
-
- Reshuffle.ViaRandomKey<T> - Class in org.apache.beam.sdk.transforms
-
Deprecated.
- ReshuffleTrigger<W extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
-
- ReshuffleTrigger() - Constructor for class org.apache.beam.sdk.transforms.windowing.ReshuffleTrigger
-
Deprecated.
- resolve(String, ResolveOptions) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
-
- resolve(String, ResolveOptions) - Method in interface org.apache.beam.sdk.io.fs.ResourceId
-
Returns a child ResourceId
under this
.
- ResolveOptions - Interface in org.apache.beam.sdk.io.fs
-
- ResolveOptions.StandardResolveOptions - Enum in org.apache.beam.sdk.io.fs
-
Defines the standard resolve options.
- resolveType(Type) - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
Returns a
TypeDescriptor
representing the given
type, with type variables resolved according to the specialization
in this type.
- resourceId() - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata
-
- ResourceId - Interface in org.apache.beam.sdk.io.fs
-
An identifier which represents a file-like resource.
- ResourceIdCoder - Class in org.apache.beam.sdk.io.fs
-
- ResourceIdCoder() - Constructor for class org.apache.beam.sdk.io.fs.ResourceIdCoder
-
- RESTRICTION_CODER - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- RestrictionTracker<RestrictionT> - Interface in org.apache.beam.sdk.transforms.splittabledofn
-
Manages concurrent access to the restriction and keeps track of its claimed part for a
splittable DoFn
.
- resume() - Static method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
-
Indicates that there is more work to be done for the current element.
- resumeDelay() - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
-
- retryTransientErrors() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy
-
Retry all failures except for known persistent errors.
- Reversed() - Constructor for class org.apache.beam.sdk.transforms.Top.Reversed
-
- rightOuterJoin(PCollection<KV<K, V1>>, PCollection<KV<K, V2>>, V1) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
-
Right Outer Join of two collections of KV elements.
- root() - Static method in class org.apache.beam.sdk.transforms.display.DisplayData.Path
-
Path for display data registered by a top-level component.
- run(Pipeline) - Method in class org.apache.beam.runners.apex.ApexRunner
-
- run() - Method in class org.apache.beam.runners.apex.ApexYarnLauncher.ProcessWatcher
-
- run(Pipeline) - Method in class org.apache.beam.runners.apex.TestApexRunner
-
- run(Pipeline) - Method in class org.apache.beam.runners.dataflow.DataflowRunner
-
- run(Pipeline) - Method in class org.apache.beam.runners.dataflow.TestDataflowRunner
-
- run(Pipeline) - Method in class org.apache.beam.runners.direct.DirectRunner
-
- run(Pipeline) - Method in class org.apache.beam.runners.flink.FlinkRunner
-
- run(Pipeline) - Method in class org.apache.beam.runners.flink.TestFlinkRunner
-
- run(Pipeline) - Method in class org.apache.beam.runners.gearpump.GearpumpRunner
-
- run(Pipeline) - Method in class org.apache.beam.runners.gearpump.TestGearpumpRunner
-
- run(JobApi.RunJobRequest, StreamObserver<JobApi.RunJobResponse>) - Method in class org.apache.beam.runners.reference.job.ReferenceRunnerJobService
-
- run(Pipeline) - Method in class org.apache.beam.runners.spark.SparkRunner
-
- run(Pipeline) - Method in class org.apache.beam.runners.spark.SparkRunnerDebugger
-
- run(Pipeline) - Method in class org.apache.beam.runners.spark.TestSparkRunner
-
- run() - Method in class org.apache.beam.sdk.Pipeline
-
- run(PipelineOptions) - Method in class org.apache.beam.sdk.Pipeline
-
- run(Pipeline) - Method in class org.apache.beam.sdk.PipelineRunner
-
Processes the given
Pipeline
, potentially asynchronously, returning a runner-specific
type of result.
- run(Pipeline) - Method in class org.apache.beam.sdk.testing.CrashingRunner
-
- run() - Method in class org.apache.beam.sdk.testing.TestPipeline
-
Runs this
TestPipeline
, unwrapping any
AssertionError
that is raised during
testing.
- run(PipelineOptions) - Method in class org.apache.beam.sdk.testing.TestPipeline
-
- Runner() - Constructor for class org.apache.beam.runners.apex.ApexRunnerRegistrar.Runner
-
- Runner() - Constructor for class org.apache.beam.runners.dataflow.DataflowPipelineRegistrar.Runner
-
- Runner() - Constructor for class org.apache.beam.runners.direct.DirectRegistrar.Runner
-
- Runner() - Constructor for class org.apache.beam.runners.flink.FlinkRunnerRegistrar.Runner
-
- Runner() - Constructor for class org.apache.beam.runners.gearpump.GearpumpRunnerRegistrar.Runner
-
- Runner() - Constructor for class org.apache.beam.runners.spark.SparkRunnerRegistrar.Runner
-
- runWindowFn(WindowFn<T, W>, List<Long>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
-
Runs the
WindowFn
over the provided input, returning a map
of windows to the timestamps in those windows.
- runWindowFnWithValue(WindowFn<T, W>, List<TimestampedValue<T>>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
-
Runs the
WindowFn
over the provided input, returning a map
of windows to the timestamps in those windows.
- Sample - Class in org.apache.beam.sdk.transforms
-
PTransform
s for taking samples of the elements in a
PCollection
, or samples of the values associated with each
key in a PCollection
of KV
s.
- Sample() - Constructor for class org.apache.beam.sdk.transforms.Sample
-
- Sample.FixedSizedSampleFn<T> - Class in org.apache.beam.sdk.transforms
-
CombineFn
that computes a fixed-size sample of a
collection of values.
- satisfies(RelTrait) - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-
- satisfies(SerializableFunction<Iterable<T>, Void>) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
-
Applies the provided checking function (presumably containing assertions) to the
iterable in question.
- satisfies(SerializableFunction<Iterable<T>, Void>) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-
- satisfies(SerializableFunction<T, Void>) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
-
Applies the provided checking function (presumably containing assertions) to the
value in question.
- SCALAR_FIELD_NAME - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- SchemaAndRecord - Class in org.apache.beam.sdk.io.gcp.bigquery
-
A wrapper for a
GenericRecord
and the
TableSchema
representing the schema of the
table (or query) it was generated from.
- SchemaAndRecord(GenericRecord, TableSchema) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.SchemaAndRecord
-
- scopedMetricsContainer(MetricsContainer) - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
-
- SerializableCoder<T extends java.io.Serializable> - Class in org.apache.beam.sdk.coders
-
A
Coder
for Java classes that implement
Serializable
.
- SerializableCoder(Class<T>, TypeDescriptor<T>) - Constructor for class org.apache.beam.sdk.coders.SerializableCoder
-
- SerializableCoder.SerializableCoderProviderRegistrar - Class in org.apache.beam.sdk.coders
-
- SerializableCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.coders.SerializableCoder.SerializableCoderProviderRegistrar
-
- SerializableComparator<T> - Interface in org.apache.beam.sdk.transforms
-
A Comparator
that is also Serializable
.
- SerializableConfiguration - Class in org.apache.beam.sdk.io.hadoop
-
A wrapper to allow Hadoop Configuration
s to be serialized using Java's standard
serialization mechanisms.
- SerializableConfiguration() - Constructor for class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
-
- SerializableConfiguration(Configuration) - Constructor for class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
-
- SerializableFunction<InputT,OutputT> - Interface in org.apache.beam.sdk.transforms
-
A function that computes an output value of type OutputT
from an input value of type
InputT
and is Serializable
.
- SerializableFunctions - Class in org.apache.beam.sdk.transforms
-
- SerializableFunctions() - Constructor for class org.apache.beam.sdk.transforms.SerializableFunctions
-
- SerializableMatcher<T> - Interface in org.apache.beam.sdk.testing
-
A
Matcher
that is also
Serializable
.
- SerializableSplit() - Constructor for class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.SerializableSplit
-
- SerializableSplit(InputSplit) - Constructor for class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.SerializableSplit
-
- serialize(String, Instant) - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantSerializer
-
- serialize(ValueProvider<?>, JsonGenerator, SerializerProvider) - Method in class org.apache.beam.sdk.options.ValueProvider.Serializer
-
- SERIALIZED_FN - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- Serializer() - Constructor for class org.apache.beam.sdk.options.ValueProvider.Serializer
-
- serializeTimers(Collection<TimerInternals.TimerData>, TimerInternals.TimerDataCoder) - Static method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-
- Sessions - Class in org.apache.beam.sdk.transforms.windowing
-
- set(long) - Method in interface org.apache.beam.sdk.metrics.Gauge
-
Set current value for this gauge.
- set() - Static method in class org.apache.beam.sdk.state.StateSpecs
-
- set(Coder<T>) - Static method in class org.apache.beam.sdk.state.StateSpecs
-
- set(Instant) - Method in interface org.apache.beam.sdk.state.Timer
-
Sets or resets the time in the timer's
TimeDomain
at which it should fire.
- set(long...) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
-
- setApiRootUrl(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
- setApplicationName(String) - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
-
- setApplicationName(String) - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
-
- setAppName(String) - Method in interface org.apache.beam.sdk.options.ApplicationNameOptions
-
- setAutoscalingAlgorithm(DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- setBatchIntervalMillis(Long) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- setBlockOnRun(boolean) - Method in interface org.apache.beam.runners.direct.DirectOptions
-
- setCheckpointDir(String) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- setCheckpointDurationMillis(Long) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- setCheckpointingInterval(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setCheckpointingMode(CheckpointingMode) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setCheckpointTimeoutMillis(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setClientContext(ClientContext) - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
-
- setCloningBehavior(DoFnTester.CloningBehavior) - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
- SetCoder<T> - Class in org.apache.beam.sdk.coders
-
- SetCoder(Coder<T>) - Constructor for class org.apache.beam.sdk.coders.SetCoder
-
- setCoder(Coder<T>) - Method in class org.apache.beam.sdk.values.PCollection
-
Sets the
Coder
used by this
PCollection
to encode and decode the
values stored in it.
- setCoderRegistry(CoderRegistry) - Method in class org.apache.beam.sdk.Pipeline
-
- setConfigFile(String) - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
-
- setContentTypeHint(String) - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- setCredentialFactoryClass(Class<? extends CredentialFactory>) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-
- setCurrentContainer(MetricsContainer) - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
-
- setCurrentTransform(TransformHierarchy.Node, Pipeline) - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
-
- setDataflowClient(Dataflow) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
- setDataflowEndpoint(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
- setDataflowJobFile(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
- setDebuggee(Debuggee) - Method in interface org.apache.beam.runners.dataflow.options.CloudDebuggerOptions
-
- setDefaultPipelineOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.io.FileSystems
-
Sets the default configuration in workers.
- setDefaultWorkerLogLevel(DataflowWorkerLoggingOptions.Level) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
-
- setDiskSizeGb(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- setDumpHeapOnOOM(boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
- setEmbeddedCluster(EmbeddedCluster) - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
-
- setEmbeddedExecution(boolean) - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
-
- setEmbeddedExecutionDebugMode(boolean) - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
-
- setEnableCloudDebugger(boolean) - Method in interface org.apache.beam.runners.dataflow.options.CloudDebuggerOptions
-
- setEnableMetrics(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setEnableSparkMetricSinks(Boolean) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- setEnforceEncodability(boolean) - Method in interface org.apache.beam.runners.direct.DirectOptions
-
- setEnforceImmutability(boolean) - Method in interface org.apache.beam.runners.direct.DirectOptions
-
- setExecutionRetryDelay(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setExecutorService(ExecutorService) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-
- setExpectedAssertions(Integer) - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
-
- setExperiments(List<String>) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
- setExternalizedCheckpointsEnabled(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setFilesToStage(List<String>) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- setFilesToStage(List<String>) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setFlinkMaster(String) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setForceStreaming(boolean) - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
-
- setGcpCredential(Credentials) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-
- setGcpTempLocation(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-
- setGcsEndpoint(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-
- setGcsUploadBufferSizeBytes(Integer) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-
- setGcsUploadBufferSizeBytes(Integer) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions.Builder
-
- setGcsUtil(GcsUtil) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-
- setGoogleApiTrace(GoogleApiDebugOptions.GoogleApiTracer) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions
-
- setHdfsConfiguration(List<Configuration>) - Method in interface org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptions
-
- setHooks(DataflowRunnerHooks) - Method in class org.apache.beam.runners.dataflow.DataflowRunner
-
Sets callbacks to invoke during execution see DataflowRunnerHooks
.
- setInput(String) - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- setIsBoundedInternal(PCollection.IsBounded) - Method in class org.apache.beam.sdk.values.PCollection
-
For internal use only; no backwards-compatibility guarantees.
- setIsReadSeekEfficient(boolean) - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
-
- setJobId(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
-
- setJobName(String) - Method in interface org.apache.beam.sdk.options.PipelineOptions
-
- setLabels(Map<String, String>) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-
- setListeners(List<JavaStreamingListener>) - Method in interface org.apache.beam.runners.spark.SparkContextOptions
-
- setMaxBundleSize(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setMaxBundleTimeMills(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setMaxConditionCost(double) - Method in interface org.apache.beam.runners.dataflow.options.CloudDebuggerOptions
-
- setMaxNumWorkers(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- setMaxRecordsPerBatch(Long) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- setMetricsSupported(boolean) - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
-
Called by the run to indicate whether metrics reporting is supported.
- setMimeType(String) - Method in class org.apache.beam.sdk.io.fs.CreateOptions.Builder
-
- setMinimumTextLength(Integer) - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- setMinReadTimeMillis(Long) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- setName(String) - Method in class org.apache.beam.sdk.values.PCollection
-
- setName(String) - Method in class org.apache.beam.sdk.values.PValueBase
-
- setNetwork(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- setNumberOfExecutionRetries(Integer) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setNumberOfWorkerHarnessThreads(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
- setNumWorkers(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- setObjectReuse(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setOnCreateMatcher(SerializableMatcher<PipelineResult>) - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-
- setOnSuccessMatcher(SerializableMatcher<PipelineResult>) - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-
- SetOperatorFilteringDoFn(TupleTag<BeamRecord>, TupleTag<BeamRecord>, BeamSetOperatorRelBase.OpType, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSetOperatorsTransforms.SetOperatorFilteringDoFn
-
- setOptionsId(long) - Method in interface org.apache.beam.sdk.options.PipelineOptions
-
- setOutput(String) - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- setOutputStream(PValue, JavaStream<OutputT>) - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
-
- setOverrideWindmillBinary(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
- setParallelism(Integer) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setParallelism(int) - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
-
- setParameters(T, PreparedStatement) - Method in interface org.apache.beam.sdk.io.jdbc.JdbcIO.PreparedStatementSetter
-
- setParameters(PreparedStatement) - Method in interface org.apache.beam.sdk.io.jdbc.JdbcIO.StatementPreparator
-
- setParDoFusionEnabled(boolean) - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
-
- setParseSynchronously(Boolean) - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- setPathValidator(PathValidator) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-
- setPathValidatorClass(Class<? extends PathValidator>) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-
- setPipelineName(String) - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- setProfilingAgentConfiguration(DataflowProfilingOptions.DataflowProfilingAgentConfiguration) - Method in interface org.apache.beam.runners.dataflow.options.DataflowProfilingOptions
-
- setProject(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-
- setProject(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-
- setProtoTranslation(boolean) - Method in interface org.apache.beam.runners.direct.DirectOptions
-
- setProvidedSparkContext(JavaSparkContext) - Method in interface org.apache.beam.runners.spark.SparkContextOptions
-
- setProviderRuntimeValues(ValueProvider<Map<String, Object>>) - Method in interface org.apache.beam.sdk.testing.TestPipeline.TestValueProviderOptions
-
- setPubsubRootUrl(String) - Method in interface org.apache.beam.sdk.io.gcp.pubsub.PubsubOptions
-
- setQueueMaxPollTime(Long) - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- setQueuePollTime(Long) - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- setReadOutputMetadata(Boolean) - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- setReadTimePercentage(Double) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- setRegion(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-
- setRelative() - Method in interface org.apache.beam.sdk.state.Timer
-
Sets the timer relative to the current time, according to any offset and alignment specified.
- setResourceId(ResourceId) - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
-
- setRetainExternalizedCheckpointsOnCancellation(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setRunMillis(long) - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
-
- setRunner(Class<? extends PipelineRunner<?>>) - Method in interface org.apache.beam.sdk.options.PipelineOptions
-
- setRunnerDeterminedSharding(boolean) - Method in interface org.apache.beam.runners.direct.DirectTestOptions
-
- sets(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-
- setSaveProfilesToGcs(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowProfilingOptions
-
- setSerializers(Map<String, String>) - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
-
- setServiceAccount(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-
- setSideInput(PCollectionView<T>, BoundedWindow, T) - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
- setSideInputs(Map<PCollectionView<?>, Map<BoundedWindow, ?>>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
Registers the tuple of values of the side input
PCollectionView
s to
pass to the
DoFn
under test.
- setSizeBytes(long) - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
-
- setSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction.Builder
-
- setSparkMaster(String) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- setStableUniqueNames(PipelineOptions.CheckEnabled) - Method in interface org.apache.beam.sdk.options.PipelineOptions
-
- setStager(Stager) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
- setStagerClass(Class<? extends Stager>) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
- setStagingLocation(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-
- SetState<T> - Interface in org.apache.beam.sdk.state
-
- setStateBackend(AbstractStateBackend) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-
- setStopPipelineWatermark(Long) - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
-
- setStorageLevel(String) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- setStreaming(boolean) - Method in interface org.apache.beam.sdk.options.StreamingOptions
-
- setSubnetwork(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- setTargetParallelism(int) - Method in interface org.apache.beam.runners.direct.DirectOptions
-
- setTempDatasetId(String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-
- setTemplateLocation(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-
- setTempLocation(String) - Method in interface org.apache.beam.sdk.options.PipelineOptions
-
- setTempRoot(String) - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-
- setTestTimeoutSeconds(Long) - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-
- setTikaConfigPath(String) - Method in interface org.apache.beam.sdk.io.tika.TikaOptions
-
- setTimer(TimerInternals.TimerData) - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-
- setTimer(StateNamespace, String, Instant, TimeDomain) - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-
- setTimer(Instant, TimeDomain) - Method in interface org.apache.beam.sdk.state.Timers
-
Sets a timer to fire when the event time watermark, the current processing time, or
the synchronized processing time watermark surpasses a given timestamp.
- setTimestampBound(TimestampBound) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction.Builder
-
- setTransformNameMapping(Map<String, String>) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
- setTrigger(Trigger) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamAggregationRel
-
- setTupleTracingEnabled(boolean) - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
-
- setTypeDescriptor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.values.PCollection
-
- setup() - Method in class org.apache.beam.runners.gearpump.translators.functions.DoFnFunction
-
- setup() - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSqlFilterFn
-
- setup() - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSqlProjectFn
-
- setUpdate(boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-
- setUsePublicIps(Boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- setUserAgent(String) - Method in interface org.apache.beam.sdk.options.PipelineOptions
-
- setUsesProvidedSparkContext(boolean) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-
- setWindmillServiceEndpoint(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
- setWindmillServicePort(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-
- setWindowedWrites(boolean) - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-
Indicates that the operation will be performing windowed writes.
- setWindowFn(WindowFn) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamAggregationRel
-
- setWindowingStrategyInternal(WindowingStrategy<?, ?>) - Method in class org.apache.beam.sdk.values.PCollection
-
For internal use only; no backwards-compatibility guarantees.
- setWorkerCacheMb(Integer) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
-
- setWorkerDiskType(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- setWorkerHarnessContainerImage(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- setWorkerId(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
-
- setWorkerLogLevelOverrides(DataflowWorkerLoggingOptions.WorkerLogLevelOverrides) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
-
- setWorkerMachineType(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- setWorkerSystemErrMessageLevel(DataflowWorkerLoggingOptions.Level) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
-
- setWorkerSystemOutMessageLevel(DataflowWorkerLoggingOptions.Level) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
-
- setZone(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-
- setZone(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-
- ShardedKey<K> - Class in org.apache.beam.sdk.values
-
A key and a shard number.
- ShardedKeyCoder<KeyT> - Class in org.apache.beam.sdk.coders
-
- ShardedKeyCoder(Coder<KeyT>) - Constructor for class org.apache.beam.sdk.coders.ShardedKeyCoder
-
- ShardNameTemplate - Class in org.apache.beam.sdk.io
-
Standard shard naming templates.
- ShardNameTemplate() - Constructor for class org.apache.beam.sdk.io.ShardNameTemplate
-
- shorts() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-
- shouldDefer(TransformHierarchy.Node) - Method in class org.apache.beam.runners.spark.SparkRunner.Evaluator
-
- shouldResume() - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
-
If false, the
DoFn
promises that there is no more work remaining for the current
element, so the runner should not resume the
DoFn.ProcessElement
call.
- shouldRetry(InsertRetryPolicy.Context) - Method in class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy
-
Return true if this failure should be retried.
- sideInput(PCollectionView<SideInputT>) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
-
Returns the value of a given side input.
- sideInput(PCollectionView<SideInputT>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
-
Returns the value of a given side input.
- sideInput(PCollectionView<T>) - Method in interface org.apache.beam.sdk.state.StateContext
-
Returns the value of the side input for the corresponding state window.
- sideInput(PCollectionView<T>) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.Context
-
Returns the value of the side input for the window corresponding to the
main input's window in which values are being combined.
- sideInput(PCollectionView<T>) - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContext
-
Returns the value of the side input.
- SideInputBroadcast<T> - Class in org.apache.beam.runners.spark.util
-
Broadcast helper for side inputs.
- sideInputJoin(PCollection<KV<BeamRecord, BeamRecord>>, PCollection<KV<BeamRecord, BeamRecord>>, BeamRecord, BeamRecord) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
-
- SideInputJoinDoFn(JoinRelType, BeamRecord, PCollectionView<Map<BeamRecord, Iterable<BeamRecord>>>, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamJoinTransforms.SideInputJoinDoFn
-
- SimpleCombineFn(SerializableFunction<Iterable<V>, V>) - Constructor for class org.apache.beam.sdk.transforms.Combine.SimpleCombineFn
-
Deprecated.
- SimpleFunction<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
-
- SimpleFunction() - Constructor for class org.apache.beam.sdk.transforms.SimpleFunction
-
- SimpleFunction(SerializableFunction<InputT, OutputT>) - Constructor for class org.apache.beam.sdk.transforms.SimpleFunction
-
- SimplePCollectionView() - Constructor for class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
-
For serialization only.
- SimpleQueryTransform(String) - Constructor for class org.apache.beam.sdk.extensions.sql.BeamSql.SimpleQueryTransform
-
- singleOutputOverrideFactory(DataflowPipelineOptions) - Static method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides
-
Returns a
PTransformOverrideFactory
that replaces a single-output
ParDo
with a
composite transform specialized for the
DataflowRunner
.
- singletonView(PCollection<T>, WindowingStrategy<?, W>, boolean, T, Coder<T>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
-
Returns a
PCollectionView<T>
capable of processing elements encoded using the provided
Coder
and windowed using the provided *
WindowingStrategy
.
- sink - Variable in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-
The Sink that this WriteOperation will write to.
- SinkMetrics - Class in org.apache.beam.sdk.metrics
-
Standard Sink Metrics.
- SinkMetrics() - Constructor for class org.apache.beam.sdk.metrics.SinkMetrics
-
- size() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
-
Returns the number of bytes in the backing array that are valid.
- size() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
-
Returns the number of columns for this schema.
- size() - Method in class org.apache.beam.sdk.values.PCollectionList
-
- size() - Method in class org.apache.beam.sdk.values.TupleTagList
-
Returns the number of TupleTags in this TupleTagList.
- sizeBytes() - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata
-
- SlidingWindows - Class in org.apache.beam.sdk.transforms.windowing
-
A
WindowFn
that windows values into possibly overlapping fixed-size
timestamp-based windows.
- smallest(int) - Static method in class org.apache.beam.sdk.transforms.Top
-
Returns a PTransform
that takes an input
PCollection<T>
and returns a PCollection<List<T>>
with a
single element containing the smallest count
elements of the input
PCollection<T>
, in increasing order, sorted according to
their natural order.
- Smallest() - Constructor for class org.apache.beam.sdk.transforms.Top.Smallest
-
Deprecated.
- smallestPerKey(int) - Static method in class org.apache.beam.sdk.transforms.Top
-
Returns a PTransform
that takes an input
PCollection<KV<K, V>>
and returns a
PCollection<KV<K, List<V>>>
that contains an output
element mapping each distinct key in the input
PCollection
to the smallest count
values
associated with that key in the input
PCollection<KV<K, V>>
, in increasing order, sorted
according to their natural order.
- SolrIO - Class in org.apache.beam.sdk.io.solr
-
Transforms for reading and writing data from/to Solr.
- SolrIO.ConnectionConfiguration - Class in org.apache.beam.sdk.io.solr
-
A POJO describing a connection configuration to Solr.
- SolrIO.Read - Class in org.apache.beam.sdk.io.solr
-
- SolrIO.Write - Class in org.apache.beam.sdk.io.solr
-
- sort() - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter
-
- SORT_VALUES - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- SortValues<PrimaryKeyT,SecondaryKeyT,ValueT> - Class in org.apache.beam.sdk.extensions.sorter
-
SortValues<PrimaryKeyT, SecondaryKeyT, ValueT>
takes a PCollection<KV<PrimaryKeyT,
Iterable<KV<SecondaryKeyT, ValueT>>>>
with elements consisting of a primary key and iterables
over <secondary key, value>
pairs, and returns a PCollection<KV<PrimaryKeyT,
Iterable<KV<SecondaryKeyT, ValueT>>>
of the same elements but with values sorted by a secondary
key.
- Source<T> - Class in org.apache.beam.sdk.io
-
Base class for defining input formats and creating a Source
for reading the input.
- Source() - Constructor for class org.apache.beam.sdk.io.Source
-
- Source.Reader<T> - Class in org.apache.beam.sdk.io
-
The interface that readers of custom input sources must implement.
- SOURCE_DOES_NOT_NEED_SPLITTING - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- SOURCE_ESTIMATED_SIZE_BYTES - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- SOURCE_IS_INFINITE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- SOURCE_METADATA - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- SOURCE_SPEC - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- SOURCE_STEP_INPUT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- SourceMetrics - Class in org.apache.beam.sdk.metrics
-
- SourceMetrics() - Constructor for class org.apache.beam.sdk.metrics.SourceMetrics
-
- sourceName() - Method in class org.apache.beam.runners.spark.metrics.AggregatorMetricSource
-
- sourceName() - Method in class org.apache.beam.runners.spark.metrics.CompositeSource
-
- sourceName() - Method in class org.apache.beam.runners.spark.metrics.SparkBeamMetricSource
-
- SourceRDD - Class in org.apache.beam.runners.spark.io
-
Classes implementing Beam
Source
RDD
s.
- SourceRDD() - Constructor for class org.apache.beam.runners.spark.io.SourceRDD
-
- SourceRDD.Bounded<T> - Class in org.apache.beam.runners.spark.io
-
- SourceRDD.Unbounded<T,CheckpointMarkT extends UnboundedSource.CheckpointMark> - Class in org.apache.beam.runners.spark.io
-
- SourceTestUtils - Class in org.apache.beam.sdk.testing
-
Helper functions and test harnesses for checking correctness of
Source
implementations.
- SourceTestUtils() - Constructor for class org.apache.beam.sdk.testing.SourceTestUtils
-
- SourceTestUtils.ExpectedSplitOutcome - Enum in org.apache.beam.sdk.testing
-
- span(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
-
Returns the minimal window that includes both this window and
the given window.
- SpannerAccessor - Class in org.apache.beam.sdk.io.gcp.spanner
-
Manages lifecycle of DatabaseClient
and Spanner
instances.
- SpannerConfig - Class in org.apache.beam.sdk.io.gcp.spanner
-
Configuration for a Cloud Spanner client.
- SpannerConfig() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-
- SpannerConfig.Builder - Class in org.apache.beam.sdk.io.gcp.spanner
-
- SpannerIO - Class in org.apache.beam.sdk.io.gcp.spanner
-
- SpannerIO.CreateTransaction - Class in org.apache.beam.sdk.io.gcp.spanner
-
- SpannerIO.CreateTransaction.Builder - Class in org.apache.beam.sdk.io.gcp.spanner
-
- SpannerIO.Read - Class in org.apache.beam.sdk.io.gcp.spanner
-
- SpannerIO.ReadAll - Class in org.apache.beam.sdk.io.gcp.spanner
-
- SpannerIO.Write - Class in org.apache.beam.sdk.io.gcp.spanner
-
A
PTransform
that writes
Mutation
objects to Google Cloud Spanner.
- SpannerIO.WriteGrouped - Class in org.apache.beam.sdk.io.gcp.spanner
-
- SparkBeamMetricSource - Class in org.apache.beam.runners.spark.metrics
-
A Spark
Source
that is tailored to expose a
SparkBeamMetric
,
wrapping an underlying
MetricResults
instance.
- SparkBeamMetricSource(String) - Constructor for class org.apache.beam.runners.spark.metrics.SparkBeamMetricSource
-
- SparkContextOptions - Interface in org.apache.beam.runners.spark
-
A custom
PipelineOptions
to work with properties related to
JavaSparkContext
.
- SparkContextOptions.EmptyListenersList - Class in org.apache.beam.runners.spark
-
Returns an empty list, top avoid handling null.
- SparkGroupAlsoByWindowViaWindowSet - Class in org.apache.beam.runners.spark.stateful
-
An implementation of GroupByKeyViaGroupByKeyOnly.GroupAlsoByWindow
logic for grouping by windows and controlling
trigger firings and pane accumulation.
- SparkGroupAlsoByWindowViaWindowSet() - Constructor for class org.apache.beam.runners.spark.stateful.SparkGroupAlsoByWindowViaWindowSet
-
- SparkNativePipelineVisitor - Class in org.apache.beam.runners.spark
-
Pipeline visitor for translating a Beam pipeline into equivalent Spark operations.
- SparkPipelineOptions - Interface in org.apache.beam.runners.spark
-
Spark runner
PipelineOptions
handles Spark execution-related configurations,
such as the master address, batch-interval, and other user-related knobs.
- SparkPipelineOptions.TmpCheckpointDirFactory - Class in org.apache.beam.runners.spark
-
Returns the default checkpoint directory of /tmp/${job.name}.
- SparkPipelineResult - Class in org.apache.beam.runners.spark
-
Represents a Spark pipeline execution result.
- SparkRunner - Class in org.apache.beam.runners.spark
-
The SparkRunner translate operations defined on a pipeline to a representation executable by
Spark, and then submitting the job to Spark to be executed.
- SparkRunner.Evaluator - Class in org.apache.beam.runners.spark
-
Evaluator on the pipeline.
- SparkRunnerDebugger - Class in org.apache.beam.runners.spark
-
Pipeline runner which translates a Beam pipeline into equivalent Spark operations, without
running them.
- SparkRunnerDebugger.DebugSparkPipelineResult - Class in org.apache.beam.runners.spark
-
- SparkRunnerRegistrar - Class in org.apache.beam.runners.spark
-
- SparkRunnerRegistrar.Options - Class in org.apache.beam.runners.spark
-
- SparkRunnerRegistrar.Runner - Class in org.apache.beam.runners.spark
-
- SparkSideInputReader - Class in org.apache.beam.runners.spark.util
-
A SideInputReader
for thw SparkRunner.
- SparkSideInputReader(Map<TupleTag<?>, KV<WindowingStrategy<?, ?>, SideInputBroadcast<?>>>) - Constructor for class org.apache.beam.runners.spark.util.SparkSideInputReader
-
- SparkTimerInternals - Class in org.apache.beam.runners.spark.stateful
-
An implementation of TimerInternals
for the SparkRunner.
- SparkUnboundedSource - Class in org.apache.beam.runners.spark.io
-
- SparkUnboundedSource() - Constructor for class org.apache.beam.runners.spark.io.SparkUnboundedSource
-
- SparkUnboundedSource.Metadata - Class in org.apache.beam.runners.spark.io
-
A metadata holder for an input stream partition.
- SparkWatermarks(Instant, Instant, Instant) - Constructor for class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.SparkWatermarks
-
- split(int, PipelineOptions) - Method in class org.apache.beam.runners.gearpump.translators.io.ValuesSource
-
- split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.BoundedSource
-
Splits the source into bundles of approximately desiredBundleSizeBytes
.
- split(CassandraIO.Read<T>, long) - Method in interface org.apache.beam.sdk.io.cassandra.CassandraService
-
Split a table read into several sources.
- split(CassandraIO.Read<T>, long) - Method in class org.apache.beam.sdk.io.cassandra.CassandraServiceImpl
-
- split(CassandraIO.Read<T>, long, long) - Method in class org.apache.beam.sdk.io.cassandra.CassandraServiceImpl
-
Compute the number of splits based on the estimated size and the desired bundle size, and
create several sources.
- split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
-
- split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
-
- split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.HadoopInputFormatBoundedSource
-
- split(int, PipelineOptions) - Method in class org.apache.beam.sdk.io.jms.JmsIO.UnboundedJmsSource
-
- split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
-
- split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-
- split(int) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
-
Returns a list of up to
numSplits + 1
ByteKeys
in ascending order,
where the keys have been interpolated to form roughly equal sub-ranges of this
ByteKeyRange
, assuming a uniform distribution of keys within this range.
- split(long, long) - Method in class org.apache.beam.sdk.io.range.OffsetRange
-
- split(int, PipelineOptions) - Method in class org.apache.beam.sdk.io.UnboundedSource
-
Returns a list of UnboundedSource
objects representing the instances of this source
that should be used when executing the workflow.
- split(String) - Static method in class org.apache.beam.sdk.transforms.Regex
-
Returns a
Regex.Split
PTransform
that splits a string on the regular expression
and then outputs each item.
- split(Pattern) - Static method in class org.apache.beam.sdk.transforms.Regex
-
Returns a
Regex.Split
PTransform
that splits a string on the regular expression
and then outputs each item.
- split(String, boolean) - Static method in class org.apache.beam.sdk.transforms.Regex
-
Returns a
Regex.Split
PTransform
that splits a string on the regular expression
and then outputs each item.
- split(Pattern, boolean) - Static method in class org.apache.beam.sdk.transforms.Regex
-
Returns a
Regex.Split
PTransform
that splits a string on the regular expression
and then outputs each item.
- Split(Pattern, boolean) - Constructor for class org.apache.beam.sdk.transforms.Regex.Split
-
- SPLIT_POINTS_UNKNOWN - Static variable in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
-
- splitAtFraction(double) - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
-
Tells the reader to narrow the range of the input it's going to read and give up
the remainder, so that the new range would contain approximately the given
fraction of the amount of data in the current range.
- splitAtFraction(double) - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
- stageFiles() - Method in class org.apache.beam.runners.dataflow.util.GcsStager
-
- stageFiles() - Method in interface org.apache.beam.runners.dataflow.util.Stager
-
- Stager - Interface in org.apache.beam.runners.dataflow.util
-
Interface for staging files needed for running a Dataflow pipeline.
- StagerFactory() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions.StagerFactory
-
- StagingLocationFactory() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowPipelineOptions.StagingLocationFactory
-
- StandardCreateOptions() - Constructor for class org.apache.beam.sdk.io.fs.CreateOptions.StandardCreateOptions
-
- start() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-
- start() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
- start() - Method in class org.apache.beam.sdk.io.Source.Reader
-
Initializes the reader and advances the reader to the first record.
- start() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
-
Initializes the reader and advances the reader to the first record.
- start() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
-
Returns the start of this window, inclusive.
- startBundle() - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
-
- startBundle(DoFn<T, Void>.StartBundleContext) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write.PubsubBoundedWriter
-
- startBundle() - Method in class org.apache.beam.sdk.transforms.DoFnTester
-
- StartBundleContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.StartBundleContext
-
- startImpl() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
-
- startImpl() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-
- startReading(ReadableByteChannel) - Method in class org.apache.beam.sdk.io.AvroSource.AvroReader
-
- startReading(ReadableByteChannel) - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-
Creates a decompressing channel from the input channel and passes it to its delegate reader's
FileBasedReader#startReading(ReadableByteChannel)
.
- startReading(ReadableByteChannel) - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
-
Performs any initialization of the subclass of FileBasedReader
that involves IO
operations.
- state - Variable in class org.apache.beam.runners.spark.SparkPipelineResult
-
- State - Interface in org.apache.beam.sdk.state
-
- StateBinder - Interface in org.apache.beam.sdk.state
-
For internal use only; no backwards-compatibility guarantees.
- StateContext<W extends BoundedWindow> - Interface in org.apache.beam.sdk.state
-
For internal use only; no backwards-compatibility guarantees.
- StateContexts - Class in org.apache.beam.sdk.state
-
For internal use only; no backwards-compatibility guarantees.
- StateContexts() - Constructor for class org.apache.beam.sdk.state.StateContexts
-
- stateInternals() - Method in class org.apache.beam.runners.gearpump.translators.utils.NoOpStepContext
-
- StateSpec<StateT extends State> - Interface in org.apache.beam.sdk.state
-
A specification of a persistent state cell.
- StateSpec.Cases<ResultT> - Interface in org.apache.beam.sdk.state
-
Cases for doing a "switch" on the type of
StateSpec
.
- StateSpec.Cases.WithDefault<ResultT> - Class in org.apache.beam.sdk.state
-
A base class for a visitor with a default method for cases it is not interested in.
- StateSpecFunctions - Class in org.apache.beam.runners.spark.stateful
-
A class containing StateSpec
mappingFunctions.
- StateSpecFunctions() - Constructor for class org.apache.beam.runners.spark.stateful.StateSpecFunctions
-
- StateSpecs - Class in org.apache.beam.sdk.state
-
- status() - Method in class org.apache.beam.sdk.io.fs.MatchResult
-
- STATUS_BACKOFF_FACTORY - Static variable in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-
- step() - Method in interface org.apache.beam.sdk.metrics.MetricResult
-
Return the step context to which this metric result applies.
- steps() - Method in class org.apache.beam.sdk.metrics.MetricsFilter
-
- stop() - Method in class org.apache.beam.runners.spark.SparkPipelineResult
-
- stop() - Method in class org.apache.beam.runners.spark.SparkRunnerDebugger.DebugSparkPipelineResult
-
- stop() - Static method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
-
Indicates that there is no more work to be done for the current element.
- StreamingInserts<DestinationT> - Class in org.apache.beam.sdk.io.gcp.bigquery
-
PTransform that performs streaming BigQuery write.
- StreamingInserts(BigQueryIO.Write.CreateDisposition, DynamicDestinations<?, DestinationT>) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StreamingInserts
-
Constructor.
- StreamingIT - Interface in org.apache.beam.sdk.testing
-
- StreamingOptions - Interface in org.apache.beam.sdk.options
-
Options used to configure streaming.
- StreamingWriteTables - Class in org.apache.beam.sdk.io.gcp.bigquery
-
- StreamingWriteTables() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StreamingWriteTables
-
- StringDelegateCoder<T> - Class in org.apache.beam.sdk.coders
-
A
Coder
that wraps a
Coder<String>
and encodes/decodes values via string representations.
- StringDelegateCoder(Class<T>, TypeDescriptor<T>) - Constructor for class org.apache.beam.sdk.coders.StringDelegateCoder
-
- strings() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-
- StringUtf8Coder - Class in org.apache.beam.sdk.coders
-
A
Coder
that encodes
Strings
in UTF-8 encoding.
- StripIdsDoFn() - Constructor for class org.apache.beam.sdk.values.ValueWithRecordId.StripIdsDoFn
-
- stripPartitionDecorator(String) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
-
Strip off any partition decorator information from a tablespec.
- Structs - Class in org.apache.beam.runners.dataflow.util
-
A collection of static methods for manipulating datastructure representations transferred via the
Dataflow API.
- StructuralByteArray - Class in org.apache.beam.sdk.coders
-
A wrapper around a byte[] that uses structural, value-based
equality rather than byte[]'s normal object identity.
- StructuralByteArray(byte[]) - Constructor for class org.apache.beam.sdk.coders.StructuralByteArray
-
- structuralValue(byte[]) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
-
Returns an object with an Object.equals()
method that represents structural equality on
the argument.
- structuralValue(T) - Method in class org.apache.beam.sdk.coders.Coder
-
Returns an object with an Object.equals()
method that represents structural equality on
the argument.
- structuralValue(T) - Method in class org.apache.beam.sdk.coders.DelegateCoder
-
Returns an object with an Object.equals()
method that represents structural equality on
the argument.
- structuralValue(KV<K, V>) - Method in class org.apache.beam.sdk.coders.KvCoder
-
- structuralValue(T) - Method in class org.apache.beam.sdk.coders.NullableCoder
-
- structuralValue(T) - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
-
- structuralValue(KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
-
- structuralValueConsistentWithEquals(Coder<T>, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that for the given Coder<T>
and values of
type T
, the structural values are equal if and only if the
encoded bytes are equal.
- structuralValueConsistentWithEqualsInContext(Coder<T>, Coder.Context, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that for the given Coder<T>
, Coder.Context
, and
values of type T
, the structural values are equal if and only if the
encoded bytes are equal, in any Coder.Context
.
- structuralValueDecodeEncodeEqual(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that for the given Coder<T>
and value of type T
,
the structural value is equal to the structural value yield by encoding
and decoding the original value.
- structuralValueDecodeEncodeEqualInContext(Coder<T>, Coder.Context, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-
Verifies that for the given Coder<T>
, Coder.Context
,
and value of type T
, the structural value is equal to the
structural value yield by encoding and decoding the original value,
in any Coder.Context
.
- StructuredCoder<T> - Class in org.apache.beam.sdk.coders
-
- StructuredCoder() - Constructor for class org.apache.beam.sdk.coders.StructuredCoder
-
- subTriggers - Variable in class org.apache.beam.sdk.transforms.windowing.Trigger
-
- subTriggers() - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
-
- success() - Static method in class org.apache.beam.sdk.testing.SuccessOrFailure
-
- SUCCESS_COUNTER - Static variable in class org.apache.beam.sdk.testing.PAssert
-
- SuccessOrFailure - Class in org.apache.beam.sdk.testing
-
- sum() - Method in class org.apache.beam.sdk.metrics.DistributionResult
-
- Sum - Class in org.apache.beam.sdk.transforms
-
PTransform
s for computing the sum of the elements in a
PCollection
, or the sum of the values associated with
each key in a PCollection
of KV
s.
- v1() - Static method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreIO
-
Returns a
DatastoreV1
that provides an API for accessing Cloud Datastore through v1
version of Datastore Client library.
- validate() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
-
- validate() - Method in class org.apache.beam.sdk.io.AvroSource
-
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
-
- validate() - Method in class org.apache.beam.sdk.io.CompressedSource
-
Validates that the delegate source is a valid source and that the channel factory is not null.
- validate() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
-
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSink
-
- validate() - Method in class org.apache.beam.sdk.io.FileBasedSource
-
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
-
- validate() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-
- validate() - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.HadoopInputFormatBoundedSource
-
- validate(T) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-
- validate() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-
- validate() - Method in class org.apache.beam.sdk.io.Source
-
Checks that this source is valid, before it can be used in a pipeline.
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.WriteFiles
-
- validate(Class<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.options.PipelineOptionsValidator
-
Validates that the passed
PipelineOptions
conforms to all the validation criteria from
the passed in interface.
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.transforms.PTransform
-
Called before running the Pipeline to verify this transform is fully and correctly
specified.
- validateCli(Class<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.options.PipelineOptionsValidator
-
Validates that the passed
PipelineOptions
from command line interface (CLI)
conforms to all the validation criteria from the passed in interface.
- validateGetOutputTimestamp(WindowFn<T, W>, long) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
-
Assigns the given
timestamp
to windows using the specified
windowFn
, and
verifies that result of
windowFn.getOutputTime
for later windows
(as defined by
maxTimestamp
won't prevent the watermark from passing the end of earlier
windows.
- validateGetOutputTimestamps(WindowFn<T, W>, TimestampCombiner, List<List<Long>>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
-
Verifies that later-ending merged windows from any of the timestamps hold up output of
earlier-ending windows, using the provided
WindowFn
and
TimestampCombiner
.
- validateGetOutputTimestampsWithValue(WindowFn<T, W>, TimestampCombiner, List<List<TimestampedValue<T>>>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
-
Verifies that later-ending merged windows from any of the timestampValues hold up output of
earlier-ending windows, using the provided
WindowFn
and
TimestampCombiner
.
- validateGetOutputTimestampWithValue(WindowFn<T, W>, TimestampedValue<T>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
-
Assigns the given
timestampedValue
to windows using the specified
windowFn
, and
verifies that result of
windowFn.getOutputTime
for later windows
(as defined by
maxTimestamp
won't prevent the watermark from passing the end of earlier
windows.
- validateInputFilePatternSupported(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
-
Validates the the input GCS path is accessible and that the path
is well formed.
- validateInputFilePatternSupported(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
-
- validateInputFilePatternSupported(String) - Method in interface org.apache.beam.sdk.extensions.gcp.storage.PathValidator
-
Validate that a file pattern is conforming.
- validateNonInterferingOutputTimes(WindowFn<T, W>, long) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
-
Assigns the given timestamp
to windows using the specified windowFn
, and
verifies that result of windowFn.getOutputTimestamp
for each window is within the
proper bound.
- validateNonInterferingOutputTimesWithValue(WindowFn<T, W>, TimestampedValue<T>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
-
Assigns the given timestampedValue
to windows using the specified windowFn
, and
verifies that result of windowFn.getOutputTimestamp
for each window is within the
proper bound.
- validateOutputFilePrefixSupported(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
-
Validates the the output GCS path is accessible and that the path
is well formed.
- validateOutputFilePrefixSupported(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
-
- validateOutputFilePrefixSupported(String) - Method in interface org.apache.beam.sdk.extensions.gcp.storage.PathValidator
-
Validate that an output file prefix is conforming.
- validateOutputResourceSupported(ResourceId) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
-
- validateOutputResourceSupported(ResourceId) - Method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
-
- validateOutputResourceSupported(ResourceId) - Method in interface org.apache.beam.sdk.extensions.gcp.storage.PathValidator
-
Validates that an output path is conforming.
- ValidatesRunner - Interface in org.apache.beam.sdk.testing
-
Category tag for tests which validate that a Beam runner is correctly implemented.
- validateValueType(int, Object) - Method in class org.apache.beam.sdk.extensions.sql.BeamRecordSqlType
-
- validateValueType(int, Object) - Method in class org.apache.beam.sdk.values.BeamRecordType
-
Validate input fieldValue for a field.
- Validation - Annotation Type in org.apache.beam.sdk.options
-
- Validation.Required - Annotation Type in org.apache.beam.sdk.options
-
This criteria specifies that the value must be not null.
- VALUE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-
- value() - Method in class org.apache.beam.sdk.metrics.GaugeResult.EmptyGaugeResult
-
- value() - Method in class org.apache.beam.sdk.metrics.GaugeResult
-
- value() - Static method in class org.apache.beam.sdk.state.StateSpecs
-
Create a
StateSpec
for a single value of type
T
.
- value(Coder<T>) - Static method in class org.apache.beam.sdk.state.StateSpecs
-
- ValueInSingleWindow<T> - Class in org.apache.beam.sdk.values
-
An immutable tuple of value, timestamp, window, and pane.
- ValueInSingleWindow() - Constructor for class org.apache.beam.sdk.values.ValueInSingleWindow
-
- ValueInSingleWindow.Coder<T> - Class in org.apache.beam.sdk.values
-
- valueOf(String) - Static method in enum org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.Level
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.annotations.Experimental.Kind
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.coders.CannotProvideCoderException.ReasonCode
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamSetOperatorRelBase.OpType
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sql.impl.schema.BeamIOType
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.CompressedSource.CompressionMode
-
Deprecated.
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.Compression
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
-
Deprecated.
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.FileBasedSource.Mode
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.fs.EmptyMatchTreatment
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.fs.MatchResult.Status
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.fs.MoveOptions.StandardMoveOptions
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.fs.ResolveOptions.StandardResolveOptions
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.CreateDisposition
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.Method
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.WriteDisposition
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.TextIO.CompressionType
-
Deprecated.
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.TFRecordIO.CompressionType
-
Deprecated.
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.xml.XmlIO.Read.CompressionType
-
Deprecated.
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.options.PipelineOptions.CheckEnabled
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.Pipeline.PipelineVisitor.CompositeBehavior
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.PipelineResult.State
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.state.TimeDomain
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.testing.SourceTestUtils.ExpectedSplitOutcome
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.testing.TestStream.EventType
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.display.DisplayData.Type
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.DoFnTester.CloningBehavior
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.windowing.PaneInfo.Timing
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.windowing.Window.ClosingBehavior
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.windowing.Window.OnTimeBehavior
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.values.PCollection.IsBounded
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.beam.sdk.values.WindowingStrategy.AccumulationMode
-
Returns the enum constant of this type with the specified name.
- ValueProvider<T> - Interface in org.apache.beam.sdk.options
-
A
ValueProvider
abstracts the notion of fetching a value that may or may not be currently
available.
- ValueProvider.Deserializer - Class in org.apache.beam.sdk.options
-
For internal use only; no backwards compatibility guarantees.
- ValueProvider.NestedValueProvider<T,X> - Class in org.apache.beam.sdk.options
-
- ValueProvider.RuntimeValueProvider<T> - Class in org.apache.beam.sdk.options
-
- ValueProvider.Serializer - Class in org.apache.beam.sdk.options
-
For internal use only; no backwards compatibility guarantees.
- ValueProvider.StaticValueProvider<T> - Class in org.apache.beam.sdk.options
-
- ValueProviders - Class in org.apache.beam.sdk.options
-
- values() - Static method in enum org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.Level
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.annotations.Experimental.Kind
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.coders.CannotProvideCoderException.ReasonCode
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamSetOperatorRelBase.OpType
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.extensions.sql.impl.schema.BeamIOType
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.io.CompressedSource.CompressionMode
-
Deprecated.
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.io.Compression
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
-
Deprecated.
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.io.FileBasedSource.Mode
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.io.fs.EmptyMatchTreatment
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.io.fs.MatchResult.Status
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.io.fs.MoveOptions.StandardMoveOptions
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.io.fs.ResolveOptions.StandardResolveOptions
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.CreateDisposition
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.Method
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.WriteDisposition
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
-
Writes just the values to Kafka.
- values() - Static method in enum org.apache.beam.sdk.io.TextIO.CompressionType
-
Deprecated.
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.io.TFRecordIO.CompressionType
-
Deprecated.
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.io.xml.XmlIO.Read.CompressionType
-
Deprecated.
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.options.PipelineOptions.CheckEnabled
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.Pipeline.PipelineVisitor.CompositeBehavior
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.PipelineResult.State
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Method in interface org.apache.beam.sdk.state.MapState
-
Returns an Iterable
over the values contained in this map.
- values() - Static method in enum org.apache.beam.sdk.state.TimeDomain
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.testing.SourceTestUtils.ExpectedSplitOutcome
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.testing.TestStream.EventType
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.transforms.display.DisplayData.Type
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.transforms.DoFnTester.CloningBehavior
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- Values<V> - Class in org.apache.beam.sdk.transforms
-
Values<V>
takes a PCollection
of KV<K, V>
s and
returns a PCollection<V>
of the values.
- values() - Static method in enum org.apache.beam.sdk.transforms.windowing.PaneInfo.Timing
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.transforms.windowing.Window.ClosingBehavior
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.transforms.windowing.Window.OnTimeBehavior
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.values.PCollection.IsBounded
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.beam.sdk.values.WindowingStrategy.AccumulationMode
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- ValuesSource<T> - Class in org.apache.beam.runners.gearpump.translators.io
-
unbounded source that reads from a Java Iterable
.
- ValuesSource(Iterable<T>, Coder<T>) - Constructor for class org.apache.beam.runners.gearpump.translators.io.ValuesSource
-
- ValueState<T> - Interface in org.apache.beam.sdk.state
-
- ValueWithRecordId<ValueT> - Class in org.apache.beam.sdk.values
-
For internal use only; no backwards compatibility guarantees.
- ValueWithRecordId(ValueT, byte[]) - Constructor for class org.apache.beam.sdk.values.ValueWithRecordId
-
- ValueWithRecordId.StripIdsDoFn<T> - Class in org.apache.beam.sdk.values
-
DoFn
to turn a
ValueWithRecordId<T>
back to the value
T
.
- ValueWithRecordId.ValueWithRecordIdCoder<ValueT> - Class in org.apache.beam.sdk.values
-
A
Coder
for
ValueWithRecordId
, using a wrapped value
Coder
.
- ValueWithRecordIdCoder(Coder<ValueT>) - Constructor for class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
-
- VarIntCoder - Class in org.apache.beam.sdk.coders
-
A
Coder
that encodes
Integers
using between 1 and 5 bytes.
- VarLongCoder - Class in org.apache.beam.sdk.coders
-
A
Coder
that encodes
Longs
using between 1 and 10 bytes.
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
-
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
-
- verifyDeterministic() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.AtomicCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.AvroCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BeamRecordCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BitSetCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.ByteCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.Coder
-
- verifyDeterministic(Coder<?>, String, Iterable<Coder<?>>) - Static method in class org.apache.beam.sdk.coders.Coder
-
Verifies all of the provided coders are deterministic.
- verifyDeterministic(Coder<?>, String, Coder<?>...) - Static method in class org.apache.beam.sdk.coders.Coder
-
Verifies all of the provided coders are deterministic.
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.CustomCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.DelegateCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.DoubleCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.DurationCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.InstantCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.KvCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
-
LengthPrefixCoder
is deterministic if the nested Coder
is.
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.ListCoder
-
List sizes are always known, so ListIterable may be deterministic while
the general IterableLikeCoder is not.
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.MapCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.NullableCoder
-
NullableCoder
is deterministic if the nested Coder
is.
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.SerializableCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.SetCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.ShardedKeyCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.VarIntCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.VarLongCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.VoidCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.BooleanCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.DateCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.DoubleCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.FloatCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.ShortCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlRecordHelper.TimeCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV2
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.hadoop.WritableCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo.PaneInfoCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
-
- verifyDeterministic() - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
-
- verifyPAssertsSucceeded(Pipeline, PipelineResult) - Static method in class org.apache.beam.sdk.testing.TestPipeline
-
Verifies all {
PAsserts
} in the pipeline have been executed and were successful.
- verifyPath(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
-
- verifyPath(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
-
- verifyPath(String) - Method in interface org.apache.beam.sdk.extensions.gcp.storage.PathValidator
-
Validate that a path is a valid path and that the path
is accessible.
- via(SimpleFunction<? super InputT, ? extends Iterable<OutputT>>) - Static method in class org.apache.beam.sdk.transforms.FlatMapElements
-
For a
SimpleFunction<InputT, ? extends Iterable<OutputT>>
fn
,
return a
PTransform
that applies
fn
to every element of the input
PCollection<InputT>
and outputs all of the elements to the output
PCollection<OutputT>
.
- via(SerializableFunction<NewInputT, ? extends Iterable<OutputT>>) - Method in class org.apache.beam.sdk.transforms.FlatMapElements
-
For a
SerializableFunction<InputT, ? extends Iterable<OutputT>>
fn
,
returns a
PTransform
that applies
fn
to every element of the input
PCollection<InputT>
and outputs all of the elements to the output
PCollection<OutputT>
.
- via(SimpleFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.MapElements
-
For a SimpleFunction<InputT, OutputT>
fn
, returns a PTransform
that
takes an input PCollection<InputT>
and returns a PCollection<OutputT>
containing fn.apply(v)
for every element v
in the input.
- via(SerializableFunction<NewInputT, OutputT>) - Method in class org.apache.beam.sdk.transforms.MapElements
-
For a SerializableFunction<InputT, OutputT>
fn
and output type descriptor,
returns a PTransform
that takes an input PCollection<InputT>
and returns a
PCollection<OutputT>
containing fn.apply(v)
for every element v
in the
input.
- viaRandomKey() - Static method in class org.apache.beam.sdk.transforms.Reshuffle
-
Deprecated.
Encapsulates the sequence "pair input with unique key, apply
Reshuffle.of()
, drop the key" commonly used to break fusion.
- View - Class in org.apache.beam.sdk.transforms
-
- View.AsIterable<T> - Class in org.apache.beam.sdk.transforms
-
For internal use only; no backwards-compatibility guarantees.
- View.AsList<T> - Class in org.apache.beam.sdk.transforms
-
For internal use only; no backwards-compatibility guarantees.
- View.AsMap<K,V> - Class in org.apache.beam.sdk.transforms
-
For internal use only; no backwards-compatibility guarantees.
- View.AsMultimap<K,V> - Class in org.apache.beam.sdk.transforms
-
For internal use only; no backwards-compatibility guarantees.
- View.AsSingleton<T> - Class in org.apache.beam.sdk.transforms
-
For internal use only; no backwards-compatibility guarantees.
- View.CreatePCollectionView<ElemT,ViewT> - Class in org.apache.beam.sdk.transforms
-
For internal use only; no backwards-compatibility guarantees.
- ViewFn<PrimitiveViewT,ViewT> - Class in org.apache.beam.sdk.transforms
-
For internal use only; no backwards-compatibility guarantees.
- ViewFn() - Constructor for class org.apache.beam.sdk.transforms.ViewFn
-
- visitPrimitiveTransform(TransformHierarchy.Node) - Method in class org.apache.beam.runners.gearpump.translators.GearpumpPipelineTranslator
-
- visitPrimitiveTransform(TransformHierarchy.Node) - Method in class org.apache.beam.runners.spark.SparkRunner.Evaluator
-
- visitPrimitiveTransform(TransformHierarchy.Node) - Method in class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
-
- visitPrimitiveTransform(TransformHierarchy.Node) - Method in interface org.apache.beam.sdk.Pipeline.PipelineVisitor
-
Called for each primitive transform after all of its topological predecessors
and inputs have been visited.
- visitValue(PValue, TransformHierarchy.Node) - Method in class org.apache.beam.runners.gearpump.translators.GearpumpPipelineTranslator
-
- visitValue(PValue, TransformHierarchy.Node) - Method in class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
-
- visitValue(PValue, TransformHierarchy.Node) - Method in interface org.apache.beam.sdk.Pipeline.PipelineVisitor
-
Called for each value after the transform that produced the value has been
visited.
- VoidCoder - Class in org.apache.beam.sdk.coders
-
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.apex.ApexRunnerResult
-
- waitUntilFinish() - Method in class org.apache.beam.runners.apex.ApexRunnerResult
-
- waitUntilFinish() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-
- waitUntilFinish(Duration, MonitoringUtil.JobMessagesHandler) - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-
Waits until the pipeline finishes and returns the final status.
- waitUntilFinish() - Method in class org.apache.beam.runners.direct.DirectRunner.DirectPipelineResult
-
Waits until the pipeline finishes and returns the final status.
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.direct.DirectRunner.DirectPipelineResult
-
Waits until the pipeline finishes and returns the final status.
- waitUntilFinish() - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
-
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
-
- waitUntilFinish() - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
-
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
-
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.gearpump.GearpumpPipelineResult
-
- waitUntilFinish() - Method in class org.apache.beam.runners.gearpump.GearpumpPipelineResult
-
- waitUntilFinish() - Method in class org.apache.beam.runners.spark.SparkPipelineResult
-
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.spark.SparkPipelineResult
-
- waitUntilFinish(Duration) - Method in interface org.apache.beam.sdk.PipelineResult
-
Waits until the pipeline finishes and returns the final status.
- waitUntilFinish() - Method in interface org.apache.beam.sdk.PipelineResult
-
Waits until the pipeline finishes and returns the final status.
- Watch - Class in org.apache.beam.sdk.transforms
-
Given a "poll function" that produces a potentially growing set of outputs for an input, this
transform simultaneously continuously watches the growth of output sets of all inputs, until a
per-input termination condition is reached.
- Watch() - Constructor for class org.apache.beam.sdk.transforms.Watch
-
- Watch.Growth<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
-
- Watch.Growth.PollFn<InputT,OutputT> - Interface in org.apache.beam.sdk.transforms
-
- Watch.Growth.PollResult<OutputT> - Class in org.apache.beam.sdk.transforms
-
- Watch.Growth.TerminationCondition<InputT,StateT> - Interface in org.apache.beam.sdk.transforms
-
A strategy for determining whether it is time to stop polling the current input regardless of
whether its output is complete or not.
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
-
Like Read#watchForNewFiles
.
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.AvroIO.ParseAll
-
Like Read#watchForNewFiles
.
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.AvroIO.Read
-
Continuously watches for new files matching the filepattern, polling it at the given
interval, until the given termination condition is reached.
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.AvroIO.ReadAll
-
Like Read#watchForNewFiles
.
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.TextIO.Read
-
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
-
Same as Read#watchForNewFiles(Duration, TerminationCondition)
.
- WatermarkAdvancingStreamingListener() - Constructor for class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.WatermarkAdvancingStreamingListener
-
- WatermarkEvent() - Constructor for class org.apache.beam.sdk.testing.TestStream.WatermarkEvent
-
- WatermarkHoldState - Interface in org.apache.beam.sdk.state
-
For internal use only; no backwards-compatibility guarantees.
- watermarkStateInternal(TimestampCombiner) - Static method in class org.apache.beam.sdk.state.StateSpecs
-
For internal use only; no backwards-compatibility guarantees.
- weeks(int, int) - Static method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows
-
Returns a
WindowFn
that windows elements into periods measured by weeks.
- where(TypeParameter<X>, TypeDescriptor<X>) - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
Returns a new TypeDescriptor
where the type variable represented by typeParameter
are substituted by typeDescriptor
.
- where(Type, Type) - Method in class org.apache.beam.sdk.values.TypeDescriptor
-
- window() - Method in interface org.apache.beam.sdk.state.StateContext
-
Returns the window corresponding to the state.
- window() - Method in class org.apache.beam.sdk.transforms.DoFn.OnTimerContext
-
Returns the window in which the timer is firing.
- Window<T> - Class in org.apache.beam.sdk.transforms.windowing
-
- Window() - Constructor for class org.apache.beam.sdk.transforms.windowing.Window
-
- window() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn.AssignContext
-
Returns the window of the current element prior to this
WindowFn
being called.
- Window.Assign<T> - Class in org.apache.beam.sdk.transforms.windowing
-
- Window.ClosingBehavior - Enum in org.apache.beam.sdk.transforms.windowing
-
Specifies the conditions under which a final pane will be created when a window is permanently
closed.
- Window.OnTimeBehavior - Enum in org.apache.beam.sdk.transforms.windowing
-
Specifies the conditions under which an on-time pane will be created when a window is closed.
- WindowAssignTranslator<T> - Class in org.apache.beam.runners.gearpump.translators
-
Window.Assign
is translated to Gearpump flatMap function.
- WindowAssignTranslator() - Constructor for class org.apache.beam.runners.gearpump.translators.WindowAssignTranslator
-
- WindowAssignTranslator.AssignWindows<T> - Class in org.apache.beam.runners.gearpump.translators
-
A Function used internally by Gearpump to wrap the actual Beam's WindowFn.
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
-
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
-
Returns the
Coder
used for serializing the windows used
by this windowFn.
- WindowedContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.WindowedContext
-
- windowedFilename(int, int, BoundedWindow, PaneInfo, FileBasedSink.OutputFileHints) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy
-
- windowedFilename(int, int, BoundedWindow, PaneInfo, FileBasedSink.OutputFileHints) - Method in class org.apache.beam.sdk.io.FileBasedSink.FilenamePolicy
-
When a sink has requested windowed or triggered output, this method will be invoked to return
the file
resource
to be created given the base output directory and a
FileBasedSink.OutputFileHints
containing information about the file, including a suggested
extension (e.g.
- windowedWrites - Variable in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-
Whether windowed writes are being used.
- WindowFn<T,W extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
-
The argument to the
Window
transform used to assign elements into
windows and to determine how windows are merged.
- WindowFn() - Constructor for class org.apache.beam.sdk.transforms.windowing.WindowFn
-
- WindowFn.AssignContext - Class in org.apache.beam.sdk.transforms.windowing
-
- WindowFn.MergeContext - Class in org.apache.beam.sdk.transforms.windowing
-
- WindowFnTestUtils - Class in org.apache.beam.sdk.testing
-
- WindowFnTestUtils() - Constructor for class org.apache.beam.sdk.testing.WindowFnTestUtils
-
- WindowingStrategy<T,W extends BoundedWindow> - Class in org.apache.beam.sdk.values
-
A WindowingStrategy
describes the windowing behavior for a specific collection of values.
- WindowingStrategy.AccumulationMode - Enum in org.apache.beam.sdk.values
-
The accumulation modes that can be used with windowing.
- WindowMappingFn<TargetWindowT extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
-
Experimental! This will be ready for users eventually, but should be considered internal for
now.
- WindowMappingFn() - Constructor for class org.apache.beam.sdk.transforms.windowing.WindowMappingFn
-
- WindowMappingFn(Duration) - Constructor for class org.apache.beam.sdk.transforms.windowing.WindowMappingFn
-
- windows() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn.MergeContext
-
Returns the current set of windows.
- WindowTimestampFn(int) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamAggregationTransforms.WindowTimestampFn
-
- with(SimpleFunction<DataT, InputT>, Combine.CombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposeCombineFnBuilder
-
- with(SimpleFunction<DataT, InputT>, CombineWithContext.CombineFnWithContext<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposeCombineFnBuilder
-
- with(SimpleFunction<DataT, InputT>, Combine.CombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
-
- with(SimpleFunction<DataT, InputT>, CombineWithContext.CombineFnWithContext<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
-
- with(SimpleFunction<DataT, InputT>, CombineFnBase.GlobalCombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
-
- withAddresses(List<String>) - Method in class org.apache.beam.sdk.io.amqp.AmqpIO.Read
-
Define the AMQP addresses where to receive messages.
- withAllowedLateness(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.Window
-
Override the amount of lateness allowed for data elements in the output
PCollection
and downstream
PCollections
until explicitly set again.
- withAllowedLateness(Duration, Window.ClosingBehavior) - Method in class org.apache.beam.sdk.transforms.windowing.Window
-
Override the amount of lateness allowed for data elements in the pipeline.
- withAllowedLateness(Duration) - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
Returns a
WindowingStrategy
identical to
this
but with the allowed lateness
set to
allowedLateness
.
- withAllowedTimestampSkew(Duration) - Method in class org.apache.beam.sdk.transforms.WithTimestamps
-
- withAuth(String) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
-
Define the password to authenticate on the Redis server.
- withAuth(String) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
-
- withAuth(String) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadAll
-
- withAuth(String) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
-
- withAWSClientsProvider(AWSClientsProvider) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
-
- withAWSClientsProvider(String, String, Regions) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
-
Specify credential details and region to be used to read from Kinesis.
- withBaseFilename(ResourceId) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
-
Sets the base filename.
- withBaseFilename(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
-
- withBasicCredentials(String, String) - Method in class org.apache.beam.sdk.io.solr.SolrIO.ConnectionConfiguration
-
If Solr basic authentication is enabled, provide the username and password.
- withBatchSize(long) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
-
Provide a size for the scroll read.
- withBatchSize(long) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
-
Sets batch size for the write operation.
- withBatchSize(long) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
-
Define the size of the batch to group write operations.
- withBatchSizeBytes(long) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-
Specifies the batch size limit.
- withBigtableOptions(BigtableOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-
Returns a new
BigtableIO.Read
that will read from the Cloud Bigtable instance
indicated by the given options, and using any other specified customizations.
- withBigtableOptions(BigtableOptions.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-
Returns a new
BigtableIO.Read
that will read from the Cloud Bigtable instance
indicated by the given options, and using any other specified customizations.
- withBigtableOptions(BigtableOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
-
Returns a new
BigtableIO.Write
that will write to the Cloud Bigtable instance
indicated by the given options, and using any other specified customizations.
- withBigtableOptions(BigtableOptions.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
-
Returns a new
BigtableIO.Write
that will write to the Cloud Bigtable instance
indicated by the given options, and using any other specified customizations.
- withBootstrapServers(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
Sets the bootstrap servers for the Kafka consumer.
- withBootstrapServers(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
-
Returns a new
KafkaIO.Write
transform with Kafka producer pointing to
bootstrapServers
.
- withBucket(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-
- withBucket(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-
- withCassandraService(CassandraService<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-
Specify an instance of
CassandraService
used to connect and read from Cassandra
database.
- withCassandraService(CassandraService<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
-
Specify the
CassandraService
used to connect and write into the Cassandra database.
- withCharset(Charset) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
-
Sets the XML file charset.
- withCharset(Charset) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
-
- withCharset(Charset) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
-
Sets the charset used to write the file.
- withChunkSize(Long) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-
- withClosingBehavior(Window.ClosingBehavior) - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
- withCodec(CodecFactory) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
-
Writes to Avro file(s) compressed using specified codec.
- withCodec(CodecFactory) - Method in class org.apache.beam.sdk.io.AvroIO.Write
-
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
-
Sets a coder for the result of the parse function.
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.AvroIO.ParseAll
-
Specifies the coder for the result of the parseFn
.
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-
Sets a
Coder
for the result of the parse function.
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-
- withCoder(Coder<OutputT>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.transforms.Create.TimestampedValues
-
Returns a
Create.TimestampedValues
PTransform like this one that uses the given
Coder<T>
to decode each of the objects into a
value of type
T
.
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.transforms.Create.Values
-
Returns a
Create.Values
PTransform like this one that uses the given
Coder<T>
to decode each of the objects into a
value of type
T
.
- withCollection(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
-
Sets the collection to consider in the database.
- withCollection(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
-
Sets the collection where to write data in the database.
- withColumns(String...) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-
- withColumns(List<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-
- withColumns(String...) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
- withColumns(List<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.CompressedSource
-
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.FileIO.ReadMatches
-
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.TextIO.Read
-
Reads from input sources using the specified compression type.
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
-
Reads from input sources using the specified compression type.
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-
Returns a transform for writing to text files like this one but that compresses output using
the given
Compression
.
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
-
Returns a transform for reading TFRecord files that decompresses all input files using the
specified compression type.
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
-
Writes to output files using the specified compression type.
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
-
Decompresses all input files using the specified compression type.
- withCompressionType(TextIO.CompressionType) - Method in class org.apache.beam.sdk.io.TextIO.Read
-
- withCompressionType(TextIO.CompressionType) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
-
- withCompressionType(TFRecordIO.CompressionType) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
-
- withCompressionType(TFRecordIO.CompressionType) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
-
- withCompressionType(XmlIO.Read.CompressionType) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
-
- withConfigProperties(Map<String, String>) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
-
Sets the configuration properties like metastore URI.
- withConfigProperties(Map<String, String>) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
-
Sets the configuration properties like metastore URI.
- withConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.FileIO.Match
-
- withConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.FileIO.MatchAll
-
- withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.Read
-
Reads from the source using the options provided by the given configuration.
- withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
-
Reads from the HBase instance indicated by the* given configuration.
- withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
-
Writes to the HBase instance indicated by the* given Configuration.
- withConnectionConfiguration(ElasticsearchIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
-
Provide the Elasticsearch connection configuration object.
- withConnectionConfiguration(ElasticsearchIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
-
Provide the Elasticsearch connection configuration object.
- withConnectionConfiguration(MqttIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Read
-
Define the MQTT connection configuration used to connect to the MQTT broker.
- withConnectionConfiguration(MqttIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Write
-
Define MQTT connection configuration used to connect to the MQTT broker.
- withConnectionConfiguration(RedisConnectionConfiguration) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
-
- withConnectionConfiguration(RedisConnectionConfiguration) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadAll
-
- withConnectionConfiguration(RedisConnectionConfiguration) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
-
- withConnectionConfiguration(SolrIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Read
-
Provide the Solr connection configuration object.
- withConnectionConfiguration(SolrIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Write
-
Provide the Solr connection configuration object.
- withConnectionFactory(ConnectionFactory) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
-
Specify the JMS connection factory to connect to the JMS broker.
- withConnectionFactory(ConnectionFactory) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
-
Specify the JMS connection factory to connect to the JMS broker.
- withConnectionProperties(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
-
Sets the connection properties passed to driver.connect(...).
- withConsistencyLevel(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-
- withConsistencyLevel(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
-
- withConsumerFactoryFn(SerializableFunction<Map<String, Object>, Consumer<byte[], byte[]>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
A factory to create Kafka Consumer
from consumer configuration.
- withContentTypeHint(String) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
Returns a new transform which will use the provided content type hint
to make the file parser detection more efficient.
- withCreateDisposition(BigQueryIO.Write.CreateDisposition) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
Specifies whether the table should be created if it does not exist.
- withCustomGcsTempLocation(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
Provides a custom location on GCS for storing temporary files to be loaded via BigQuery
batch load jobs.
- withDatabase(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
-
Sets the database name.
- withDatabase(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
-
Sets the database name.
- withDatabase(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-
- withDatabase(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-
- withDatabase(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
-
Sets the database to use.
- withDatabase(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
-
Sets the database to use.
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
-
Specifies the Cloud Spanner database.
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
-
Specifies the Cloud Spanner database.
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
Specifies the Cloud Spanner database.
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
Specifies the Cloud Spanner database.
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-
Specifies the Cloud Spanner database.
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-
Specifies the Cloud Spanner database.
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-
Specifies the Cloud Spanner database.
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-
Specifies the Cloud Spanner database.
- withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-
- withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-
- withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
-
- withDecompression(CompressedSource.DecompressingChannelFactory) - Method in class org.apache.beam.sdk.io.CompressedSource
-
- WithDefault() - Constructor for class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
-
- withDefaultValue(T) - Method in class org.apache.beam.sdk.transforms.View.AsSingleton
-
Default value to return for windows with no value in them.
- withDelimiter(byte[]) - Method in class org.apache.beam.sdk.io.TextIO.Read
-
Set the custom delimiter to be used in place of the default ones ('\r', '\n' or '\r\n').
- withDelimiter(byte[]) - Method in class org.apache.beam.sdk.io.TextIO.ReadFiles
-
Like Read#withDelimiter
.
- withDirectoryTreatment(FileIO.ReadMatches.DirectoryTreatment) - Method in class org.apache.beam.sdk.io.FileIO.ReadMatches
-
Controls how to handle directories in the input
PCollection
.
- withEarlyFirings(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-
- withEarlyFirings(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
-
Creates a new Trigger
like the this, except that it fires repeatedly whenever
the given Trigger
fires before the watermark has passed the end of the window.
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
-
Like Read#withEmptyMatchTreatment
.
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.AvroIO.ParseAll
-
Like Read#withEmptyMatchTreatment
.
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.AvroIO.Read
-
Configures whether or not a filepattern matching no files is allowed.
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.AvroIO.ReadAll
-
Like Read#withEmptyMatchTreatment
.
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.AvroSource
-
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.FileIO.Match
-
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.FileIO.MatchAll
-
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
-
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.TextIO.Read
-
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
-
Same as Read#withEmptyMatchTreatment
.
- withEndKey(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
-
Returns new
ByteKeyRange
like this one, but with the specified end key.
- withEndpoint(String, int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
-
- withEndpoint(String, int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadAll
-
- withEndpoint(String, int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
-
- withEntity(Class<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-
Specify the entity class (annotated POJO).
- withEntity(Class<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
-
- withEpsilon(double) - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
-
Returns an ApproximateQuantilesCombineFn
that's like
this one except that it uses the specified epsilon
value.
- withExtensionsFrom(Iterable<Class<?>>) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
-
Returns a
ProtoCoder
like this one, but with the extensions from the given classes
registered.
- withExtensionsFrom(Class<?>...) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
-
- withFailedInsertRetryPolicy(InsertRetryPolicy) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
Specfies a policy for handling failed inserts.
- withFanout(int) - Method in class org.apache.beam.sdk.transforms.Combine.Globally
-
Returns a
PTransform
identical to this, but that uses an intermediate node
to combine parts of the data to reduce load on the final global combine step.
- withFilename(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-
- withFilter(Filter) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
-
Filters the rows read from HBase using the given* row filter.
- withFilter(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
-
Sets the filter details.
- withFilter(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-
- withFilter(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
-
Sets a filter on the documents in a collection.
- withFooter(String) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-
Adds a footer string to each file.
- withFooter(String) - Method in class org.apache.beam.sdk.io.TextIO.Write
-
- withFormatFunction(SerializableFunction<UserT, OutputT>) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
-
Specifies a format function to convert UserT
to the output type.
- withFormatFunction(SerializableFunction<T, TableRow>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
Formats the user's type into a
TableRow
to be written to BigQuery.
- withFormatFunction(SerializableFunction<UserT, String>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-
Specifies a format function to convert UserT
to the output type.
- withGapDuration(Duration) - Static method in class org.apache.beam.sdk.transforms.windowing.Sessions
-
Creates a
Sessions
WindowFn
with the specified gap duration.
- withHeader(String) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-
Adds a header string to each file.
- withHeader(String) - Method in class org.apache.beam.sdk.io.TextIO.Write
-
- withHintMatchesManyFiles() - Method in class org.apache.beam.sdk.io.AvroIO.Parse
-
Like Read#withHintMatchesManyFiles()
.
- withHintMatchesManyFiles() - Method in class org.apache.beam.sdk.io.AvroIO.Read
-
- withHintMatchesManyFiles() - Method in class org.apache.beam.sdk.io.TextIO.Read
-
- withHost(String) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
-
Define the host name of the Redis server.
- withHosts(List<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-
Specify the hosts of the Apache Cassandra instances.
- withHosts(List<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
-
Specify the Cassandra instance hosts where to write data.
- withHotKeyFanout(SerializableFunction<? super K, Integer>) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
-
If a single key has disproportionately many values, it may become a
bottleneck, especially in streaming mode.
- withHotKeyFanout(int) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
-
- withIdAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
-
When reading from Cloud Pub/Sub where unique record identifiers are provided as Pub/Sub
message attributes, specifies the name of the attribute containing the unique identifier.
- withIdAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
-
Writes to Pub/Sub, adding each record's unique identifier to the published messages in an
attribute with the specified name.
- withIndex(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-
- withIndex(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
- withInitialPositionInStream(InitialPositionInStream) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
-
Specify reading from some initial position in stream.
- withInitialTimestampInStream(Instant) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
-
Specify reading beginning at given
Instant
.
- withInputMetadata(Metadata) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
Returns a new transform which will use the provided input metadata
for parsing the files.
- withInsertRetryPolicy(InsertRetryPolicy) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StreamingInserts
-
Specify a retry policy for failed inserts.
- withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-
- withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-
- withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
-
Specifies the Cloud Spanner instance.
- withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
-
Specifies the Cloud Spanner instance.
- withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
Specifies the Cloud Spanner instance.
- withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
Specifies the Cloud Spanner instance.
- withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-
Specifies the Cloud Spanner instance.
- withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-
Specifies the Cloud Spanner instance.
- withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-
Specifies the Cloud Spanner instance.
- withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-
Specifies the Cloud Spanner instance.
- withJsonSchema(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
- withJsonSchema(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
- withJsonTimePartitioning(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
- withKeepAlive(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
-
Sets whether socket keep alive is enabled.
- withKeepAlive(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
-
Sets whether socket keep alive is enabled.
- withKeyDeserializer(Class<? extends Deserializer<K>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
Sets a Kafka Deserializer
to interpret key bytes read from Kafka.
- withKeyDeserializerAndCoder(Class<? extends Deserializer<K>>, Coder<K>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
Sets a Kafka
Deserializer
for interpreting key bytes read from Kafka along with a
Coder
for helping the Beam runner materialize key objects at runtime if necessary.
- withKeyPattern(String) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
-
- withKeyRange(ByteKeyRange) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-
Returns a new
BigtableIO.Read
that will read only rows in the specified range.
- withKeyRange(ByteKeyRange) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
-
Reads only rows in the specified range.
- withKeyRange(byte[], byte[]) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
-
Reads only rows in the specified range.
- WithKeys<K,V> - Class in org.apache.beam.sdk.transforms
-
WithKeys<K, V>
takes a PCollection<V>
, and either a
constant key of type K
or a function from V
to
K
, and returns a PCollection<KV<K, V>>
, where each
of the values in the input PCollection
has been paired with
either the constant key or a key computed from the value.
- withKeySerializer(Class<? extends Serializer<K>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
-
Sets a Serializer
for serializing key (if any) to bytes.
- withKeySet(KeySet) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-
- withKeySet(KeySet) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
- withKeyspace(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-
Specify the Cassandra keyspace where to read data.
- withKeyspace(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
-
Specify the Cassandra keyspace where to write data.
- withKeystorePassword(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-
If Elasticsearch uses SSL/TLS with mutual authentication (via shield),
provide the password to open the client keystore.
- withKeystorePath(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-
If Elasticsearch uses SSL/TLS with mutual authentication (via shield),
provide the keystore containing the client key.
- withKeyTranslation(SimpleFunction<?, K>) - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.Read
-
Transforms the keys read from the source using the given key translation function.
- withKeyType(TypeDescriptor<K>) - Method in class org.apache.beam.sdk.transforms.WithKeys
-
Return a
WithKeys
that is like this one with the specified key type descriptor.
- withLabel(String) - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
-
- withLateFirings(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-
- withLateFirings(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
-
Creates a new Trigger
like the this, except that it fires repeatedly whenever
the given Trigger
fires after the watermark has passed the end of the window.
- withLinkUrl(String) - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
-
- withLiteralGqlQuery(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
Returns a new
DatastoreV1.Read
that reads the results of the specified GQL query.
- withLiteralGqlQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
- withLocalDc(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-
Specify the local DC used for the load balancing.
- withLocalDc(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
-
Specify the local DC used by the load balancing policy.
- withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
-
Returns a new
DatastoreV1.DeleteEntity
that deletes entities from the Cloud Datastore Emulator
running locally on the specified host port.
- withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
-
Returns a new
DatastoreV1.DeleteKey
that deletes entities from the Cloud Datastore Emulator
running locally on the specified host port.
- withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
Returns a new
DatastoreV1.Read
that reads from a Datastore Emulator running at the
given localhost address.
- withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
-
Returns a new
DatastoreV1.Write
that writes to the Cloud Datastore Emulator running locally on
the specified host port.
- withMapper(ObjectMapper) - Method in class org.apache.beam.sdk.extensions.jackson.AsJsons
-
- withMapper(ObjectMapper) - Method in class org.apache.beam.sdk.extensions.jackson.ParseJsons
-
- withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
-
- withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.AvroIO.ParseAll
-
- withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.AvroIO.Read
-
- withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.AvroIO.ReadAll
-
- withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.TextIO.Read
-
- withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
-
- withMaxBatchSize(long) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
-
Provide a maximum size in number of documents for the batch see bulk API
(https://www.elastic.co/guide/en/elasticsearch/reference/2.4/docs-bulk.html).
- withMaxBatchSizeBytes(long) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
-
Provide a maximum size in bytes for the batch see bulk API
(https://www.elastic.co/guide/en/elasticsearch/reference/2.4/docs-bulk.html).
- withMaxConnectionIdleTime(int) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
-
Sets the maximum idle time for a pooled connection.
- withMaxConnectionIdleTime(int) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
-
Sets the maximum idle time for a pooled connection.
- withMaxInputSize(long) - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
-
Returns an ApproximateQuantilesCombineFn
that's like
this one except that it uses the specified maxNumElements
value.
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.amqp.AmqpIO.Read
-
Define the max number of records received by the
AmqpIO.Read
.
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.BoundedReadFromUnboundedSource
-
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
-
Define the max number of records that the source will read.
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
- withMaxNumRecords(int) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
-
Specifies to read at most a given number of records.
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Read
-
Define the max number of records received by the
MqttIO.Read
.
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.Read.Unbounded
-
- withMaxNumWritersPerBundle(int) - Method in class org.apache.beam.sdk.io.WriteFiles
-
Set the maximum number of writers created in a bundle before spilling to shuffle.
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.amqp.AmqpIO.Read
-
Define the max read time (duration) while the
AmqpIO.Read
will receive messages.
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.BoundedReadFromUnboundedSource
-
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.GenerateSequence
-
Specifies to stop generating elements after the given time.
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
-
Define the max read time that the source will read.
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
-
Specifies to read at most a given number of records.
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Read
-
Define the max read time (duration) while the
MqttIO.Read
will receive messages.
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.Read.Unbounded
-
- withMemoryMB(int) - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
-
Sets the size of the memory buffer in megabytes.
- withMetadata(Map<String, Object>) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
-
Writes to Avro file(s) with the specified metadata.
- withMetadata(Map<String, Object>) - Method in class org.apache.beam.sdk.io.AvroIO.Write
-
- withMethod(BigQueryIO.Write.Method) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
Choose the method used to write data to BigQuery.
- WithMetricsSupport - Class in org.apache.beam.runners.spark.metrics
-
A
MetricRegistry
decorator-like that supports
AggregatorMetric
and
SparkBeamMetric
as
Gauges
.
- withMinBundleSize(long) - Method in class org.apache.beam.sdk.io.AvroSource
-
Sets the minimum bundle size.
- withMinBundleSize(long) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
-
Sets a parameter minBundleSize
for the minimum bundle size of the source.
- withMinimumTextlength(Integer) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
Returns a new transform which will operate on the text blocks with the
given minimum text length.
- withMode(WindowingStrategy.AccumulationMode) - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
Returns a
WindowingStrategy
identical to
this
but with the accumulation mode
set to
mode
.
- withNamespace(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
- withNamespace(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
- withNamespace(Class<?>) - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
-
- withNumFileShards(int) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
Control how many file shards are written when using BigQuery load jobs.
- withNumQuerySplits(int) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
Returns a new
DatastoreV1.Read
that reads by splitting the given
query
into
numQuerySplits
.
- withNumShards(int) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
-
Configures the number of output shards produced overall (when using unwindowed writes) or
per-window (when using windowed writes).
- withNumShards(int) - Method in class org.apache.beam.sdk.io.AvroIO.Write
-
- withNumShards(int) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-
Configures the number of output shards produced overall (when using unwindowed writes) or
per-window (when using windowed writes).
- withNumShards(int) - Method in class org.apache.beam.sdk.io.TextIO.Write
-
- withNumShards(int) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
-
Writes to the provided number of shards.
- withNumShards(int) - Method in class org.apache.beam.sdk.io.WriteFiles
-
- withNumShards(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.WriteFiles
-
- withNumSplits(int) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
-
Sets the user defined number of splits.
- withOffset(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-
Partitions the timestamp space into half-open intervals of the form
[N * size + offset, (N + 1) * size + offset),
where 0 is the epoch.
- withOffset(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-
Assigns timestamps into half-open intervals of the form
[N * period + offset, N * period + offset + size).
- withOnTimeBehavior(Window.OnTimeBehavior) - Method in class org.apache.beam.sdk.transforms.windowing.Window
-
(Experimental) Override the default
Window.OnTimeBehavior
, to control
whether to output an empty on-time pane.
- withOnTimeBehavior(Window.OnTimeBehavior) - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
- withOptions(TikaOptions) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
Path to Tika configuration resource.
- withoutDefaults() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
-
Returns a
PTransform
identical to this, but that does not attempt to
provide a default value in the case of empty input.
- withoutMetadata() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
Returns a
PTransform
for PCollection of
KV
, dropping Kafka metatdata.
- withOutputCoder(Coder<OutputT>) - Method in class org.apache.beam.sdk.transforms.Watch.Growth
-
Specifies a
Coder
to use for the outputs.
- withOutputFilenames() - Method in class org.apache.beam.sdk.io.AvroIO.Write
-
Specify that output filenames are wanted.
- withOutputFilenames() - Method in class org.apache.beam.sdk.io.TextIO.Write
-
Specify that output filenames are wanted.
- withOutputTags(TupleTag<OutputT>, TupleTagList) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
-
- withoutResultFlattening() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
-
- withoutResultFlattening() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-
- withoutSharding() - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
-
Forces a single file as output and empty shard name template.
- withoutSharding() - Method in class org.apache.beam.sdk.io.AvroIO.Write
-
- withoutSharding() - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-
Forces a single file as output and empty shard name template.
- withoutSharding() - Method in class org.apache.beam.sdk.io.TextIO.Write
-
- withoutSharding() - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
-
Forces a single file as output.
- withoutStrictParsing() - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
-
During parsing of the arguments, we will skip over improperly formatted and unknown
arguments.
- withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
-
Disable validation that the table exists or the query succeeds prior to pipeline submission.
- withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-
- withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
Disables BigQuery table validation.
- withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-
Disables validation that the table being read from exists.
- withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
-
Disables validation that the table being written to exists.
- withoutValidation() - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
-
Returns a transform for reading TFRecord files that has GCS path validation on
pipeline creation disabled.
- withParameterSetter(JdbcIO.PreparedStatementSetter<ParameterT>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-
- withParseFn(SerializableFunction<GenericRecord, X>, Coder<X>) - Method in class org.apache.beam.sdk.io.AvroSource
-
Reads
GenericRecord
of unspecified schema and maps them to instances of a custom type
using the given
parseFn
and encoded using the given coder.
- withParser(MongoDbGridFSIO.Parser<X>) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-
- withParseSynchronously(Boolean) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
Returns a new transform which will use the synchronous reader.
- withPartition(Map<String, String>) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
-
Sets the partition details.
- withPassword(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-
Specify the password for authentication.
- withPassword(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
-
Specify the password used for authentication.
- withPassword(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-
If Elasticsearch authentication is enabled, provide the password.
- withPassword(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
-
- withPassword(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
-
Define the password to connect to the JMS broker (authenticated).
- withPassword(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
-
Define the password to connect to the JMS broker (authenticated).
- withPassword(String) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
-
- withPollInterval(Duration) - Method in class org.apache.beam.sdk.transforms.Watch.Growth
-
- withPort(int) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-
Specify the port number of the Apache Cassandra instances.
- withPort(int) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
-
Specify the Cassandra instance port number where to write data.
- withPort(int) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
-
Define the port number of the Redis server.
- withPreparedStatementSetter(JdbcIO.PreparedStatementSetter<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
-
- withProducerFactoryFn(SerializableFunction<Map<String, Object>, Producer<K, V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
-
Sets a custom function to create Kafka producer.
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
-
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
-
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
-
Returns a new
DatastoreV1.DeleteKey
that deletes entities from the Cloud Datastore for the
specified project.
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
-
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
Returns a new
DatastoreV1.Read
that reads from the Cloud Datastore for the specified
project.
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
-
Returns a new
DatastoreV1.Write
that writes to the Cloud Datastore for the specified project.
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
-
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
-
Specifies the Cloud Spanner project.
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
-
Specifies the Cloud Spanner project.
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
Specifies the Cloud Spanner project.
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
Specifies the Cloud Spanner project.
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-
Specifies the Cloud Spanner project.
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-
Specifies the Cloud Spanner project.
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-
Specifies the Cloud Spanner project.
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-
Specifies the Cloud Spanner project.
- withQuery(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
-
Provide a query used while reading from Elasticsearch.
- withQuery(Query) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-
- withQuery(Statement) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-
- withQuery(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-
- withQuery(Statement) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
- withQuery(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
- withQuery(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-
- withQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-
- withQuery(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-
- withQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-
- withQuery(String) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Read
-
Provide a query used while reading from Solr.
- withQueue(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
-
Specify the JMS queue destination name where to read messages from.
- withQueue(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
-
Specify the JMS queue destination name where to send messages to.
- withQueueMaxPollTime(Long) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
Returns a new transform which will use the specified queue max poll time.
- withQueuePollTime(Long) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
Returns a new transform which will use the specified queue poll time.
- withRate(long, Duration) - Method in class org.apache.beam.sdk.io.GenerateSequence
-
Specifies to generate at most a given number of elements per a given period.
- withReadOperation(ReadOperation) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
- withReadOutputMetadata(Boolean) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
Returns a new transform which will report the metadata.
- withRecordClass(Class<T>) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
-
Sets a JAXB annotated class that can be populated using a record of the provided XML file.
- withRecordClass(Class<T>) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
-
- withRecordClass(Class<T>) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
-
Writes objects of the given class mapped to XML elements using JAXB.
- withRecordElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
-
Sets name of the record element of the XML document.
- withRecordElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
-
- withRepresentativeType(TypeDescriptor<IdT>) - Method in class org.apache.beam.sdk.transforms.Distinct.WithRepresentativeValues
-
Return a
WithRepresentativeValues
PTransform
that is like this one, but with
the specified output type descriptor.
- withRepresentativeValueFn(SerializableFunction<T, IdT>) - Static method in class org.apache.beam.sdk.transforms.Distinct
-
Returns a Distinct<T, IdT>
PTransform
.
- withResumeDelay(Duration) - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
-
- withRetained(boolean) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Write
-
Whether or not the publish message should be retained by the messaging engine.
- withRootDirectory(File) - Static method in class org.apache.beam.artifact.local.LocalFileSystemArtifactStagerService
-
- withRootElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
-
Sets name of the root element of the XML document.
- withRootElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
-
- withRootElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
-
Sets the enclosing root element for the generated XML files.
- withRowFilter(RowFilter) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-
Returns a new
BigtableIO.Read
that will filter the rows read from Cloud Bigtable
using the given row filter.
- withRowMapper(JdbcIO.RowMapper<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-
- withRowMapper(JdbcIO.RowMapper<OutputT>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-
- withRunnerDeterminedSharding() - Method in class org.apache.beam.sdk.io.WriteFiles
-
- withScan(Scan) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
-
Filters the rows read from HBase using the given* scan.
- withSchema(Schema) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
-
Sets the the output schema.
- withSchema(Schema) - Method in class org.apache.beam.sdk.io.AvroIO.Write
-
- withSchema(String) - Method in class org.apache.beam.sdk.io.AvroSource
-
Reads files containing records that conform to the given schema.
- withSchema(Schema) - Method in class org.apache.beam.sdk.io.AvroSource
-
- withSchema(Class<X>) - Method in class org.apache.beam.sdk.io.AvroSource
-
Reads files containing records of the given class.
- withSchema(TableSchema) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
Uses the specified schema for rows to be written.
- withSchema(ValueProvider<TableSchema>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
- withSchemaFromView(PCollectionView<Map<String, String>>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
Allows the schemas for each table to be computed within the pipeline itself.
- withScrollKeepalive(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
-
Provide a scroll keepalive.
- withShard(int) - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-
- withSharding(PTransform<PCollection<UserT>, PCollectionView<Integer>>) - Method in class org.apache.beam.sdk.io.WriteFiles
-
- withShardNameTemplate(String) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
-
- withShardNameTemplate(String) - Method in class org.apache.beam.sdk.io.AvroIO.Write
-
- withShardNameTemplate(String) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-
- withShardNameTemplate(String) - Method in class org.apache.beam.sdk.io.TextIO.Write
-
- withShardNameTemplate(String) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
-
Uses the given shard name template.
- withShardTemplate(String) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
-
Sets the shard template.
- withSideInputs(List<PCollectionView<?>>) - Method in class org.apache.beam.sdk.io.WriteFiles
-
- withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.Combine.Globally
-
- withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.Combine.Globally
-
- withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
-
- withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
-
- withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
-
- withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
-
- withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-
- withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-
- withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
-
- withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
-
- withSideInputStream(TranslationContext, JavaStream<WindowedValue<InputT>>, Map<String, PCollectionView<?>>) - Static method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils
-
- withSingletonValues() - Method in class org.apache.beam.sdk.transforms.View.AsMap
-
- withSkew(Duration) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-
- withSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
-
Specifies the Cloud Spanner configuration.
- withSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
Specifies the Cloud Spanner configuration.
- withSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-
Specifies the Cloud Spanner configuration.
- withSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-
Specifies the Cloud Spanner configuration.
- withStartingDay(int, int, int) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-
- withStartingMonth(int, int) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-
- withStartingYear(int) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-
- withStartKey(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
-
Returns new
ByteKeyRange
like this one, but with the specified start key.
- withStartReadTime(Instant) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
Use timestamp to set up start offset.
- withStatement(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
-
- withStatementPreparator(JdbcIO.StatementPreparator) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-
- withStreamName(String) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
-
Specify reading from streamName.
- withSuffix(String) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
-
Configures the filename suffix for written files.
- withSuffix(String) - Method in class org.apache.beam.sdk.io.AvroIO.Write
-
- withSuffix(String) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
-
Sets the suffix.
- withSuffix(String) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-
Configures the filename suffix for written files.
- withSuffix(String) - Method in class org.apache.beam.sdk.io.TextIO.Write
-
- withSuffix(String) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
-
Writes to the file(s) with the given filename suffix.
- withTable(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-
Specify the Cassandra table where to read data.
- withTable(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-
- withTable(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
- withTable(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
-
Sets the table name to read from.
- withTable(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
-
Sets the table name to write to, the table should exist beforehand.
- withTableDescription(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
Specifies the table description.
- withTableId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-
- withTableId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
-
- withTableId(String) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
-
Reads from the specified table.
- withTableId(String) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
-
Writes to the specified table.
- withTempDirectory(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
-
Set the base directory used to generate temporary files.
- withTempDirectory(ResourceId) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
-
Set the base directory used to generate temporary files.
- withTempDirectory(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.AvroIO.Write
-
- withTempDirectory(ResourceId) - Method in class org.apache.beam.sdk.io.AvroIO.Write
-
- withTempDirectory(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-
Set the base directory used to generate temporary files.
- withTempDirectory(ResourceId) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-
Set the base directory used to generate temporary files.
- withTempDirectory(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.TextIO.Write
-
- withTempDirectory(ResourceId) - Method in class org.apache.beam.sdk.io.TextIO.Write
-
- withTemplateCompatibility() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
-
Use new template-compatible source implementation.
- withTemplateCompatibility() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-
- withTempLocation(String) - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
-
Sets the path to a temporary location where the sorter writes intermediate files.
- withTerminationPerInput(Watch.Growth.TerminationCondition<InputT, ?>) - Method in class org.apache.beam.sdk.transforms.Watch.Growth
-
- withTikaConfigPath(String) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
Returns a new transform which will use the custom TikaConfig.
- withTikaConfigPath(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Read
-
- withTimeout(int) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
-
Define the Redis connection timeout.
- withTimeout(int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
-
- withTimeout(int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadAll
-
- withTimeout(int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
-
- withTimePartitioning(TimePartitioning) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
- withTimePartitioning(ValueProvider<TimePartitioning>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
- withTimestamp(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
- withTimestampAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
-
When reading from Cloud Pub/Sub where record timestamps are provided as Pub/Sub message
attributes, specifies the name of the attribute that contains the timestamp.
- withTimestampAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
-
Writes to Pub/Sub and adds each record's timestamp to the published messages in an attribute
with the specified name.
- withTimestampBound(TimestampBound) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
-
- withTimestampBound(TimestampBound) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
- withTimestampCombiner(TimestampCombiner) - Method in class org.apache.beam.sdk.transforms.windowing.Window
-
- withTimestampCombiner(TimestampCombiner) - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
- withTimestampFn(SerializableFunction<Long, Instant>) - Method in class org.apache.beam.sdk.io.GenerateSequence
-
Specifies the function to use to assign timestamps to the elements.
- withTimestampFn(SerializableFunction<KV<K, V>, Instant>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
A function to assign a timestamp to a record.
- withTimestampFn2(SerializableFunction<KafkaRecord<K, V>, Instant>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
A function to assign a timestamp to a record.
- WithTimestamps<T> - Class in org.apache.beam.sdk.transforms
-
- withTimeZone(DateTimeZone) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-
- withTimeZone(DateTimeZone) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-
- withTimeZone(DateTimeZone) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-
- withTopic(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
-
Specify the JMS topic destination name where to receive messages from.
- withTopic(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
-
Specify the JMS topic destination name where to send messages to.
- withTopic(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
Sets the topic to read from.
- withTopic(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
-
Sets the Kafka topic to write to.
- withTopicPartitions(List<TopicPartition>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
Sets a list of partitions to read from.
- withTopics(List<String>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
Sets a list of topics to read from.
- withTransaction(PCollectionView<Transaction>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-
- withTransaction(PCollectionView<Transaction>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-
- withTrigger(Trigger) - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
Returns a
WindowingStrategy
identical to
this
but with the trigger set to
wildcardTrigger
.
- withTriggeringFrequency(Duration) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
Choose the frequency at which file writes are triggered.
- withType(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.transforms.Create.TimestampedValues
-
Returns a
Create.TimestampedValues
PTransform like this one that uses the given
TypeDescriptor<T>
to determine the
Coder
to use to decode each of the
objects into a value of type
T
.
- withType(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.transforms.Create.Values
-
Returns a
Create.Values
PTransform like this one that uses the given
TypeDescriptor<T>
to determine the
Coder
to use to decode each of the
objects into a value of type
T
.
- withUdaf(String, Combine.CombineFn) - Method in class org.apache.beam.sdk.extensions.sql.BeamSql.QueryTransform
-
- withUdaf(String, Combine.CombineFn) - Method in class org.apache.beam.sdk.extensions.sql.BeamSql.SimpleQueryTransform
-
- withUdf(String, Class<? extends BeamSqlUdf>) - Method in class org.apache.beam.sdk.extensions.sql.BeamSql.QueryTransform
-
register a UDF function used in this query.
- withUdf(String, SerializableFunction) - Method in class org.apache.beam.sdk.extensions.sql.BeamSql.QueryTransform
-
- withUdf(String, Class<? extends BeamSqlUdf>) - Method in class org.apache.beam.sdk.extensions.sql.BeamSql.SimpleQueryTransform
-
register a UDF function used in this query.
- withUdf(String, SerializableFunction) - Method in class org.apache.beam.sdk.extensions.sql.BeamSql.SimpleQueryTransform
-
- withUpToDateThreshold(Duration) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
-
Specifies how late records consumed by this source can be to still be considered on time.
- withUri(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-
- withUri(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-
- withUri(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
-
Define the location of the MongoDB instances using an URI.
- withUri(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
-
Define the location of the MongoDB instances using an URI.
- withUsername(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-
Specify the username for authentication.
- withUsername(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
-
Specify the username used for authentication.
- withUsername(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-
If Elasticsearch authentication is enabled, provide the username.
- withUsername(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
-
- withUsername(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
-
Define the username to connect to the JMS broker (authenticated).
- withUsername(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
-
Define the username to connect to the JMS broker (authenticated).
- withUsername(String) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
-
- withValidation() - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
-
After creation we will validate that
PipelineOptions
conforms to all the
validation criteria from
<T>
.
- withValidation() - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory
-
After creation we will validate that <T>
conforms to all the
validation criteria.
- withValidationEventHandler(ValidationEventHandler) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
-
Sets the ValidationEventHandler
to use with JAXB.
- withValidationEventHandler(ValidationEventHandler) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
-
- withValueDeserializer(Class<? extends Deserializer<V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
Sets a Kafka Deserializer
to interpret value bytes read from Kafka.
- withValueDeserializerAndCoder(Class<? extends Deserializer<V>>, Coder<V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
Sets a Kafka
Deserializer
for interpreting value bytes read from Kafka along with a
Coder
for helping the Beam runner materialize value objects at runtime if necessary.
- withValueSerializer(Class<? extends Serializer<V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
-
Sets a Serializer
for serializing value to bytes.
- withValueTranslation(SimpleFunction<?, V>) - Method in class org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIO.Read
-
Transforms the values read from the source using the given value translation function.
- withWatermark(Instant) - Method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
-
Sets the watermark - an approximate lower bound on timestamps of future new outputs from
this
Watch.Growth.PollFn
.
- withWatermarkFn(SerializableFunction<KV<K, V>, Instant>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
A function to calculate watermark after a record.
- withWatermarkFn2(SerializableFunction<KafkaRecord<K, V>, Instant>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-
A function to calculate watermark after a record.
- withWindowedWrites() - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
-
Preserves windowing of input elements and writes them to files based on the element's window.
- withWindowedWrites() - Method in class org.apache.beam.sdk.io.AvroIO.Write
-
- withWindowedWrites() - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
-
Specify that writes are windowed.
- withWindowedWrites() - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-
Preserves windowing of input elements and writes them to files based on the element's window.
- withWindowedWrites() - Method in class org.apache.beam.sdk.io.TextIO.Write
-
- withWindowedWrites() - Method in class org.apache.beam.sdk.io.WriteFiles
-
Returns a new
WriteFiles
that writes preserves windowing on it's input.
- withWindowFn(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.values.WindowingStrategy
-
Returns a
WindowingStrategy
identical to
this
but with the window function
set to
wildcardWindowFn
.
- withWritableByteChannelFactory(FileBasedSink.WritableByteChannelFactory) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-
- withWritableByteChannelFactory(FileBasedSink.WritableByteChannelFactory) - Method in class org.apache.beam.sdk.io.TextIO.Write
-
See TypedWrite#withWritableByteChannelFactory(WritableByteChannelFactory)
.
- withWriteDisposition(BigQueryIO.Write.WriteDisposition) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
Specifies what to do with existing data in the table, in case the table already exists.
- WorkerHarnessContainerImageFactory() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions.WorkerHarnessContainerImageFactory
-
- WorkerLogLevelOverrides() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
-
- wrap(Throwable) - Method in class org.apache.beam.sdk.testing.PAssert.PAssertionSite
-
- wrap(String) - Method in class org.apache.beam.sdk.testing.PAssert.PAssertionSite
-
- WritableCoder<T extends org.apache.hadoop.io.Writable> - Class in org.apache.beam.sdk.io.hadoop
-
A
WritableCoder
is a
Coder
for a Java class that implements
Writable
.
- WritableCoder(Class<T>) - Constructor for class org.apache.beam.sdk.io.hadoop.WritableCoder
-
- WritableCoder.WritableCoderProviderRegistrar - Class in org.apache.beam.sdk.io.hadoop
-
- WritableCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.hadoop.WritableCoder.WritableCoderProviderRegistrar
-
- write() - Static method in class org.apache.beam.sdk.io.amqp.AmqpIO
-
- Write() - Constructor for class org.apache.beam.sdk.io.amqp.AmqpIO.Write
-
- write(Class<T>) - Static method in class org.apache.beam.sdk.io.AvroIO
-
Writes a
PCollection
to an Avro file (or multiple Avro files matching a sharding
pattern).
- write() - Static method in class org.apache.beam.sdk.io.cassandra.CassandraIO
-
- Write() - Constructor for class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
-
- write(T) - Method in interface org.apache.beam.sdk.io.cassandra.CassandraService.Writer
-
This method should be synchronous.
- write(T) - Method in class org.apache.beam.sdk.io.cassandra.CassandraServiceImpl.WriterImpl
-
Write the entity to the Cassandra instance, using Mapper
obtained with the
MappingManager
.
- write() - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO
-
- Write() - Constructor for class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
-
- write(OutputT) - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
-
Called for each value in the bundle.
- write() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
-
- Write() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-
- write() - Static method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO
-
- Write() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
-
- write() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1
-
- Write() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
-
- write() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO
-
- Write() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-
- write() - Static method in class org.apache.beam.sdk.io.hbase.HBaseIO
-
- write() - Static method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO
-
Write data to Hive.
- Write() - Constructor for class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
-
- write() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
-
Write data to a JDBC datasource.
- Write() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
-
- write() - Static method in class org.apache.beam.sdk.io.jms.JmsIO
-
- Write() - Constructor for class org.apache.beam.sdk.io.jms.JmsIO.Write
-
- write() - Static method in class org.apache.beam.sdk.io.kafka.KafkaIO
-
- Write() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.Write
-
- write() - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO
-
Write data to GridFS.
- write(MongoDbGridFSIO.WriteFn<T>) - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO
-
- Write() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-
- write(T, OutputStream) - Method in interface org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.WriteFn
-
Output the object to the given OutputStream.
- write() - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbIO
-
Write data to MongoDB.
- Write() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
-
- write() - Static method in class org.apache.beam.sdk.io.mqtt.MqttIO
-
- Write() - Constructor for class org.apache.beam.sdk.io.mqtt.MqttIO.Write
-
- write() - Static method in class org.apache.beam.sdk.io.redis.RedisIO
-
Write data to a Redis server.
- Write() - Constructor for class org.apache.beam.sdk.io.redis.RedisIO.Write
-
- write() - Static method in class org.apache.beam.sdk.io.solr.SolrIO
-
- Write() - Constructor for class org.apache.beam.sdk.io.solr.SolrIO.Write
-
- write() - Static method in class org.apache.beam.sdk.io.TextIO
-
A
PTransform
that writes a
PCollection
to a text file (or multiple text files
matching a sharding pattern), with each element of the input collection encoded into its own
line.
- write() - Static method in class org.apache.beam.sdk.io.TFRecordIO
-
A
PTransform
that writes a
PCollection
to TFRecord file (or
multiple TFRecord files matching a sharding pattern), with each
element of the input collection encoded into its own record.
- Write() - Constructor for class org.apache.beam.sdk.io.TFRecordIO.Write
-
- write() - Static method in class org.apache.beam.sdk.io.xml.XmlIO
-
- Write() - Constructor for class org.apache.beam.sdk.io.xml.XmlIO.Write
-
- write(T) - Method in interface org.apache.beam.sdk.state.ValueState
-
Set the value.
- writeAvros(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
-
Returns A
PTransform
that writes binary encoded Avro messages of a given type
to a Google Cloud Pub/Sub stream.
- writeCompressed(WritableByteChannel) - Method in enum org.apache.beam.sdk.io.Compression
-
- writeCustomType() - Static method in class org.apache.beam.sdk.io.AvroIO
-
A
PTransform
that writes a
PCollection
to an avro file (or multiple avro files
matching a sharding pattern), with each element of the input collection encoded into its own
record of type OutputT.
- writeCustomType() - Static method in class org.apache.beam.sdk.io.TextIO
-
A
PTransform
that writes a
PCollection
to a text file (or multiple text files
matching a sharding pattern), with each element of the input collection encoded into its own
line.
- writeCustomTypeToGenericRecords() - Static method in class org.apache.beam.sdk.io.AvroIO
-
- writeExternal(ObjectOutput) - Method in class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
-
- WriteFiles<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io
-
- WriteFilesResult<DestinationT> - Class in org.apache.beam.sdk.io
-
- writeFooter() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
-
Writes footer at the end of output files.
- writeGenericRecords(Schema) - Static method in class org.apache.beam.sdk.io.AvroIO
-
Writes Avro records of the specified schema.
- writeGenericRecords(String) - Static method in class org.apache.beam.sdk.io.AvroIO
-
Writes Avro records of the specified schema.
- WriteGrouped(SpannerIO.Write) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.WriteGrouped
-
- writeHeader() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
-
Writes header at the beginning of output files.
- writeMessages() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
-
Returns A
PTransform
that writes to a Google Cloud Pub/Sub stream.
- WriteOperation(FileBasedSink<?, DestinationT, OutputT>) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-
Constructs a WriteOperation using the default strategy for generating a temporary directory
from the base output filename.
- WriteOperation(FileBasedSink<?, DestinationT, OutputT>, ResourceId) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-
Create a new WriteOperation.
- writeProtos(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
-
Returns A
PTransform
that writes binary encoded protobuf messages of a given type
to a Google Cloud Pub/Sub stream.
- Writer(FileBasedSink.WriteOperation<DestinationT, OutputT>, String) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.Writer
-
- WriteResult - Class in org.apache.beam.sdk.io.gcp.bigquery
-
- WriterImpl(CassandraIO.Write<T>) - Constructor for class org.apache.beam.sdk.io.cassandra.CassandraServiceImpl.WriterImpl
-
- writeStrings() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
-
Returns A
PTransform
that writes UTF-8 encoded strings to a Google Cloud Pub/Sub
stream.
- writeTableRows() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
-
- writeTo(OutputStream, int, int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
-
Writes length
bytes starting at offset
from the backing data store to the
specified output stream.