- abort() - Method in interface org.apache.beam.runners.fnexecution.state.StateDelegator.Registration
- 
De-registers the handler for all future requests for state for the registered process bundle
 instruction id. 
- abort(Executor) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- absolute(String, String...) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData.Path
- 
Construct a path from an absolute component path hierarchy. 
- AbstractBeamCalcRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
- 
BeamRelNode to replace ProjectandFilternode.
 
- AbstractBeamCalcRel(RelOptCluster, RelTraitSet, RelNode, RexProgram) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.AbstractBeamCalcRel
-  
- AbstractGetterTypeSupplier() - Constructor for class org.apache.beam.sdk.schemas.AutoValueSchema.AbstractGetterTypeSupplier
-  
- AbstractReadFileRangesFn(SerializableFunction<String, ? extends FileBasedSource<InT>>, ReadAllViaFileBasedSource.ReadFileRangesFnExceptionHandler) - Constructor for class org.apache.beam.sdk.io.ReadAllViaFileBasedSourceTransform.AbstractReadFileRangesFn
-  
- AbstractResult() - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write.AbstractResult
-  
- accept(BeamFnApi.Elements) - Method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
-  
- accept(ByteString) - Method in class org.apache.beam.sdk.fn.data.DecodingFnDataReceiver
-  
- accept(T) - Method in interface org.apache.beam.sdk.fn.data.FnDataReceiver
-  
- accept(T1, T2) - Method in interface org.apache.beam.sdk.function.ThrowingBiConsumer
-  
- accept(T) - Method in interface org.apache.beam.sdk.function.ThrowingConsumer
-  
- accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierContext
-  
- accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierListContext
-  
- accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionContext
-  
- accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.FieldSpecifierContext
-  
- accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierContext
-  
- accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierListContext
-  
- accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifiedComponentContext
-  
- accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifyComponentContext
-  
- accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.SimpleIdentifierContext
-  
- accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.WildcardContext
-  
- accept(SchemaZipFold.Context, Schema.FieldType, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold
- 
Accepts two components, context.parent() is always ROW, MAP, ARRAY or absent. 
- accept(SchemaZipFold.Context, Optional<Schema.Field>, Optional<Schema.Field>) - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold
- 
Accepts two fields, context.parent() is always ROW. 
- accessPattern() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.SideInputSpec
-  
- accessType() - Method in class org.apache.beam.sdk.io.solace.data.Semp.QueueData
-  
- accumulate(T, T) - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold
- 
Accumulate two results together. 
- accumulateWeight(long) - Method in class org.apache.beam.sdk.fn.data.WeightedList
-  
- AccumulatingCombineFn() - Constructor for class org.apache.beam.sdk.transforms.Combine.AccumulatingCombineFn
-  
- accumulatingFiredPanes() - Method in class org.apache.beam.sdk.transforms.windowing.Window
- 
Returns a new WindowPTransformthat uses the registered WindowFn and
 Triggering behavior, and that accumulates elements in a pane after they are triggered.
 
- ACCUMULATOR_NAME - Static variable in class org.apache.beam.runners.flink.metrics.FlinkMetricContainer
-  
- AccumulatorCheckpointingSparkListener() - Constructor for class org.apache.beam.runners.spark.metrics.MetricsAccumulator.AccumulatorCheckpointingSparkListener
-  
- ackDeadlineSeconds(PubsubClient.SubscriptionPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
- 
Return the ack deadline, in seconds, for subscription.
 
- ackDeadlineSeconds(PubsubClient.SubscriptionPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
-  
- ackDeadlineSeconds(PubsubClient.SubscriptionPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
-  
- ackDeadlineSeconds(PubsubClient.SubscriptionPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- ackId() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.IncomingMessage
- 
Id to pass back to Pubsub to acknowledge receipt of this message. 
- acknowledge(PubsubClient.SubscriptionPath, List<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
- 
Acknowldege messages from subscriptionwithackIds.
 
- acknowledge(PubsubClient.SubscriptionPath, List<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
-  
- acknowledge(PubsubClient.SubscriptionPath, List<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
-  
- acknowledge(PubsubClient.SubscriptionPath, List<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- acquireTaskAttemptIdLock(Configuration, int) - Method in interface org.apache.beam.sdk.io.hadoop.format.ExternalSynchronization
- 
Creates unique TaskAttemptIDfor given taskId.
 
- acquireTaskAttemptIdLock(Configuration, int) - Method in class org.apache.beam.sdk.io.hadoop.format.HDFSSynchronization
-  
- acquireTaskIdLock(Configuration) - Method in interface org.apache.beam.sdk.io.hadoop.format.ExternalSynchronization
- 
Creates TaskIDwith unique id among given job.
 
- acquireTaskIdLock(Configuration) - Method in class org.apache.beam.sdk.io.hadoop.format.HDFSSynchronization
-  
- ActionFactory - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.action
- 
Factory class for creating instances that will handle different functions of DoFns. 
- ActionFactory() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.ActionFactory
-  
- ActionFactory - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.action
- 
Factory class for creating instances that will handle each type of record within a change stream
 query. 
- ActionFactory() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.action.ActionFactory
-  
- activate(MetricsContainer) - Method in class org.apache.beam.sdk.metrics.MetricsEnvironment.MetricsContainerHolder
-  
- activate(MetricsContainer) - Method in interface org.apache.beam.sdk.metrics.MetricsEnvironment.MetricsEnvironmentState
-  
- ACTIVE_PARTITION_READ_COUNT - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
Counter for the active partition reads during the execution of the Connector. 
- actuateProjectionPushdown(Map<TupleTag<?>, FieldAccessDescriptor>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-  
- actuateProjectionPushdown(Map<TupleTag<?>, FieldAccessDescriptor>) - Method in interface org.apache.beam.sdk.schemas.ProjectionProducer
- 
Actuate a projection pushdown. 
- add(MetricsContainerStepMap) - Method in class org.apache.beam.runners.flink.metrics.MetricsAccumulator
-  
- add(Iterable<String>) - Method in class org.apache.beam.runners.jet.metrics.BoundedTrieImpl
-  
- add(String...) - Method in class org.apache.beam.runners.jet.metrics.BoundedTrieImpl
-  
- add(String) - Method in class org.apache.beam.runners.jet.metrics.StringSetImpl
-  
- add(String...) - Method in class org.apache.beam.runners.jet.metrics.StringSetImpl
-  
- add(MetricsContainerStepMap) - Method in class org.apache.beam.runners.spark.metrics.MetricsContainerStepMapAccumulator
-  
- add(MetricsContainerStepMap) - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsAccumulator
-  
- add(int, GlobalWatermarkHolder.SparkWatermarks) - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
-  
- add(T) - Method in class org.apache.beam.runners.twister2.translators.functions.Twister2SinkFunction
-  
- add(long, Instant, boolean) - Method in class org.apache.beam.sdk.extensions.ordered.combiner.SequenceRangeAccumulator
-  
- add(T, long, Coder<T>) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.Sketch
-  
- add(T, Coder<T>) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.Sketch
-  
- add(KV<byte[], byte[]>) - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter
-  
- add(Class<?>, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase.ParameterListBuilder
-  
- add(Type, String, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase.ParameterListBuilder
-  
- add(T, long) - Method in class org.apache.beam.sdk.fn.data.WeightedList
-  
- add(List<ValueInSingleWindow<T>>, TableDataInsertAllResponse.InsertErrors, TableReference, FailsafeValueInSingleWindow<TableRow, TableRow>) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.ErrorContainer
-  
- add(Iterable<String>) - Method in interface org.apache.beam.sdk.metrics.BoundedTrie
- 
Adds a path to the trie. 
- add(String...) - Method in interface org.apache.beam.sdk.metrics.BoundedTrie
- 
Adds a path to the trie. 
- add(String, String, Iterable<String>) - Method in class org.apache.beam.sdk.metrics.Lineage
- 
Add a FQN (fully-qualified name) to Lineage. 
- add(String, Iterable<String>) - Method in class org.apache.beam.sdk.metrics.Lineage
- 
Add a FQN (fully-qualified name) to Lineage. 
- add(String) - Method in class org.apache.beam.sdk.metrics.Lineage
- 
Adds the given details as Lineage. 
- add(String) - Method in interface org.apache.beam.sdk.metrics.StringSet
- 
Add a value to this set. 
- add(String...) - Method in interface org.apache.beam.sdk.metrics.StringSet
- 
Add values to this set. 
- add(InputT) - Method in interface org.apache.beam.sdk.state.GroupingState
- 
Add a value to the buffer. 
- add(Iterable<TimestampedValue<T>>) - Static method in class org.apache.beam.sdk.testing.TestStream.ElementEvent
- 
For internal use only: no backwards compatibility guarantees. 
- add(long) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn.LargestUnique
- 
Deprecated. Adds a value to the heap, returning whether the value is (large enough to be) in the heap. 
- add(DisplayData.ItemSpec<?>) - Method in interface org.apache.beam.sdk.transforms.display.DisplayData.Builder
- 
Register the given display item. 
- addAccum(AccumT) - Method in interface org.apache.beam.sdk.state.CombiningState
- 
Add an accumulator to this state cell. 
- addAll(Map<Integer, Queue<GlobalWatermarkHolder.SparkWatermarks>>) - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
-  
- addAll(WeightedList<T>) - Method in class org.apache.beam.sdk.fn.data.WeightedList
-  
- addAll(List<T>, long) - Method in class org.apache.beam.sdk.fn.data.WeightedList
-  
- addAnnotation(String, byte[]) - Method in class org.apache.beam.sdk.transforms.PTransform
-  
- addArray(Collection<T>) - Method in class org.apache.beam.sdk.values.Row.Builder
-  
- addArray(Object...) - Method in class org.apache.beam.sdk.values.Row.Builder
-  
- addArrayField(String, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addAttempted(T, BiFunction<T, T, T>) - Method in class org.apache.beam.sdk.metrics.MetricResult
-  
- addBatchWriteRequest(long, boolean) - Method in interface org.apache.beam.sdk.io.aws2.common.AsyncBatchWriteHandler.Stats
-  
- addBoolean(Map<String, Object>, String, boolean) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- addBooleanField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addByteArrayField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addByteField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addCoderAndEncodedRecord(Coder<T>, T) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Record.Builder
-  
- addCollectionToSingletonOutput(PCollection<?>, String, PCollectionView<?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
- 
Adds an output to this CollectionToSingletonDataflow step, consuming the specified
 inputPValueand producing the specified outputPValue.
 
- addCommitted(T, BiFunction<T, T, T>) - Method in class org.apache.beam.sdk.metrics.MetricResult
-  
- addDataSet(String, DataSet<T>) - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.BatchTranslationContext
-  
- addDataStream(String, DataStream<T>) - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext
-  
- addDateTimeField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addDecimalField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addDictionary(Map<String, Object>, String, Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- addDouble(Map<String, Object>, String, Double) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- addDoubleField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addElements(T, T...) - Method in class org.apache.beam.sdk.testing.TestStream.Builder
- 
Adds the specified elements to the source with timestamp equal to the current watermark. 
- addElements(TimestampedValue<T>, TimestampedValue<T>...) - Method in class org.apache.beam.sdk.testing.TestStream.Builder
- 
Adds the specified elements to the source with the provided timestamps. 
- addEncodingInput(Coder<?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
- 
Sets the encoding for this Dataflow step. 
- addErrorCollection(PCollection<ErrorT>) - Method in interface org.apache.beam.sdk.transforms.errorhandling.ErrorHandler
-  
- addErrorCollection(PCollection<ErrorT>) - Method in class org.apache.beam.sdk.transforms.errorhandling.ErrorHandler.DefaultErrorHandler
-  
- addErrorCollection(PCollection<ErrorT>) - Method in class org.apache.beam.sdk.transforms.errorhandling.ErrorHandler.PTransformErrorHandler
-  
- addErrorForCode(int, String) - Method in class org.apache.beam.sdk.extensions.gcp.util.CustomHttpErrors.Builder
- 
Adds a matcher to log the provided string if the error matches a particular status code. 
- addErrorForCodeAndUrlContains(int, String, String) - Method in class org.apache.beam.sdk.extensions.gcp.util.CustomHttpErrors.Builder
- 
Adds a matcher to log the provided string if the error matches a particular status code and
 the url contains a certain string. 
- addExceptionStackTrace(Exception) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Failure.Builder
-  
- addExperiment(ExperimentalOptions, String) - Static method in interface org.apache.beam.sdk.options.ExperimentalOptions
- 
Adds experiment to options if not already present. 
- addFailure(String, String) - Method in class org.apache.beam.sdk.io.cdap.context.FailureCollectorWrapper
-  
- addField(Schema.Field) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addField(String, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addFields(List<Schema.Field>) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addFields(Schema.Field...) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- AddFields - Class in org.apache.beam.sdk.schemas.transforms
- 
A transform to add new nullable fields to a PCollection's schema. 
- AddFields() - Constructor for class org.apache.beam.sdk.schemas.transforms.AddFields
-  
- AddFields.Inner<T> - Class in org.apache.beam.sdk.schemas.transforms
- 
Inner PTransform for AddFields. 
- addFloatField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- AddHarnessIdInterceptor - Class in org.apache.beam.sdk.fn.channel
- 
A ClientInterceptorthat attaches a provided SDK Harness ID to outgoing messages.
 
- addHumanReadableJson(Object) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Record.Builder
-  
- addIfAbsent(T) - Method in interface org.apache.beam.sdk.state.SetState
- 
Ensures a value is a member of the set, returning trueif it was added andfalseotherwise.
 
- addIfNotDefault(DisplayData.ItemSpec<T>, T) - Method in interface org.apache.beam.sdk.transforms.display.DisplayData.Builder
- 
Register the given display item if the value is different than the specified default. 
- addIfNotNull(DisplayData.ItemSpec<?>) - Method in interface org.apache.beam.sdk.transforms.display.DisplayData.Builder
- 
Register the given display item if the value is not null. 
- addIncompleteNewPartitions(NewPartition) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.reconciler.OrphanedMetadataCleaner
- 
Add NewPartition if it hasn't been updated for 15 minutes. 
- addIncompleteNewPartitions(NewPartition) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.reconciler.PartitionReconciler
- 
Capture NewPartition row that cannot merge on its own. 
- addInput(String, Boolean) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
- 
Adds an input with the given name and value to this Dataflow step. 
- addInput(String, String) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
- 
Adds an input with the given name and value to this Dataflow step. 
- addInput(String, Long) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
- 
Adds an input with the given name and value to this Dataflow step. 
- addInput(String, PInput) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
- 
Adds an input with the given name to this Dataflow step, coming from the specified input
 PValue. 
- addInput(String, Map<String, Object>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
- 
Adds an input that is a dictionary of strings to objects. 
- addInput(String, List<? extends Map<String, Object>>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
- 
Adds an input that is a list of objects. 
- addInput(Iterable<T>, T) - Method in class org.apache.beam.runners.flink.FlinkStreamingAggregationsTranslators.ConcatenateAsIterable
-  
- addInput(SequenceRangeAccumulator, TimestampedValue<KV<EventKeyT, KV<Long, EventT>>>) - Method in class org.apache.beam.sdk.extensions.ordered.combiner.DefaultSequenceCombiner
-  
- addInput(HyperLogLogPlus, InputT) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.ApproximateDistinctFn
-  
- addInput(SketchFrequencies.Sketch<InputT>, InputT) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.CountMinSketchFn
-  
- addInput(MergingDigest, Double) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.TDigestQuantilesFn
-  
- addInput(AccumT, InputT) - Method in class org.apache.beam.sdk.extensions.sql.impl.LazyAggregateCombineFn
-  
- addInput(long[], Boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CountIf.CountIfFn
-  
- addInput(CovarianceAccumulator, Row) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CovarianceFn
-  
- addInput(VarianceAccumulator, T) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.VarianceFn
-  
- addInput(BeamBuiltinAggregations.BitXOr.Accum, T) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamBuiltinAggregations.BitXOr
-  
- addInput(AccumT, InputT, Long, Long, Long) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamBuiltinAnalyticFunctions.PositionAwareCombineFn
-  
- addInput(AccumT, InputT) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamBuiltinAnalyticFunctions.PositionAwareCombineFn
-  
- addInput(List<T>, T) - Method in class org.apache.beam.sdk.extensions.sql.impl.udaf.ArrayAgg.ArrayAggArray
-  
- addInput(String, byte[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.udaf.StringAgg.StringAggByte
-  
- addInput(String, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udaf.StringAgg.StringAggString
-  
- addInput(Long, Long) - Method in class org.apache.beam.sdk.extensions.sql.provider.UdfTestProvider.Sum
-  
- addInput(AccumT, InputT) - Method in class org.apache.beam.sdk.extensions.sql.TypedCombineFnDelegate
-  
- addInput(AccumT, InputT) - Method in interface org.apache.beam.sdk.extensions.sql.udf.AggregateFn
- 
Adds the given input value to the given accumulator, returning the new accumulator value. 
- addInput(List<String>, String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Concatenate
-  
- addInput(ApproximateUnique.ApproximateUniqueCombineFn.LargestUnique, T) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn
- 
Deprecated. 
- addInput(InputT) - Method in interface org.apache.beam.sdk.transforms.Combine.AccumulatingCombineFn.Accumulator
- 
Adds the given input value to this accumulator, modifying this accumulator. 
- addInput(AccumT, InputT) - Method in class org.apache.beam.sdk.transforms.Combine.AccumulatingCombineFn
-  
- addInput(double[], Double) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
-  
- addInput(Combine.Holder<V>, V) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
-  
- addInput(int[], Integer) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
-  
- addInput(long[], Long) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
-  
- addInput(AccumT, InputT) - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
- 
Adds the given input value to the given accumulator, returning the new accumulator value. 
- addInput(List<V>, V) - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
-  
- addInput(Object[], DataT) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
-  
- addInput(Object[], DataT, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
-  
- addInput(AccumT, InputT, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
- 
Adds the given input value to the given accumulator, returning the new accumulator value. 
- addInput(Top.BoundedHeap<KV<Integer, T>, SerializableComparator<KV<Integer, T>>>, T) - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
-  
- addInt16Field(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addInt32Field(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addInt64Field(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addIterable(Iterable<T>) - Method in class org.apache.beam.sdk.values.Row.Builder
-  
- addIterableField(String, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addKnownCoderUrn(String) - Static method in class org.apache.beam.runners.fnexecution.wire.LengthPrefixUnknownCoders
- 
Registers a coder as being of known type and as such not meriting length prefixing. 
- addLabel(String, String) - Method in class org.apache.beam.sdk.metrics.LabeledMetricNameUtils.MetricNameBuilder
- 
Add a metric label KV pair to the metric name. 
- addLengthPrefixedCoder(String, RunnerApi.Components.Builder, boolean) - Static method in class org.apache.beam.runners.fnexecution.wire.LengthPrefixUnknownCoders
- 
- addList(Map<String, Object>, String, List<? extends Map<String, Object>>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- addList(Map<String, Object>, String, T[]) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- addLogicalTypeConversions(GenericData) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
-  
- addLogicalTypeField(String, Schema.LogicalType<InputT, BaseT>) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addLong(Map<String, Object>, String, long) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- addLongs(Map<String, Object>, String, long...) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- addMapField(String, Schema.FieldType, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addMessageListener(Consumer<JobApi.JobMessage>) - Method in class org.apache.beam.runners.jobsubmission.JobInvocation
- 
Listen for job messages with a Consumer.
 
- addMethodParameters(Method) - Method in class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase.ParameterListBuilder
-  
- addMissingPartitions(List<Range.ByteStringRange>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.reconciler.OrphanedMetadataCleaner
- 
Add all the missingPartitions. 
- addMissingPartitions(List<Range.ByteStringRange>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.reconciler.PartitionReconciler
- 
Capture partitions that are not currently being streamed. 
- addNameFilter(MetricNameFilter) - Method in class org.apache.beam.sdk.metrics.MetricsFilter.Builder
- 
- addNull(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- addNullableArrayField(String, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableBooleanField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableByteArrayField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableByteField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableDateTimeField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableDecimalField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableDoubleField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableField(String, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableFloatField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableInt16Field(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableInt32Field(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableInt64Field(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableIterableField(String, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableLogicalTypeField(String, Schema.LogicalType<InputT, BaseT>) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableMapField(String, Schema.FieldType, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableRowField(String, Schema) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addNullableStringField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addObject(Map<String, Object>, String, Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- addOptions(Schema.Options) - Method in class org.apache.beam.sdk.schemas.Schema.Options.Builder
-  
- addOutput(String, PCollection<?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
- 
Adds a primitive output to this Dataflow step with the given name as the local output name,
 producing the specified output PValue, including itsCoderif aTypedPValue.
 
- addOutput(Output) - Method in class org.apache.beam.sdk.io.cdap.context.BatchSinkContextImpl
- 
Overrides the output configuration of this Batch job to the specified  Output. 
- addOutputColumnList(List<ResolvedNodes.ResolvedOutputColumn>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
-  
- addOverrideForClass(Class<?>, DataflowWorkerLoggingOptions.Level) - Method in class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
- 
Deprecated. Overrides the default log level for the passed in class. 
- addOverrideForClass(Class<?>, SdkHarnessOptions.LogLevel) - Method in class org.apache.beam.sdk.options.SdkHarnessOptions.SdkHarnessLogLevelOverrides
- 
Overrides the default log level for the passed in class. 
- addOverrideForName(String, DataflowWorkerLoggingOptions.Level) - Method in class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
- 
Deprecated. Overrides the default log level for the passed in name. 
- addOverrideForName(String, SdkHarnessOptions.LogLevel) - Method in class org.apache.beam.sdk.options.SdkHarnessOptions.SdkHarnessLogLevelOverrides
- 
Overrides the default log logLevel for the passed in name. 
- addOverrideForPackage(Package, DataflowWorkerLoggingOptions.Level) - Method in class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
- 
Deprecated. Overrides the default log level for the passed in package. 
- addOverrideForPackage(Package, SdkHarnessOptions.LogLevel) - Method in class org.apache.beam.sdk.options.SdkHarnessOptions.SdkHarnessLogLevelOverrides
- 
Overrides the default log level for the passed in package. 
- addProperties(MetadataEntity, Map<String, String>) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- addResolvedTable(TableResolution.SimpleTableWithPath) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
- 
Store a table together with its full path for repeated resolutions. 
- addRowField(String, Schema) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addRows(Object...) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestBoundedTable
- 
Add rows to the builder. 
- addRows(String, Row...) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
-  
- addRows(Duration, Object...) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestUnboundedTable
- 
Add rows to the builder. 
- addRunnerWireCoder(PipelineNode.PCollectionNode, RunnerApi.Components.Builder, RunnerApi.ExecutableStagePayload.WireCoderSetting) - Static method in class org.apache.beam.runners.fnexecution.wire.WireCoders
- 
Creates a runner-side wire coder for a port read/write for the given PCollection. 
- addSchema(String, TableProvider) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
- 
Add a top-level schema backed by the table provider. 
- addSdkWireCoder(PipelineNode.PCollectionNode, RunnerApi.Components.Builder, RunnerApi.ExecutableStagePayload.WireCoderSetting) - Static method in class org.apache.beam.runners.fnexecution.wire.WireCoders
- 
Creates an SDK-side wire coder for a port read/write for the given PCollection. 
- AddShardKeyDoFn - Class in org.apache.beam.sdk.io.solace.write
- 
This class adds pseudo-key with a given cardinality. 
- AddShardKeyDoFn(int) - Constructor for class org.apache.beam.sdk.io.solace.write.AddShardKeyDoFn
-  
- addStateListener(Consumer<JobApi.JobStateEvent>) - Method in class org.apache.beam.runners.jobsubmission.JobInvocation
- 
Listen for job state changes with a Consumer.
 
- addStep(PTransform<?, ?>, String) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
- 
Adds a step to the Dataflow workflow for the given transform, with the given Dataflow step
 type. 
- addStep(String) - Method in class org.apache.beam.sdk.metrics.MetricsFilter.Builder
- 
Add a step filter. 
- addString(Map<String, Object>, String, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- addStringField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- addStringList(Map<String, Object>, String, List<String>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- addTags(MetadataEntity, String...) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- addTags(MetadataEntity, Iterable<String>) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- addTimers(Iterator<TimerInternals.TimerData>) - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-  
- addToCurrentBundle(Solace.Record) - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- addTraceFor(AbstractGoogleClient, String) - Method in class org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions.GoogleApiTracer
- 
- addTraceFor(AbstractGoogleClientRequest<?>, String) - Method in class org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions.GoogleApiTracer
- 
- addUdaf(String, Combine.CombineFn) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
- 
Register a UDAF function which can be used in GROUP-BY expression. 
- addUdf(String, Class<?>, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
- 
Register a UDF function which can be used in SQL expression. 
- addUdf(String, Class<? extends BeamSqlUdf>) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
- 
Register a UDF function which can be used in SQL expression. 
- addUdf(String, SerializableFunction) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
- 
Register a UDF function which can be used in SQL expression. 
- addUuids() - Static method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteIO
- 
Add Uuids to to-be-published messages that ensures that uniqueness is maintained. 
- AddUuidsTransform - Class in org.apache.beam.sdk.io.gcp.pubsublite.internal
- 
A transform to add UUIDs to each message to be written to Pub/Sub Lite. 
- AddUuidsTransform() - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.internal.AddUuidsTransform
-  
- addValue(Object) - Method in class org.apache.beam.sdk.values.Row.Builder
-  
- addValues(List<Object>) - Method in class org.apache.beam.sdk.values.Row.Builder
-  
- addValues(Object...) - Method in class org.apache.beam.sdk.values.Row.Builder
-  
- advance() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-  
- advance() - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
- 
- advance() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
- 
For subscription mode only: Track progression of time according to the Clockpassed .
 
- advance() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.UnboundedReaderImpl
-  
- advance() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-  
- advance() - Method in class org.apache.beam.sdk.io.Source.Reader
- 
Advances the reader to the next valid record. 
- advance() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
- 
Advances the reader to the next valid record. 
- advanceBy(Duration) - Static method in class org.apache.beam.sdk.testing.TestStream.ProcessingTimeEvent
- 
For internal use only: no backwards compatibility guarantees. 
- advanceImpl() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
-  
- advanceImpl() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
- 
Advances to the next record and returns true, or returns false if there is no next
 record.
 
- advanceNextBatchWatermarkToInfinity() - Method in class org.apache.beam.runners.spark.io.CreateStream
- 
Advances the watermark in the next batch to the end-of-time. 
- advanceProcessingTime(Duration) - Method in class org.apache.beam.sdk.testing.TestStream.Builder
- 
Advance the processing time by the specified amount. 
- advanceTo(Instant) - Static method in class org.apache.beam.sdk.testing.TestStream.WatermarkEvent
- 
For internal use only: no backwards compatibility guarantees. 
- advanceWatermark() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
- 
Advances the watermark. 
- advanceWatermarkForNextBatch(Instant) - Method in class org.apache.beam.runners.spark.io.CreateStream
- 
Advances the watermark in the next batch. 
- advanceWatermarkTo(Instant) - Method in class org.apache.beam.sdk.testing.TestStream.Builder
- 
Advance the watermark of this source to the specified instant. 
- advanceWatermarkToInfinity() - Method in class org.apache.beam.sdk.testing.TestStream.Builder
- 
Advance the watermark to infinity, completing this  TestStream. 
- AdvancingPhaser - Class in org.apache.beam.sdk.fn.stream
- 
A Phaserwhich never terminates.
 
- AdvancingPhaser(int) - Constructor for class org.apache.beam.sdk.fn.stream.AdvancingPhaser
-  
- AfterAll - Class in org.apache.beam.sdk.transforms.windowing
- 
A composite  Trigger that fires when all of its sub-triggers are ready. 
- afterBundleCommit(Instant, DoFn.BundleFinalizer.Callback) - Method in interface org.apache.beam.sdk.transforms.DoFn.BundleFinalizer
- 
The provided function will be called after the runner successfully commits the output of a
 successful bundle. 
- AfterEach - Class in org.apache.beam.sdk.transforms.windowing
- 
A composite  Trigger that executes its sub-triggers in order. 
- AfterFirst - Class in org.apache.beam.sdk.transforms.windowing
- 
A composite  Trigger that fires once after at least one of its sub-triggers have fired. 
- afterIterations(int) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
- 
- AfterPane - Class in org.apache.beam.sdk.transforms.windowing
- 
A  Trigger that fires at some point after a specified number of input elements have
 arrived. 
- AfterProcessingTime - Class in org.apache.beam.sdk.transforms.windowing
- 
A  Trigger trigger that fires at a specified point in processing time, relative to when
 input first arrives. 
- AfterSynchronizedProcessingTime - Class in org.apache.beam.sdk.transforms.windowing
- 
FOR INTERNAL USE ONLY. 
- afterTimeSinceNewOutput(ReadableDuration) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
- 
- afterTimeSinceNewOutput(SerializableFunction<InputT, ReadableDuration>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
- 
- afterTotalOf(ReadableDuration) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
- 
- afterTotalOf(SerializableFunction<InputT, ReadableDuration>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
- 
- AfterWatermark - Class in org.apache.beam.sdk.transforms.windowing
- 
AfterWatermarktriggers fire based on progress of the system watermark.
 
- AfterWatermark.AfterWatermarkEarlyAndLate - Class in org.apache.beam.sdk.transforms.windowing
-  
- AfterWatermark.FromEndOfWindow - Class in org.apache.beam.sdk.transforms.windowing
- 
A watermark trigger targeted relative to the end of the window. 
- aggregate(Combine.CombineFn<InputT, ?, OutputT>) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
- 
- AggregateCombiner() - Constructor for class org.apache.beam.sdk.schemas.transforms.Group.AggregateCombiner
-  
- aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.AggregateCombiner
- 
Build up an aggregation function over the input elements. 
- aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.AggregateCombiner
- 
Build up an aggregation function over the input elements. 
- aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
- 
Build up an aggregation function over the input elements. 
- aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
-  
- aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
- 
Build up an aggregation function over the input elements. 
- aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
-  
- aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
- 
Build up an aggregation function over the input elements. 
- aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
-  
- aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
- 
Build up an aggregation function over the input elements. 
- aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
-  
- aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
- 
Build up an aggregation function over the input elements. 
- aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
-  
- aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
- 
Build up an aggregation function over the input elements. 
- aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
-  
- aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
- 
Build up an aggregation function over the input elements. 
- aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
- 
- aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
- 
Build up an aggregation function over the input elements. 
- aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
- 
- aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
-  
- aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
-  
- aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
-  
- aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
-  
- aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
-  
- aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
-  
- aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
-  
- aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
-  
- aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
-  
- aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
-  
- aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
-  
- aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
-  
- aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
-  
- aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
-  
- aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
-  
- aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
-  
- aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
- 
Build up an aggregation function over the input elements. 
- aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
- 
Build up an aggregation function over the input elements. 
- aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
- 
Build up an aggregation function over the input elements. 
- aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
- 
Build up an aggregation function over the input elements. 
- aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
- 
Build up an aggregation function over the input elements. 
- aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
- 
Build up an aggregation function over the input elements. 
- aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
- 
Build up an aggregation function over the input elements. 
- aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
- 
Build up an aggregation function over the input elements. 
- aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
- 
Build up an aggregation function over the input elements. 
- aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
- 
Build up an aggregation function over the input elements. 
- aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
- 
Build up an aggregation function over the input elements. 
- aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
- 
Build up an aggregation function over the input elements. 
- aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
- 
Build up an aggregation function over the input elements. 
- aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
- 
Build up an aggregation function over the input elements. 
- aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
- 
Build up an aggregation function over the input elements. 
- aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
- 
Build up an aggregation function over the input elements. 
- aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.AggregateCombiner
- 
Build up an aggregation function over the input elements by field id. 
- aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
-  
- aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
-  
- aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
-  
- aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
-  
- aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
-  
- aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
- 
- aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
-  
- AggregateFn<InputT,AccumT,OutputT> - Interface in org.apache.beam.sdk.extensions.sql.udf
- 
An aggregate function that can be executed as part of a SQL query. 
- AggregationCombineFnAdapter<T> - Class in org.apache.beam.sdk.extensions.sql.impl.transform.agg
- 
- AggregationCombineFnAdapter() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.agg.AggregationCombineFnAdapter
-  
- AggregationQuery - Class in org.apache.beam.sdk.io.mongodb
- 
Builds a MongoDB AggregateIterable object. 
- AggregationQuery() - Constructor for class org.apache.beam.sdk.io.mongodb.AggregationQuery
-  
- algorithm(String) - Method in class org.apache.beam.sdk.io.aws2.s3.SSECustomerKey.Builder
-  
- align(Duration) - Method in interface org.apache.beam.sdk.state.Timer
- 
- alignedTo(Duration, Instant) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
- 
Aligns timestamps to the smallest multiple of periodsince theoffsetgreater
 than the timestamp.
 
- alignedTo(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
- 
Aligns the time to be the smallest multiple of periodgreater than the epoch boundary
 (akanew Instant(0)).
 
- alignTo(Duration, Instant) - Static method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform
- 
For internal use only; no backwards-compatibility guarantees. 
- alignTo(Duration) - Static method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform
- 
For internal use only; no backwards-compatibility guarantees. 
- AlignTo() - Constructor for class org.apache.beam.sdk.transforms.windowing.TimestampTransform.AlignTo
-  
- ALL_CONTEXTS - Static variable in class org.apache.beam.sdk.testing.CoderProperties
- 
All the contexts, for use in test cases. 
- ALL_KEYS - Static variable in class org.apache.beam.sdk.io.range.ByteKeyRange
- 
The range of all keys, with empty start and end keys. 
- allLeavesDescriptor(Schema, SerializableFunction<List<String>, String>) - Static method in class org.apache.beam.sdk.schemas.utils.SelectHelpers
-  
- allMatches(String) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
- allMatches(Pattern) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
- AllMatches(Pattern) - Constructor for class org.apache.beam.sdk.transforms.Regex.AllMatches
-  
- allMetrics() - Method in class org.apache.beam.sdk.metrics.MetricResults
-  
- allocateAddressAndCreate(List<BindableService>, Endpoints.ApiServiceDescriptor.Builder) - Method in class org.apache.beam.sdk.fn.server.InProcessServerFactory
-  
- allocateAddressAndCreate(List<BindableService>, Endpoints.ApiServiceDescriptor.Builder) - Method in class org.apache.beam.sdk.fn.server.ServerFactory
- 
Creates an instance of this server using an ephemeral address. 
- allocateAddressAndCreate(List<BindableService>, Endpoints.ApiServiceDescriptor.Builder) - Method in class org.apache.beam.sdk.fn.server.ServerFactory.InetSocketAddressServerFactory
-  
- allocatePortAndCreateFor(ServiceT, ServerFactory) - Static method in class org.apache.beam.sdk.fn.server.GrpcFnServer
- 
- allocatePortAndCreateFor(List<? extends FnService>, ServerFactory) - Static method in class org.apache.beam.sdk.fn.server.GrpcFnServer
- 
- allOf(Iterable<SerializableMatcher<? super T>>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- allOf(SerializableMatcher<T>...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- allOf(Watch.Growth.TerminationCondition<InputT, FirstStateT>, Watch.Growth.TerminationCondition<InputT, SecondStateT>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
- 
- ALLOW_DUPLICATES - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- ALLOWS_SHARDABLE_STATE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- allowsDynamicSplitting() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-  
- allowsDynamicSplitting() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
-  
- allowsDynamicSplitting() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
- 
Whether this reader should allow dynamic splitting of the offset ranges. 
- AlwaysPassMatcher() - Constructor for class org.apache.beam.sdk.testing.TestPipelineOptions.AlwaysPassMatcher
-  
- AlwaysPassMatcherFactory() - Constructor for class org.apache.beam.sdk.testing.TestPipelineOptions.AlwaysPassMatcherFactory
-  
- alwaysRetry() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy
- 
Always retry all failures. 
- alwaysUseRead() - Method in class org.apache.beam.sdk.transforms.Create.Values
-  
- AmqpIO - Class in org.apache.beam.sdk.io.amqp
- 
AmqpIO supports AMQP 1.0 protocol using the Apache QPid Proton-J library. 
- AmqpIO.Read - Class in org.apache.beam.sdk.io.amqp
- 
A  PTransform to read/receive messages using AMQP 1.0 protocol. 
- AmqpIO.Write - Class in org.apache.beam.sdk.io.amqp
- 
A  PTransform to send messages using AMQP 1.0 protocol. 
- AmqpMessageCoder - Class in org.apache.beam.sdk.io.amqp
- 
A coder for AMQP message. 
- AmqpMessageCoder() - Constructor for class org.apache.beam.sdk.io.amqp.AmqpMessageCoder
-  
- AmqpMessageCoderProviderRegistrar - Class in org.apache.beam.sdk.io.amqp
- 
- AmqpMessageCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.amqp.AmqpMessageCoderProviderRegistrar
-  
- and(TupleTag<V>, List<V>) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
- 
Returns a new  CoGbkResult based on this, with the given tag and given data added to it. 
- and(TupleTag<V>, PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
- 
Returns a new KeyedPCollectionTuple<K>that is the same as this, appended with the
 given PCollection.
 
- and(String, PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
- 
- and(PCollection.IsBounded) - Method in enum org.apache.beam.sdk.values.PCollection.IsBounded
- 
Returns the composed IsBounded property. 
- and(PCollection<T>) - Method in class org.apache.beam.sdk.values.PCollectionList
- 
- and(Iterable<PCollection<T>>) - Method in class org.apache.beam.sdk.values.PCollectionList
- 
- and(String, PCollection<Row>) - Method in class org.apache.beam.sdk.values.PCollectionRowTuple
- 
- and(TupleTag<T>, PCollection<T>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
- 
- and(String, PCollection<T>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
- 
- and(TupleTag<?>) - Method in class org.apache.beam.sdk.values.TupleTagList
- 
- and(List<TupleTag<?>>) - Method in class org.apache.beam.sdk.values.TupleTagList
- 
- annotateFromBytes(PCollectionView<Map<ByteString, VideoContext>>, List<Feature>) - Static method in class org.apache.beam.sdk.extensions.ml.VideoIntelligence
- 
Annotates videos from ByteStrings of their contents. 
- annotateFromBytesWithContext(List<Feature>) - Static method in class org.apache.beam.sdk.extensions.ml.VideoIntelligence
- 
Annotates videos from key-value pairs of ByteStrings and VideoContext. 
- annotateFromURI(List<Feature>, PCollectionView<Map<String, VideoContext>>) - Static method in class org.apache.beam.sdk.extensions.ml.VideoIntelligence
- 
Annotates videos from GCS URIs. 
- annotateFromUriWithContext(List<Feature>) - Static method in class org.apache.beam.sdk.extensions.ml.VideoIntelligence
- 
Annotates videos from key-value pairs of GCS URI and VideoContext. 
- annotateImagesFromBytes(PCollectionView<Map<ByteString, ImageContext>>, List<Feature>, long, int) - Static method in class org.apache.beam.sdk.extensions.ml.CloudVision
- 
Creates a  PTransform that annotates images from their
 contents encoded in  ByteStrings. 
- annotateImagesFromBytes(PCollectionView<Map<ByteString, ImageContext>>, List<Feature>, long) - Static method in class org.apache.beam.sdk.extensions.ml.CloudVision
- 
Creates a  PTransform that annotates images from their
 contents encoded in  ByteStrings. 
- AnnotateImagesFromBytes(PCollectionView<Map<ByteString, ImageContext>>, List<Feature>, long, int) - Constructor for class org.apache.beam.sdk.extensions.ml.CloudVision.AnnotateImagesFromBytes
-  
- annotateImagesFromBytesWithContext(List<Feature>, long, int) - Static method in class org.apache.beam.sdk.extensions.ml.CloudVision
- 
Creates a  PTransform that annotates images from KVs of
 their GCS addresses in Strings and  ImageContext for each image. 
- annotateImagesFromBytesWithContext(List<Feature>, long) - Static method in class org.apache.beam.sdk.extensions.ml.CloudVision
- 
Creates a  PTransform that annotates images from KVs of
 their GCS addresses in Strings and  ImageContext for each image. 
- AnnotateImagesFromBytesWithContext(List<Feature>, long, int) - Constructor for class org.apache.beam.sdk.extensions.ml.CloudVision.AnnotateImagesFromBytesWithContext
-  
- annotateImagesFromGcsUri(PCollectionView<Map<String, ImageContext>>, List<Feature>, long, int) - Static method in class org.apache.beam.sdk.extensions.ml.CloudVision
- 
Creates a  PTransform that annotates images from their
 GCS addresses. 
- annotateImagesFromGcsUri(PCollectionView<Map<String, ImageContext>>, List<Feature>, long) - Static method in class org.apache.beam.sdk.extensions.ml.CloudVision
- 
Creates a  PTransform that annotates images from their
 GCS addresses. 
- AnnotateImagesFromGcsUri(PCollectionView<Map<String, ImageContext>>, List<Feature>, long, int) - Constructor for class org.apache.beam.sdk.extensions.ml.CloudVision.AnnotateImagesFromGcsUri
-  
- annotateImagesFromGcsUriWithContext(List<Feature>, long, int) - Static method in class org.apache.beam.sdk.extensions.ml.CloudVision
- 
Creates a  PTransform that annotates images from KVs of
 their String-encoded contents and  ImageContext for each image. 
- annotateImagesFromGcsUriWithContext(List<Feature>, long) - Static method in class org.apache.beam.sdk.extensions.ml.CloudVision
- 
Creates a  PTransform that annotates images from KVs of
 their String-encoded contents and  ImageContext for each image. 
- AnnotateImagesFromGcsUriWithContext(List<Feature>, long, int) - Constructor for class org.apache.beam.sdk.extensions.ml.CloudVision.AnnotateImagesFromGcsUriWithContext
-  
- AnnotateText - Class in org.apache.beam.sdk.extensions.ml
- 
A  PTransform using the Cloud AI Natural language processing capability. 
- AnnotateText() - Constructor for class org.apache.beam.sdk.extensions.ml.AnnotateText
-  
- AnnotateText.Builder - Class in org.apache.beam.sdk.extensions.ml
-  
- AnnotateVideoFromBytes(PCollectionView<Map<ByteString, VideoContext>>, List<Feature>) - Constructor for class org.apache.beam.sdk.extensions.ml.VideoIntelligence.AnnotateVideoFromBytes
-  
- AnnotateVideoFromBytesWithContext(List<Feature>) - Constructor for class org.apache.beam.sdk.extensions.ml.VideoIntelligence.AnnotateVideoFromBytesWithContext
-  
- AnnotateVideoFromUri(PCollectionView<Map<String, VideoContext>>, List<Feature>) - Constructor for class org.apache.beam.sdk.extensions.ml.VideoIntelligence.AnnotateVideoFromUri
-  
- AnnotateVideoFromURIWithContext(List<Feature>) - Constructor for class org.apache.beam.sdk.extensions.ml.VideoIntelligence.AnnotateVideoFromURIWithContext
-  
- annotations - Variable in class org.apache.beam.sdk.transforms.PTransform
-  
- any(long) - Static method in class org.apache.beam.sdk.transforms.Sample
- 
Sample#any(long)takes aPCollection<T>and a limit, and produces a newPCollection<T>containing up to limit elements of the inputPCollection.
 
- anyCombineFn(int) - Static method in class org.apache.beam.sdk.transforms.Sample
- 
Returns a  Combine.CombineFn that computes a fixed-sized potentially non-uniform sample of its
 inputs. 
- anyOf(Iterable<SerializableMatcher<? super T>>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- anyOf(SerializableMatcher<T>...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- anything() - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- anyValueCombineFn() - Static method in class org.apache.beam.sdk.transforms.Sample
- 
Returns a  Combine.CombineFn that computes a single and potentially non-uniform sample value of
 its inputs. 
- API_METRIC_LABEL - Static variable in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl
-  
- ApiIOError - Class in org.apache.beam.io.requestresponse
- 
ApiIOError is a data class for storing details about an error. 
- ApiIOError() - Constructor for class org.apache.beam.io.requestresponse.ApiIOError
-  
- append(K, W, Iterator<V>) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.BagUserStateHandler
- 
Appends the values to the bag user state for the given key and window. 
- appendRows(long, ProtoRows) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StreamAppendClient
- 
Append rows to a Storage API write stream at the given offset. 
- appendRowsRowStatusCounter(BigQuerySinkMetrics.RowStatus, String, String) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySinkMetrics
-  
- applicableTo(PCollection<?>) - Static method in class org.apache.beam.sdk.transforms.GroupByKey
-  
- ApplicationNameOptions - Interface in org.apache.beam.sdk.options
- 
Options that allow setting the application name. 
- apply(Tuple2<ByteArray, byte[]>) - Method in class org.apache.beam.runners.spark.coders.CoderHelpers.FromByteFunction
-  
- apply(KV<String, Long>) - Method in class org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.FormatAsTextFn
-  
- apply(InputT) - Method in interface org.apache.beam.sdk.coders.DelegateCoder.CodingFunction
-  
- apply(Pipeline, String, RunnerApi.FunctionSpec, Map<String, PCollection<?>>) - Method in interface org.apache.beam.sdk.expansion.service.TransformProvider
-  
- apply(Schema, Schema) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory.GenericDatumFactory
-  
- apply(Schema) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory.GenericDatumFactory
-  
- apply(Schema, Schema) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory.ReflectDatumFactory
-  
- apply(Schema) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory.ReflectDatumFactory
-  
- apply(Schema, Schema) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory.SpecificDatumFactory
-  
- apply(Schema) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory.SpecificDatumFactory
-  
- apply(Schema) - Method in interface org.apache.beam.sdk.extensions.avro.io.AvroSink.DatumWriterFactory
-  
- apply(Schema, Schema) - Method in interface org.apache.beam.sdk.extensions.avro.io.AvroSource.DatumReaderFactory
-  
- apply(Row) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSetOperatorsTransforms.BeamSqlRow2KvFn
-  
- apply(Iterable<? extends InputT>) - Method in class org.apache.beam.sdk.extensions.sql.TypedCombineFnDelegate
-  
- apply(T1, T2) - Method in interface org.apache.beam.sdk.function.ThrowingBiFunction
-  
- apply(T1) - Method in interface org.apache.beam.sdk.function.ThrowingFunction
-  
- apply(String, T) - Method in interface org.apache.beam.sdk.io.aws2.sqs.SqsIO.WriteBatches.EntryMapperFn.Builder
-  
- apply(Statement, Description) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery
-  
- apply(Row) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BeamRowToBigtableMutation.ToBigtableRowFn
-  
- apply(Row) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProvider.BigtableRowToBeamRow
-  
- apply(Row) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProvider.GetMutationsFromBeamRow
-  
- apply(HealthcareIOError<T>) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HealthcareIOErrorToTableRow
-  
- apply(ValueInSingleWindow<byte[]>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite.ParsePubsubMessageProtoAsPayloadFromWindowedValue
-  
- apply(byte[]) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessages.DeserializeBytesIntoPubsubMessagePayloadOnly
-  
- apply(PubsubMessage) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessages.ParsePayloadAsPubsubMessageProto
-  
- apply(byte[]) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessages.ParsePubsubMessageProtoAsPayload
-  
- apply(Statement, Description) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
-  
- apply(Statement, Description) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal
-  
- apply(Void) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceProviderFromDataSourceConfiguration
-  
- apply(SQLException) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DefaultRetryStrategy
-  
- apply(Void) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.PoolableDataSourceProvider
-  
- apply(SQLException) - Method in interface org.apache.beam.sdk.io.jdbc.JdbcIO.RetryStrategy
-  
- apply(String, Session) - Method in class org.apache.beam.sdk.io.jms.TextMessageMapper
-  
- apply(TopicPartition) - Method in class org.apache.beam.sdk.io.kafka.CheckStopReadingFnWrapper
-  
- apply(MongoCollection<Document>) - Method in class org.apache.beam.sdk.io.mongodb.AggregationQuery
-  
- apply(MongoCollection<Document>) - Method in class org.apache.beam.sdk.io.mongodb.FindQuery
-  
- apply(Void) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverProviderFromDriverConfiguration
-  
- apply(FileIO.ReadableFile, OffsetRange, Exception) - Method in class org.apache.beam.sdk.io.ReadAllViaFileBasedSource.ReadFileRangesFnExceptionHandler
-  
- apply(Void) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceProviderFromDataSourceConfiguration
-  
- apply(PTransform<? super PBegin, OutputT>) - Method in class org.apache.beam.sdk.Pipeline
- 
- apply(String, PTransform<? super PBegin, OutputT>) - Method in class org.apache.beam.sdk.Pipeline
- 
- apply(Schema, Schema) - Method in class org.apache.beam.sdk.schemas.transforms.Cast.Narrowing
-  
- apply(Schema, Schema) - Method in interface org.apache.beam.sdk.schemas.transforms.Cast.Validator
-  
- apply(Schema, Schema) - Method in class org.apache.beam.sdk.schemas.transforms.Cast.Widening
-  
- apply(Schema, Schema) - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold
-  
- apply(T) - Method in class org.apache.beam.sdk.testing.PAssert.MatcherCheckerFn
-  
- apply(Statement, Description) - Method in class org.apache.beam.sdk.testing.TestPipeline
-  
- apply(double, double) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
- 
Applies the binary operation to the two operands, returning the result. 
- apply(V, V) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
- 
Applies the binary operation to the two operands, returning the result. 
- apply(int, int) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
- 
Applies the binary operation to the two operands, returning the result. 
- apply(long, long) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
- 
Applies the binary operation to the two operands, returning the result. 
- apply(Iterable<? extends InputT>) - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
- 
Applies this CombineFnto a collection of input values to produce a combined output
 value.
 
- apply(Iterable<? extends InputT>, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
- 
Applies this CombineFnWithContextto a collection of input values to produce a
 combined output value.
 
- apply(InputT, Contextful.Fn.Context) - Method in interface org.apache.beam.sdk.transforms.Contextful.Fn
- 
Invokes the function on the given input with the given context. 
- apply(InputT) - Method in class org.apache.beam.sdk.transforms.InferableFunction
-  
- apply(PTransform<KeyedPCollectionTuple<K>, OutputT>) - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
- 
- apply(String, PTransform<KeyedPCollectionTuple<K>, OutputT>) - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
- 
Applies the given  PTransform to this input  KeyedPCollectionTuple and returns
 its  OutputT. 
- apply(InputT) - Method in interface org.apache.beam.sdk.transforms.ProcessFunction
- 
Returns the result of invoking this function on the given input. 
- apply(InputT) - Method in interface org.apache.beam.sdk.transforms.SerializableFunction
- 
Returns the result of invoking this function on the given input. 
- apply(InputT) - Method in class org.apache.beam.sdk.transforms.SimpleFunction
-  
- apply(PrimitiveViewT) - Method in class org.apache.beam.sdk.transforms.ViewFn
- 
A function to adapt a primitive view type to a desired view type. 
- apply(WithFailures.ExceptionElement<T>) - Method in class org.apache.beam.sdk.transforms.WithFailures.ExceptionAsMapHandler
-  
- apply(WithFailures.ExceptionElement<T>) - Method in class org.apache.beam.sdk.transforms.WithFailures.ThrowableHandler
-  
- apply(PTransform<? super PBegin, OutputT>) - Method in class org.apache.beam.sdk.values.PBegin
- 
- apply(String, PTransform<? super PBegin, OutputT>) - Method in class org.apache.beam.sdk.values.PBegin
- 
Applies the given  PTransform to this  PBegin, using  name to identify
 this specific application of the transform. 
- apply(PTransform<? super PCollection<T>, OutputT>) - Method in class org.apache.beam.sdk.values.PCollection
- 
- apply(String, PTransform<? super PCollection<T>, OutputT>) - Method in class org.apache.beam.sdk.values.PCollection
- 
Applies the given  PTransform to this input  PCollection, using  name to
 identify this specific application of the transform. 
- apply(PTransform<PCollectionList<T>, OutputT>) - Method in class org.apache.beam.sdk.values.PCollectionList
- 
- apply(String, PTransform<PCollectionList<T>, OutputT>) - Method in class org.apache.beam.sdk.values.PCollectionList
- 
Applies the given  PTransform to this input  PCollectionList, using  name
 to identify this specific application of the transform. 
- apply(PTransform<? super PCollectionRowTuple, OutputT>) - Method in class org.apache.beam.sdk.values.PCollectionRowTuple
- 
- apply(String, PTransform<? super PCollectionRowTuple, OutputT>) - Method in class org.apache.beam.sdk.values.PCollectionRowTuple
- 
- apply(PTransform<? super PCollectionTuple, OutputT>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
- 
- apply(String, PTransform<? super PCollectionTuple, OutputT>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
- 
- apply(Materializations.MultimapView<Void, T>) - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryListFromMultimapViewFn
-  
- apply(Materializations.IterableView<T>) - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryListViewFn
-  
- apply(Materializations.MultimapView<Void, KV<K, V>>) - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryMapFromVoidKeyViewFn
-  
- apply(Materializations.IterableView<KV<K, V>>) - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryMapViewFn
-  
- apply(Materializations.MultimapView<Void, KV<K, V>>) - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryMultimapFromVoidKeyViewFn
-  
- apply(Materializations.IterableView<KV<K, V>>) - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryMultimapViewFn
-  
- apply(Materializations.IterableView<T>) - Method in class org.apache.beam.sdk.values.PCollectionViews.IterableBackedListViewFn
-  
- apply(Materializations.MultimapView<Void, T>) - Method in class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn
- 
Deprecated. 
- apply(Materializations.IterableView<T>) - Method in class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn2
-  
- apply(Materializations.MultimapView<Void, T>) - Method in class org.apache.beam.sdk.values.PCollectionViews.ListViewFn
- 
Deprecated. 
- apply(Materializations.MultimapView<Long, PCollectionViews.ValueOrMetadata<T, OffsetRange>>) - Method in class org.apache.beam.sdk.values.PCollectionViews.ListViewFn2
-  
- apply(Materializations.MultimapView<Void, KV<K, V>>) - Method in class org.apache.beam.sdk.values.PCollectionViews.MapViewFn
- 
Deprecated. 
- apply(Materializations.MultimapView<K, V>) - Method in class org.apache.beam.sdk.values.PCollectionViews.MapViewFn2
-  
- apply(Materializations.MultimapView<Void, KV<K, V>>) - Method in class org.apache.beam.sdk.values.PCollectionViews.MultimapViewFn
- 
Deprecated. 
- apply(Materializations.MultimapView<K, V>) - Method in class org.apache.beam.sdk.values.PCollectionViews.MultimapViewFn2
-  
- apply(Materializations.MultimapView<Void, T>) - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn
- 
Deprecated. 
- apply(Materializations.IterableView<T>) - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn2
-  
- applyBeamPTransform(DataSet<InputT>, PTransform<CollectionT, PCollection<OutputT>>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataSetAdapter
-  
- applyBeamPTransform(Map<String, ? extends DataSet<?>>, PTransform<PCollectionTuple, PCollection<OutputT>>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataSetAdapter
-  
- applyBeamPTransform(ExecutionEnvironment, PTransform<PBegin, PCollection<OutputT>>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataSetAdapter
-  
- applyBeamPTransform(DataStream<InputT>, PTransform<CollectionT, PCollection<OutputT>>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataStreamAdapter
-  
- applyBeamPTransform(Map<String, ? extends DataStream<?>>, PTransform<PCollectionTuple, PCollection<OutputT>>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataStreamAdapter
-  
- applyBeamPTransform(StreamExecutionEnvironment, PTransform<PBegin, PCollection<OutputT>>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataStreamAdapter
-  
- applyMultiOutputBeamPTransform(DataSet<InputT>, PTransform<CollectionT, PCollectionTuple>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataSetAdapter
-  
- applyMultiOutputBeamPTransform(Map<String, ? extends DataSet<?>>, PTransform<PCollectionTuple, PCollectionTuple>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataSetAdapter
-  
- applyMultiOutputBeamPTransform(ExecutionEnvironment, PTransform<PBegin, PCollectionTuple>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataSetAdapter
-  
- applyMultiOutputBeamPTransform(DataStream<InputT>, PTransform<CollectionT, PCollectionTuple>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataStreamAdapter
-  
- applyMultiOutputBeamPTransform(Map<String, ? extends DataStream<?>>, PTransform<PCollectionTuple, PCollectionTuple>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataStreamAdapter
-  
- applyMultiOutputBeamPTransform(StreamExecutionEnvironment, PTransform<PBegin, PCollectionTuple>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataStreamAdapter
-  
- applyNoOutputBeamPTransform(DataSet<InputT>, PTransform<CollectionT, PDone>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataSetAdapter
-  
- applyNoOutputBeamPTransform(Map<String, ? extends DataSet<?>>, PTransform<PCollectionTuple, PDone>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataSetAdapter
-  
- applyNoOutputBeamPTransform(ExecutionEnvironment, PTransform<PBegin, PDone>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataSetAdapter
-  
- applyNoOutputBeamPTransform(DataStream<InputT>, PTransform<CollectionT, PDone>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataStreamAdapter
-  
- applyNoOutputBeamPTransform(Map<String, ? extends DataStream<?>>, PTransform<PCollectionTuple, PDone>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataStreamAdapter
-  
- applyNoOutputBeamPTransform(StreamExecutionEnvironment, PTransform<PBegin, PDone>) - Method in class org.apache.beam.runners.flink.adapter.BeamFlinkDataStreamAdapter
-  
- applyRowMutations() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
- 
- applySdkEnvironmentOverrides(RunnerApi.Pipeline, DataflowPipelineOptions) - Method in class org.apache.beam.runners.dataflow.DataflowRunner
-  
- applyTransform(InputT, PTransform<? super InputT, OutputT>) - Static method in class org.apache.beam.sdk.Pipeline
- 
For internal use only; no backwards-compatibility guarantees. 
- applyTransform(String, InputT, PTransform<? super InputT, OutputT>) - Static method in class org.apache.beam.sdk.Pipeline
- 
For internal use only; no backwards-compatibility guarantees. 
- applyWindowing() - Method in class org.apache.beam.sdk.transforms.PeriodicImpulse
-  
- ApproximateCountDistinct - Class in org.apache.beam.sdk.extensions.zetasketch
- 
PTransforms for estimating the number of distinct elements in aPCollection, or
 the number of distinct values associated with each key in aPCollectionofKVs.
 
- ApproximateCountDistinct() - Constructor for class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct
-  
- ApproximateCountDistinct.Globally<T> - Class in org.apache.beam.sdk.extensions.zetasketch
- 
PTransformfor estimating the number of distinct elements in aPCollection.
 
- ApproximateCountDistinct.Globally.Builder<T> - Class in org.apache.beam.sdk.extensions.zetasketch
-  
- ApproximateCountDistinct.PerKey<K,V> - Class in org.apache.beam.sdk.extensions.zetasketch
-  
- ApproximateCountDistinct.PerKey.Builder<K,V> - Class in org.apache.beam.sdk.extensions.zetasketch
-  
- ApproximateDistinct - Class in org.apache.beam.sdk.extensions.sketching
- 
PTransforms for computing the approximate number of distinct elements in a stream. 
- ApproximateDistinct() - Constructor for class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct
-  
- ApproximateDistinct.ApproximateDistinctFn<InputT> - Class in org.apache.beam.sdk.extensions.sketching
- 
- ApproximateDistinct.GloballyDistinct<InputT> - Class in org.apache.beam.sdk.extensions.sketching
- 
- ApproximateDistinct.HyperLogLogPlusCoder - Class in org.apache.beam.sdk.extensions.sketching
- 
Coder for HyperLogLogPlusclass.
 
- ApproximateDistinct.PerKeyDistinct<K,V> - Class in org.apache.beam.sdk.extensions.sketching
- 
- ApproximateQuantiles - Class in org.apache.beam.sdk.transforms
- 
PTransforms for getting an idea of aPCollection's data distribution using
 approximateN-tiles (e.g.
 
- ApproximateQuantiles.ApproximateQuantilesCombineFn<T,ComparatorT extends java.util.Comparator<T> & java.io.Serializable> - Class in org.apache.beam.sdk.transforms
- 
The ApproximateQuantilesCombineFncombiner gives an idea of the distribution of a
 collection of values using approximateN-tiles.
 
- ApproximateUnique - Class in org.apache.beam.sdk.transforms
- 
- ApproximateUnique() - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique
- 
Deprecated. 
- ApproximateUnique.ApproximateUniqueCombineFn<T> - Class in org.apache.beam.sdk.transforms
- 
Deprecated. CombineFnthat computes an estimate of the number of distinct values that were
 combined.
 
- ApproximateUnique.ApproximateUniqueCombineFn.LargestUnique - Class in org.apache.beam.sdk.transforms
- 
Deprecated. A heap utility class to efficiently track the largest added elements. 
- ApproximateUnique.Globally<T> - Class in org.apache.beam.sdk.transforms
- 
Deprecated. PTransformfor estimating the number of distinct elements in aPCollection.
 
- ApproximateUnique.PerKey<K,V> - Class in org.apache.beam.sdk.transforms
- 
Deprecated. PTransformfor estimating the number of distinct values associated with each key in aPCollectionofKVs.
 
- ApproximateUniqueCombineFn(long, Coder<T>) - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn
- 
Deprecated. 
- arbitrarily() - Static method in class org.apache.beam.sdk.transforms.Redistribute
-  
- array() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
- 
Returns the backing array. 
- array(TableSchema.ColumnType) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- array(Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
Create an array type for the given field type. 
- array(Schema.FieldType, boolean) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
- ARRAY_AGG_FN - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- ArrayAgg - Class in org.apache.beam.sdk.extensions.sql.impl.udaf
-  
- ArrayAgg() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.udaf.ArrayAgg
-  
- ArrayAgg.ArrayAggArray<T> - Class in org.apache.beam.sdk.extensions.sql.impl.udaf
-  
- ArrayAggArray() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.udaf.ArrayAgg.ArrayAggArray
-  
- arrayContaining(T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- arrayContaining(Coder<T>, T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- arrayContaining(SerializableMatcher<? super T>...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- arrayContaining(List<SerializableMatcher<? super T>>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- arrayContainingInAnyOrder(T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- arrayContainingInAnyOrder(Coder<T>, T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- arrayContainingInAnyOrder(SerializableMatcher<? super T>...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- arrayContainingInAnyOrder(Collection<SerializableMatcher<? super T>>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- ArrayCopyState() - Constructor for class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark.NewVsCopy.ArrayCopyState
-  
- arrayElementType() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- ArrayNewState() - Constructor for class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark.NewVsCopy.ArrayNewState
-  
- ArrayOfNestedStringBundle() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.ArrayOfNestedStringBundle
-  
- ArrayOfStringBundle() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.ArrayOfStringBundle
-  
- arrayQualifier() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- arrayQualifier() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierListContext
-  
- ArrayQualifierContext(ParserRuleContext, int) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierContext
-  
- ArrayQualifierListContext(FieldSpecifierNotationParser.QualifierListContext) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierListContext
-  
- arrayWithSize(int) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- arrayWithSize(SerializableMatcher<? super Integer>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- ArrowConversion - Class in org.apache.beam.sdk.extensions.arrow
- 
Utilities to create  Iterables of Beam  Row instances backed by Arrow record
 batches. 
- ArrowConversion.ArrowSchemaTranslator - Class in org.apache.beam.sdk.extensions.arrow
- 
Converts Arrow schema to Beam row schema. 
- ArrowConversion.RecordBatchRowIterator - Class in org.apache.beam.sdk.extensions.arrow
-  
- arrowSchemaFromInput(InputStream) - Static method in class org.apache.beam.sdk.extensions.arrow.ArrowConversion
-  
- ArrowSchemaTranslator() - Constructor for class org.apache.beam.sdk.extensions.arrow.ArrowConversion.ArrowSchemaTranslator
-  
- ArtifactDestination() - Constructor for class org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService.ArtifactDestination
-  
- ArtifactRetrievalService - Class in org.apache.beam.runners.fnexecution.artifact
- 
- ArtifactRetrievalService() - Constructor for class org.apache.beam.runners.fnexecution.artifact.ArtifactRetrievalService
-  
- ArtifactRetrievalService(ArtifactResolver) - Constructor for class org.apache.beam.runners.fnexecution.artifact.ArtifactRetrievalService
-  
- ArtifactRetrievalService(int) - Constructor for class org.apache.beam.runners.fnexecution.artifact.ArtifactRetrievalService
-  
- ArtifactRetrievalService(ArtifactResolver, int) - Constructor for class org.apache.beam.runners.fnexecution.artifact.ArtifactRetrievalService
-  
- ArtifactStagingService - Class in org.apache.beam.runners.fnexecution.artifact
-  
- ArtifactStagingService(ArtifactStagingService.ArtifactDestinationProvider) - Constructor for class org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService
-  
- ArtifactStagingService.ArtifactDestination - Class in org.apache.beam.runners.fnexecution.artifact
- 
A pairing of a newly created artifact type and an output stream that will be readable at that
 type. 
- ArtifactStagingService.ArtifactDestinationProvider - Interface in org.apache.beam.runners.fnexecution.artifact
- 
Provides a concrete location to which artifacts can be staged on retrieval. 
- as(Class<T>) - Method in interface org.apache.beam.sdk.options.PipelineOptions
- 
Transforms this object into an object of type <T>saving each property that has been
 manipulated.
 
- as(Class<T>) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
- 
Creates and returns an object that implements <T>.
 
- as(Class<T>) - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
- 
Creates and returns an object that implements <T>using the values configured on this
 builder during construction.
 
- asCloudObject(Coder<?>, SdkComponents) - Static method in class org.apache.beam.runners.dataflow.util.CloudObjects
- 
- asInputStream(int, int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
- 
Returns an InputStreamwrapper which supplies the portion of this backing byte buffer
 starting atoffsetand up tolengthbytes.
 
- asIterable() - Static method in class org.apache.beam.sdk.transforms.View
- 
- AsJsons<InputT> - Class in org.apache.beam.sdk.extensions.jackson
- 
PTransform for serializing objects to JSON  Strings. 
- AsJsons.AsJsonsWithFailures<FailureT> - Class in org.apache.beam.sdk.extensions.jackson
- 
A  PTransform that adds exception handling to  AsJsons. 
- asList() - Static method in class org.apache.beam.sdk.transforms.View
- 
- asMap() - Method in class org.apache.beam.sdk.transforms.display.DisplayData
-  
- asMap() - Static method in class org.apache.beam.sdk.transforms.View
- 
- asMultimap() - Static method in class org.apache.beam.sdk.transforms.View
- 
- asOutputReference(PValue, AppliedPTransform<?, ?, ?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
- 
Encode a PValue reference as an output reference. 
- asOutputStream() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
- 
Returns an output stream which writes to the backing buffer from the current position. 
- asPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription
- 
Returns the string representation of this subscription as a path used in the Cloud Pub/Sub
 API. 
- asPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic
- 
Returns the string representation of this topic as a path used in the Cloud Pub/Sub API. 
- asQueryable(QueryProvider, SchemaPlus, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteTable
-  
- asResponseObserver() - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClient
-  
- assertionError() - Method in class org.apache.beam.sdk.testing.SuccessOrFailure
-  
- assertSourcesEqualReferenceSource(BoundedSource<T>, List<? extends BoundedSource<T>>, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
- 
Given a reference Sourceand a list ofSources, assert that the union of the
 records read from the list of sources is equal to the records read from the reference source.
 
- assertSplitAtFractionBehavior(BoundedSource<T>, int, double, SourceTestUtils.ExpectedSplitOutcome, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
- 
- assertSplitAtFractionExhaustive(BoundedSource<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
- 
Asserts that for each possible start position, BoundedSource.BoundedReader#splitAtFractionat every interesting fraction (halfway between two
 fractions that differ by at least one item) can be called successfully and the results are
 consistent if a split succeeds.
 
- assertSplitAtFractionFails(BoundedSource<T>, int, double, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
- 
Asserts that the source's reader fails tosplitAtFraction(fraction)after
 readingnumItemsToReadBeforeSplititems.
 
- assertSplitAtFractionSucceedsAndConsistent(BoundedSource<T>, int, double, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
- 
Verifies some consistency properties of BoundedSource.BoundedReader#splitAtFractionon
 the given source.
 
- assertSubscriptionEventuallyCreated(String, Duration) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
- 
Block until a subscription is created for this test topic in the specified project. 
- assertThatAllRows(Schema) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery
-  
- assertThatTopicEventuallyReceives(Matcher<PubsubMessage>...) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
- 
Repeatedly pull messages from  TestPubsub.subscriptionPath() until receiving one for each matcher
 (or timeout is reached), then assert that the received messages match the expectations. 
- assertUnstartedReaderReadsSameAsItsSource(BoundedSource.BoundedReader<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
- 
Assert that a Readerreturns aSourcethat, when read from, produces the same
 records as the reader.
 
- assign(BoundedWindow, Instant) - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
- 
- assignableTo(Schema) - Method in class org.apache.beam.sdk.schemas.Schema
- 
Returns true if this Schema can be assigned to another Schema. 
- assignableToIgnoreNullable(Schema) - Method in class org.apache.beam.sdk.schemas.Schema
- 
Returns true if this Schema can be assigned to another Schema, ignoring nullable. 
- AssignContext() - Constructor for class org.apache.beam.sdk.transforms.windowing.WindowFn.AssignContext
-  
- assignedWindows(WindowFn<T, W>, long) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
- 
- assignedWindowsWithValue(WindowFn<T, W>, TimestampedValue<T>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
- 
- AssignShardFn(Integer) - Constructor for class org.apache.beam.sdk.transforms.Reshuffle.AssignShardFn
-  
- assignShardKey(DestinationT, UserT, int) - Method in interface org.apache.beam.sdk.io.ShardingFunction
-  
- assignsToOneWindow() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-  
- assignsToOneWindow() - Method in class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
-  
- assignsToOneWindow() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-  
- assignsToOneWindow() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
- 
Returns true if this  WindowFn always assigns an element to exactly one window. 
- assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-  
- assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-  
- assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-  
- assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-  
- assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
- 
Returns the single window to which elements with this timestamp belong. 
- AssignWindowP<T> - Class in org.apache.beam.runners.jet.processors
- 
/** * Jet Processorimplementation for Beam's Windowing primitive.
 
- assignWindows(WindowFn<Object, IntervalWindow>.AssignContext) - Method in class org.apache.beam.sdk.extensions.sql.impl.TVFSlidingWindowFn
-  
- assignWindows(WindowFn<Object, GlobalWindow>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-  
- assignWindows(WindowFn<T, W>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
-  
- assignWindows(WindowFn<Object, IntervalWindow>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-  
- assignWindows(WindowFn<Object, IntervalWindow>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-  
- assignWindows(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-  
- assignWindows(WindowFn<T, W>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
- 
Given a timestamp and element, returns the set of windows into which it should be placed. 
- AssignWindowsFunction<T> - Class in org.apache.beam.runners.twister2.translators.functions
- 
Assign Windows function. 
- AssignWindowsFunction(WindowFn<T, BoundedWindow>, PipelineOptions) - Constructor for class org.apache.beam.runners.twister2.translators.functions.AssignWindowsFunction
-  
- AssignWindowTranslatorBatch<T> - Class in org.apache.beam.runners.twister2.translators.batch
- 
Assign Window translator. 
- AssignWindowTranslatorBatch() - Constructor for class org.apache.beam.runners.twister2.translators.batch.AssignWindowTranslatorBatch
-  
- asSingleton() - Static method in class org.apache.beam.sdk.transforms.View
- 
- asSingletonView() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
- 
Returns a  PTransform that produces a  PCollectionView whose elements are the
 result of combining elements per-window in the input  PCollection. 
- assumeSingleMessageSchema() - Method in class org.apache.beam.sdk.extensions.sbe.SbeSchema.IrOptions
-  
- ASTERISK - Static variable in class org.apache.beam.sdk.extensions.sql.impl.cep.Quantifier
-  
- ASTERISK_RELUCTANT - Static variable in class org.apache.beam.sdk.extensions.sql.impl.cep.Quantifier
-  
- asV1Beta1Path() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription
- 
- asV1Beta1Path() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic
- 
- asV1Beta2Path() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription
- 
- asV1Beta2Path() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic
- 
- AsyncBatchWriteHandler<RecT,ResT> - Class in org.apache.beam.sdk.io.aws2.common
- 
Async handler that automatically retries unprocessed records in case of a partial success. 
- AsyncBatchWriteHandler(int, FluentBackoff, AsyncBatchWriteHandler.Stats, Function<ResT, String>, BiFunction<String, List<RecT>, CompletableFuture<List<ResT>>>) - Constructor for class org.apache.beam.sdk.io.aws2.common.AsyncBatchWriteHandler
-  
- AsyncBatchWriteHandler.Stats - Interface in org.apache.beam.sdk.io.aws2.common
- 
Statistics on the batch request. 
- AsyncWatermarkCache - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.cache
- 
Asynchronously compute the earliest partition watermark and stores it in memory. 
- AsyncWatermarkCache(PartitionMetadataDao, Duration) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.cache.AsyncWatermarkCache
-  
- atMinimumTimestamp(V) - Static method in class org.apache.beam.sdk.values.TimestampedValue
- 
- AtomicCoder<T> - Class in org.apache.beam.sdk.coders
- 
A  Coder that has no component  Coders or other configuration. 
- AtomicCoder() - Constructor for class org.apache.beam.sdk.coders.AtomicCoder
-  
- AtomicLongFactory() - Constructor for class org.apache.beam.sdk.options.PipelineOptions.AtomicLongFactory
-  
- attached() - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
-  
- attachValues(List<Object>) - Method in class org.apache.beam.sdk.values.Row.Builder
-  
- attachValues(Object...) - Method in class org.apache.beam.sdk.values.Row.Builder
-  
- attempted(MetricKey, T) - Static method in class org.apache.beam.sdk.metrics.MetricResult
-  
- ATTRIBUTE_ARRAY_ENTRY_SCHEMA - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.PubsubSchemaIOProvider
-  
- ATTRIBUTE_ARRAY_FIELD_TYPE - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.PubsubSchemaIOProvider
-  
- ATTRIBUTE_MAP_FIELD_TYPE - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.PubsubSchemaIOProvider
-  
- AttributeValueCoder - Class in org.apache.beam.sdk.io.aws2.dynamodb
- 
- AUTH_VALIDATION_GROUP - Static variable in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- AuthenticatedRetryInitializer(GoogleCredentials) - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient.AuthenticatedRetryInitializer
-  
- AUTO - Static variable in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- autoCastField(Schema.Field, Object) - Static method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamTableUtils
- 
Attempt to cast an object to a specified Schema.Field.Type. 
- autoLoadUserDefinedFunctions() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
- 
- AutoScaler - Interface in org.apache.beam.sdk.io.jms
- 
Enables users to specify their own `JMS` backlog reporters enabling  JmsIO to report
  UnboundedSource.UnboundedReader#getTotalBacklogBytes(). 
- AUTOVALUE_CLASS - Static variable in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- AUTOVALUE_CLASS - Static variable in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
-  
- AutoValueSchema - Class in org.apache.beam.sdk.schemas
- 
- AutoValueSchema() - Constructor for class org.apache.beam.sdk.schemas.AutoValueSchema
-  
- AutoValueSchema.AbstractGetterTypeSupplier - Class in org.apache.beam.sdk.schemas
- 
- AutoValueUtils - Class in org.apache.beam.sdk.schemas.utils
- 
Utilities for managing AutoValue schemas. 
- AutoValueUtils() - Constructor for class org.apache.beam.sdk.schemas.utils.AutoValueUtils
-  
- AvailableParallelismFactory() - Constructor for class org.apache.beam.runners.direct.DirectOptions.AvailableParallelismFactory
-  
- AVRO_CODER_URN - Static variable in class org.apache.beam.sdk.extensions.avro.AvroGenericCoderRegistrar
-  
- AvroCoder<T> - Class in org.apache.beam.sdk.extensions.avro.coders
- 
A  Coder using Avro binary format. 
- AvroCoder(Class<T>, Schema) - Constructor for class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
-  
- AvroCoder(Class<T>, Schema, boolean) - Constructor for class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
-  
- AvroCoder(AvroDatumFactory<T>, Schema) - Constructor for class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
-  
- AvroConvertType(boolean) - Constructor for class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.AvroConvertType
-  
- AvroDatumFactory<T> - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroDatumFactory(Class<T>) - Constructor for class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory
-  
- AvroDatumFactory.GenericDatumFactory - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroDatumFactory.ReflectDatumFactory<T> - Class in org.apache.beam.sdk.extensions.avro.io
- 
Specialized  AvroDatumFactory for java classes transforming to avro through reflection. 
- AvroDatumFactory.SpecificDatumFactory<T> - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroGenericCoder - Class in org.apache.beam.sdk.extensions.avro.coders
- 
AvroCoder specialisation for GenericRecord, needed for cross-language transforms. 
- AvroGenericCoderRegistrar - Class in org.apache.beam.sdk.extensions.avro
- 
Coder registrar for AvroGenericCoder. 
- AvroGenericCoderRegistrar() - Constructor for class org.apache.beam.sdk.extensions.avro.AvroGenericCoderRegistrar
-  
- AvroGenericCoderTranslator - Class in org.apache.beam.sdk.extensions.avro
- 
Coder translator for AvroGenericCoder. 
- AvroGenericCoderTranslator() - Constructor for class org.apache.beam.sdk.extensions.avro.AvroGenericCoderTranslator
-  
- AvroGenericRecordToStorageApiProto - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
Utility methods for converting Avro  GenericRecord objects to dynamic protocol message,
 for use with the Storage write API. 
- AvroGenericRecordToStorageApiProto() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.AvroGenericRecordToStorageApiProto
-  
- AvroIO - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroIO.Parse<T> - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroIO.ParseAll<T> - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroIO.ParseFiles<T> - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroIO.Read<T> - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroIO.ReadAll<T> - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroIO.ReadFiles<T> - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroIO.RecordFormatter<ElementT> - Interface in org.apache.beam.sdk.extensions.avro.io
- 
- AvroIO.Sink<ElementT> - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroIO.TypedWrite<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroIO.Write<T> - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroJavaTimeConversions - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
- 
Avro 1.8 ships with joda time conversions only. 
- AvroJavaTimeConversions() - Constructor for class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions
-  
- AvroJavaTimeConversions.DateConversion - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJavaTimeConversions.LocalTimestampMicros - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJavaTimeConversions.LocalTimestampMicrosConversion - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJavaTimeConversions.LocalTimestampMillis - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJavaTimeConversions.LocalTimestampMillisConversion - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJavaTimeConversions.TimeMicrosConversion - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJavaTimeConversions.TimeMillisConversion - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJavaTimeConversions.TimestampMicrosConversion - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJavaTimeConversions.TimestampMillisConversion - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJodaTimeConversions - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
- 
Avro 1.8 & 1.9 ship joda time conversions. 
- AvroJodaTimeConversions() - Constructor for class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions
-  
- AvroJodaTimeConversions.DateConversion - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJodaTimeConversions.LossyTimeMicrosConversion - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJodaTimeConversions.LossyTimestampMicrosConversion - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJodaTimeConversions.TimeConversion - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJodaTimeConversions.TimeMicrosConversion - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJodaTimeConversions.TimestampConversion - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroJodaTimeConversions.TimestampMicrosConversion - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroPayloadSerializerProvider - Class in org.apache.beam.sdk.extensions.avro.schemas.io.payloads
-  
- AvroPayloadSerializerProvider() - Constructor for class org.apache.beam.sdk.extensions.avro.schemas.io.payloads.AvroPayloadSerializerProvider
-  
- AvroReader(AvroSource<T>) - Constructor for class org.apache.beam.sdk.extensions.avro.io.AvroSource.AvroReader
- 
Reads Avro records of type Tfrom the specified source.
 
- AvroReadSchemaTransformFormatProvider - Class in org.apache.beam.sdk.io.fileschematransform
-  
- AvroReadSchemaTransformFormatProvider() - Constructor for class org.apache.beam.sdk.io.fileschematransform.AvroReadSchemaTransformFormatProvider
-  
- AvroRecordSchema - Class in org.apache.beam.sdk.extensions.avro.schemas
- 
- AvroRecordSchema() - Constructor for class org.apache.beam.sdk.extensions.avro.schemas.AvroRecordSchema
-  
- AvroSchemaInformationProvider - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroSchemaInformationProvider() - Constructor for class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroSchemaInformationProvider
-  
- AvroSchemaIOProvider - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroSchemaIOProvider() - Constructor for class org.apache.beam.sdk.extensions.avro.io.AvroSchemaIOProvider
-  
- AvroSink<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.extensions.avro.io
- 
- AvroSink.DatumWriterFactory<T> - Interface in org.apache.beam.sdk.extensions.avro.io
-  
- AvroSource<T> - Class in org.apache.beam.sdk.extensions.avro.io
- 
Do not use in pipelines directly: most users should use  AvroIO.Read. 
- AvroSource.AvroReader<T> - Class in org.apache.beam.sdk.extensions.avro.io
- 
A BlockBasedReaderfor reading blocks from Avro files.
 
- AvroSource.DatumReaderFactory<T> - Interface in org.apache.beam.sdk.extensions.avro.io
-  
- AvroTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.avro
- 
- AvroTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.avro.AvroTableProvider
-  
- AvroUtils - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
- 
Utils to convert AVRO records to Beam rows. 
- AvroUtils.AvroConvertType - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroUtils.AvroConvertValueForGetter - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroUtils.AvroConvertValueForSetter - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroUtils.FixedBytesField - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
- 
Wrapper for fixed byte fields. 
- AvroUtils.TypeWithNullability - Class in org.apache.beam.sdk.extensions.avro.schemas.utils
-  
- AvroWriteRequest<T> - Class in org.apache.beam.sdk.io.gcp.bigquery
-  
- AvroWriteRequest(T, Schema) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.AvroWriteRequest
-  
- AvroWriteSchemaTransformFormatProvider - Class in org.apache.beam.sdk.io.fileschematransform
- 
- AvroWriteSchemaTransformFormatProvider() - Constructor for class org.apache.beam.sdk.io.fileschematransform.AvroWriteSchemaTransformFormatProvider
-  
- awaitCompletion() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
- 
Uses the callers thread to process all elements received until we receive the end of the stream
 from the upstream producer for all endpoints specified. 
- awaitTermination(Duration) - Method in class org.apache.beam.runners.spark.SparkPipelineResult
-  
- awaitTermination(Duration) - Method in class org.apache.beam.runners.spark.SparkRunnerDebugger.DebugSparkPipelineResult
-  
- AwsBuilderFactory<PojoT extends SdkPojo,BuilderT extends SdkBuilder<BuilderT,PojoT> & SdkPojo> - Class in org.apache.beam.sdk.io.aws2.schemas
- 
Builder factory for AWS  SdkPojo to avoid using reflection to instantiate a builder. 
- AwsBuilderFactory() - Constructor for class org.apache.beam.sdk.io.aws2.schemas.AwsBuilderFactory
-  
- AwsModule - Class in org.apache.beam.sdk.io.aws2.options
- 
- AwsModule() - Constructor for class org.apache.beam.sdk.io.aws2.options.AwsModule
-  
- AwsOptions - Interface in org.apache.beam.sdk.io.aws2.options
- 
Options used to configure Amazon Web Services specific options such as credentials and region. 
- AwsOptions.AwsRegionFactory - Class in org.apache.beam.sdk.io.aws2.options
- 
Attempt to load default region. 
- AwsOptions.AwsUserCredentialsFactory - Class in org.apache.beam.sdk.io.aws2.options
- 
- AwsPipelineOptionsRegistrar - Class in org.apache.beam.sdk.io.aws2.options
- 
A registrar containing the default AWS options. 
- AwsPipelineOptionsRegistrar() - Constructor for class org.apache.beam.sdk.io.aws2.options.AwsPipelineOptionsRegistrar
-  
- AwsRegionFactory() - Constructor for class org.apache.beam.sdk.io.aws2.options.AwsOptions.AwsRegionFactory
-  
- AwsSchemaProvider - Class in org.apache.beam.sdk.io.aws2.schemas
- 
- AwsSchemaProvider() - Constructor for class org.apache.beam.sdk.io.aws2.schemas.AwsSchemaProvider
-  
- AwsSchemaRegistrar - Class in org.apache.beam.sdk.io.aws2.schemas
-  
- AwsSchemaRegistrar() - Constructor for class org.apache.beam.sdk.io.aws2.schemas.AwsSchemaRegistrar
-  
- AwsSerializableUtils - Class in org.apache.beam.sdk.io.aws2.options
- 
Utilities for working with AWS Serializables. 
- AwsSerializableUtils() - Constructor for class org.apache.beam.sdk.io.aws2.options.AwsSerializableUtils
-  
- AwsTypes - Class in org.apache.beam.sdk.io.aws2.schemas
-  
- AwsTypes() - Constructor for class org.apache.beam.sdk.io.aws2.schemas.AwsTypes
-  
- AwsUserCredentialsFactory() - Constructor for class org.apache.beam.sdk.io.aws2.options.AwsOptions.AwsUserCredentialsFactory
-  
- AzureBlobStoreFileSystemRegistrar - Class in org.apache.beam.sdk.io.azure.blobstore
- 
AutoServiceregistrar for theAzureBlobStoreFileSystem.
 
- AzureBlobStoreFileSystemRegistrar() - Constructor for class org.apache.beam.sdk.io.azure.blobstore.AzureBlobStoreFileSystemRegistrar
-  
- AzureModule - Class in org.apache.beam.sdk.io.azure.options
- 
- AzureModule() - Constructor for class org.apache.beam.sdk.io.azure.options.AzureModule
-  
- AzureOptions - Interface in org.apache.beam.sdk.io.azure.options
-  
- AzureOptions.AzureUserCredentialsFactory - Class in org.apache.beam.sdk.io.azure.options
- 
Attempts to load Azure credentials. 
- AzurePipelineOptionsRegistrar - Class in org.apache.beam.sdk.io.azure.options
- 
A registrar containing the default Azure options. 
- AzurePipelineOptionsRegistrar() - Constructor for class org.apache.beam.sdk.io.azure.options.AzurePipelineOptionsRegistrar
-  
- AzureUserCredentialsFactory() - Constructor for class org.apache.beam.sdk.io.azure.options.AzureOptions.AzureUserCredentialsFactory
-  
- Cache - Class in org.apache.beam.io.requestresponse
- 
Transforms for reading and writing request/response associations to a cache. 
- Cache() - Constructor for class org.apache.beam.io.requestresponse.Cache
-  
- Cache.Pair<RequestT,ResponseT> - Class in org.apache.beam.io.requestresponse
- 
A simple POJO that holds both cache read and write  PTransforms. 
- CACHED_CREATORS - Static variable in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
-  
- CACHED_CREATORS - Static variable in class org.apache.beam.sdk.schemas.utils.POJOUtils
-  
- CachedSideInputReader - Class in org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions
- 
- CachedSideInputReader - Class in org.apache.beam.runners.spark.util
- 
SideInputReaderthat caches materialized views.
 
- CacheFactory - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.cache
-  
- CacheFactory(DaoFactory, Duration) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.cache.CacheFactory
-  
- CachingFactory<CreatedT> - Class in org.apache.beam.sdk.schemas
- 
A wrapper around a  Factory that assumes the schema parameter never changes. 
- CachingFactory(Factory<CreatedT>) - Constructor for class org.apache.beam.sdk.schemas.CachingFactory
-  
- CalciteConnectionWrapper - Class in org.apache.beam.sdk.extensions.sql.impl
- 
Abstract wrapper for CalciteConnectionto simplify extension.
 
- CalciteConnectionWrapper(CalciteConnection) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- CalciteFactoryWrapper - Class in org.apache.beam.sdk.extensions.sql.impl
- 
Wrapper for CalciteFactory.
 
- CalciteQueryPlanner - Class in org.apache.beam.sdk.extensions.sql.impl
- 
The core component to handle through a SQL statement, from explain execution plan, to generate a
 Beam pipeline. 
- CalciteQueryPlanner(JdbcConnection, Collection<RuleSet>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner
- 
Called by  BeamSqlEnv.instantiatePlanner() reflectively. 
- CalciteQueryPlanner.NonCumulativeCostImpl - Class in org.apache.beam.sdk.extensions.sql.impl
-  
- CalciteUtils - Class in org.apache.beam.sdk.extensions.sql.impl.utils
- 
Utility methods for Calcite related operations. 
- CalciteUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- CalciteUtils.TimeWithLocalTzType - Class in org.apache.beam.sdk.extensions.sql.impl.utils
- 
A LogicalType corresponding to TIME_WITH_LOCAL_TIME_ZONE. 
- CalcRelSplitter - Class in org.apache.beam.sdk.extensions.sql.impl.rel
- 
CalcRelSplitter operates on a Calcwith multipleRexCallsub-expressions that
 cannot all be implemented by a single concreteRelNode.
 
- CalcRelSplitter(Calc, RelBuilder, CalcRelSplitter.RelType[]) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.CalcRelSplitter
- 
Constructs a CalcRelSplitter. 
- CalcRelSplitter.RelType - Class in org.apache.beam.sdk.extensions.sql.impl.rel
- 
Type of relational expression. 
- calculateRanges(PartitionT, PartitionT, Long) - Method in interface org.apache.beam.sdk.io.jdbc.JdbcReadWithPartitionsHelper
- 
Calculate the range of each partition from the lower and upper bound, and number of partitions. 
- CalendarWindows - Class in org.apache.beam.sdk.transforms.windowing
- 
A collection of  WindowFns that windows values into calendar-based windows such as spans
 of days, months, or years. 
- CalendarWindows() - Constructor for class org.apache.beam.sdk.transforms.windowing.CalendarWindows
-  
- CalendarWindows.DaysWindows - Class in org.apache.beam.sdk.transforms.windowing
- 
A  WindowFn that windows elements into periods measured by days. 
- CalendarWindows.MonthsWindows - Class in org.apache.beam.sdk.transforms.windowing
- 
A  WindowFn that windows elements into periods measured by months. 
- CalendarWindows.YearsWindows - Class in org.apache.beam.sdk.transforms.windowing
- 
A  WindowFn that windows elements into periods measured by years. 
- call(RequestT) - Method in interface org.apache.beam.io.requestresponse.Caller
- 
Calls a Web API with the RequestTand returns aResponseT.
 
- call(Tuple2<ByteArray, byte[]>) - Method in class org.apache.beam.runners.spark.coders.CoderHelpers.FromByteFunction
-  
- call(K, Iterator<WindowedValue<KV<K, InputT>>>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.GroupAlsoByWindowViaOutputBufferFn
-  
- Caller<RequestT,ResponseT> - Interface in org.apache.beam.io.requestresponse
- 
Caller interfaces user custom code intended for API calls. 
- CallShouldBackoff<ResponseT> - Interface in org.apache.beam.io.requestresponse
- 
Informs whether a call to an API should backoff. 
- cancel() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-  
- cancel() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
-  
- cancel() - Method in class org.apache.beam.runners.direct.DirectRunner.DirectPipelineResult
-  
- cancel() - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
-  
- cancel() - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
-  
- cancel() - Method in class org.apache.beam.runners.jet.FailedRunningPipelineResults
-  
- cancel() - Method in class org.apache.beam.runners.jet.JetPipelineResult
-  
- cancel(JobApi.CancelJobRequest, StreamObserver<JobApi.CancelJobResponse>) - Method in class org.apache.beam.runners.jobsubmission.InMemoryJobService
-  
- cancel() - Method in class org.apache.beam.runners.jobsubmission.JobInvocation
- 
Cancel the job. 
- cancel() - Method in class org.apache.beam.runners.spark.SparkPipelineResult
-  
- cancel() - Method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineResult
-  
- cancel() - Method in class org.apache.beam.runners.twister2.Twister2PipelineResult
-  
- cancel(Exception) - Method in class org.apache.beam.sdk.fn.CancellableQueue
- 
- cancel() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.BigQueryServerStream
- 
Cancels the stream, releasing any client- and server-side resources. 
- cancel() - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices.FakeBigQueryServerStream
-  
- cancel() - Method in interface org.apache.beam.sdk.PipelineResult
- 
Cancels the pipeline execution. 
- CancellableQueue<T> - Class in org.apache.beam.sdk.fn
- 
A simplified  ThreadSafe blocking queue that can be cancelled freeing any blocked  Threads and preventing future  Threads from blocking. 
- CancellableQueue(int) - Constructor for class org.apache.beam.sdk.fn.CancellableQueue
- 
Creates a  ThreadSafe blocking queue with a maximum capacity. 
- cancelled() - Method in interface org.apache.beam.runners.local.PipelineMessageReceiver
- 
Report that the pipeline has been cancelled. 
- canConvertConvention(Convention) - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-  
- canImplement(LogicalCalc, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.CalcRelSplitter
- 
- canImplement(RexFieldAccess) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.CalcRelSplitter.RelType
-  
- canImplement(RexDynamicParam) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.CalcRelSplitter.RelType
-  
- canImplement(RexLiteral) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.CalcRelSplitter.RelType
-  
- canImplement(RexCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.CalcRelSplitter.RelType
-  
- canImplement(RexNode, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.CalcRelSplitter.RelType
- 
Returns whether this RelTypecan implement a given expression.
 
- canImplement(RexProgram) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.CalcRelSplitter.RelType
- 
Returns whether this tester's RelTypecan implement a given program.
 
- CannotProvideCoderException - Exception in org.apache.beam.sdk.coders
- 
- CannotProvideCoderException(String) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
-  
- CannotProvideCoderException(String, CannotProvideCoderException.ReasonCode) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
-  
- CannotProvideCoderException(String, Throwable) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
-  
- CannotProvideCoderException(String, Throwable, CannotProvideCoderException.ReasonCode) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
-  
- CannotProvideCoderException(Throwable) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
-  
- CannotProvideCoderException(Throwable, CannotProvideCoderException.ReasonCode) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
-  
- CannotProvideCoderException.ReasonCode - Enum in org.apache.beam.sdk.coders
- 
Indicates the reason that  Coder inference failed. 
- canStopPolling(Instant, StateT) - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
- 
- canTranslate(TransformT) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator
- 
Checks if a composite / primitive transform can be translated. 
- CassandraIO - Class in org.apache.beam.sdk.io.cassandra
- 
An IO to read and write from/to Apache Cassandra 
- CassandraIO.MutationType - Enum in org.apache.beam.sdk.io.cassandra
- 
Specify the mutation type: either write or delete. 
- CassandraIO.Read<T> - Class in org.apache.beam.sdk.io.cassandra
- 
- CassandraIO.ReadAll<T> - Class in org.apache.beam.sdk.io.cassandra
- 
- CassandraIO.Write<T> - Class in org.apache.beam.sdk.io.cassandra
- 
- Cast<T> - Class in org.apache.beam.sdk.schemas.transforms
- 
Set of utilities for casting rows between schemas. 
- Cast() - Constructor for class org.apache.beam.sdk.schemas.transforms.Cast
-  
- Cast.CompatibilityError - Class in org.apache.beam.sdk.schemas.transforms
- 
Describes compatibility errors during casting. 
- Cast.Narrowing - Class in org.apache.beam.sdk.schemas.transforms
- 
Narrowing changes type without guarantee to preserve data. 
- Cast.Validator - Interface in org.apache.beam.sdk.schemas.transforms
- 
Interface for statically validating casts. 
- Cast.Widening - Class in org.apache.beam.sdk.schemas.transforms
- 
Widening changes to type that can represent any possible value of the original type. 
- CAST_OP - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- CastFunctionImpl - Class in org.apache.beam.sdk.extensions.sql.zetasql.translation.impl
- 
ZetaSQLCastFunctionImpl. 
- CastFunctionImpl() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.CastFunctionImpl
-  
- castNumber(Number, Schema.TypeName, Schema.TypeName) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
-  
- castRow(Row, Schema, Schema) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
-  
- castValue(Object, Schema.FieldType, Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
-  
- catalog() - Method in class org.apache.beam.sdk.io.iceberg.IcebergCatalogConfig
-  
- catalogName() - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAICreateCatalogItem
-  
- catalogName() - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportCatalogItems
-  
- catalogName() - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportUserEvents
-  
- catalogName() - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIPredict
-  
- catalogName() - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIWriteUserEvent
-  
- catchUpToNow(boolean) - Method in class org.apache.beam.sdk.transforms.PeriodicImpulse
- 
For internal use only; no backwards-compatibility guarantees. 
- catchUpToNow - Variable in class org.apache.beam.sdk.transforms.PeriodicSequence.SequenceDefinition
-  
- CdapIO - Class in org.apache.beam.sdk.io.cdap
- 
A  CdapIO is a Transform for reading data from source or writing data to sink of a Cdap
 Plugin. 
- CdapIO() - Constructor for class org.apache.beam.sdk.io.cdap.CdapIO
-  
- CdapIO.Read<K,V> - Class in org.apache.beam.sdk.io.cdap
- 
- CdapIO.Write<K,V> - Class in org.apache.beam.sdk.io.cdap
- 
- cdapPluginObj - Variable in class org.apache.beam.sdk.io.cdap.Plugin
-  
- CELL_SCHEMA - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProvider
-  
- CEPCall - Class in org.apache.beam.sdk.extensions.sql.impl.cep
- 
A CEPCallinstance represents an operation (node) that contains an operator and a list of
 operands.
 
- CEPFieldRef - Class in org.apache.beam.sdk.extensions.sql.impl.cep
- 
A CEPFieldRefinstance represents a node that points to a specified field in aRow.
 
- CEPKind - Enum in org.apache.beam.sdk.extensions.sql.impl.cep
- 
CEPKindcorresponds to Calcite'sSqlKind.
 
- CEPLiteral - Class in org.apache.beam.sdk.extensions.sql.impl.cep
- 
CEPLiteralrepresents a literal node.
 
- CEPMeasure - Class in org.apache.beam.sdk.extensions.sql.impl.cep
- 
The CEPMeasureclass represents the Measures clause and contains information about output
 columns.
 
- CEPMeasure(Schema, String, CEPOperation) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.cep.CEPMeasure
-  
- CEPOperation - Class in org.apache.beam.sdk.extensions.sql.impl.cep
- 
CEPOperationis the base class for the evaluation operations defined in theDEFINEsyntax ofMATCH_RECOGNIZE.
 
- CEPOperation() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.cep.CEPOperation
-  
- CEPOperator - Class in org.apache.beam.sdk.extensions.sql.impl.cep
- 
The CEPOperatorrecords the operators (i.e.
 
- CEPPattern - Class in org.apache.beam.sdk.extensions.sql.impl.cep
- 
Core pattern class that stores the definition of a single pattern. 
- CEPUtils - Class in org.apache.beam.sdk.extensions.sql.impl.cep
- 
Some utility methods for transforming Calcite's constructs into our own Beam constructs (for
 serialization purpose). 
- CEPUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.cep.CEPUtils
-  
- CF_CONTINUATION_TOKEN - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
-  
- CF_INITIAL_TOKEN - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
-  
- CF_LOCK - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
-  
- CF_MISSING_PARTITIONS - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
-  
- CF_PARENT_LOW_WATERMARKS - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
-  
- CF_PARENT_PARTITIONS - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
-  
- CF_SHOULD_DELETE - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
-  
- CF_VERSION - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
-  
- CF_WATERMARK - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
-  
- CHANGE_SQN_COLUMN - Static variable in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiCDC
-  
- CHANGE_STREAM_MUTATION_GC_COUNT - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
Counter for the total number of ChangeStreamMutations that are initiated by garbage collection
 (not user initiated) identified during the execution of the Connector. 
- CHANGE_STREAM_MUTATION_USER_COUNT - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
Counter for the total number of ChangeStreamMutations that are initiated by users (not garbage
 collection) identified during the execution of the Connector. 
- CHANGE_TYPE_COLUMN - Static variable in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiCDC
-  
- changeStreamAction(ChangeStreamMetrics) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.ActionFactory
- 
- ChangeStreamAction - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.action
- 
This class is responsible for processing individual ChangeStreamRecord. 
- ChangeStreamAction(ChangeStreamMetrics) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.ChangeStreamAction
- 
Constructs ChangeStreamAction to process individual ChangeStreamRecord. 
- ChangeStreamContinuationTokenHelper - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams
-  
- ChangeStreamContinuationTokenHelper() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamContinuationTokenHelper
-  
- ChangeStreamDao - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao
- 
Data access object to list and read stream partitions of a table. 
- ChangeStreamDao(BigtableDataClient, String) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.ChangeStreamDao
-  
- ChangeStreamDao - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.dao
- 
Responsible for making change stream queries for a given partition. 
- ChangeStreamMetrics - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams
- 
Class to aggregate metrics related functionality. 
- ChangeStreamMetrics() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
-  
- ChangeStreamMetrics - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams
- 
Class to aggregate metrics related functionality. 
- ChangeStreamMetrics() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
Constructs a ChangeStreamMetrics instance with the following metrics enabled by default. 
- ChangeStreamMetrics(Set<MetricName>) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
Constructs a ChangeStreamMetrics instance with the given metrics enabled. 
- changeStreamQuery(String, Timestamp, Timestamp, long) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.ChangeStreamDao
- 
Performs a change stream query. 
- ChangeStreamRecord - Interface in org.apache.beam.sdk.io.gcp.spanner.changestreams.model
- 
Represents a Spanner Change Stream Record. 
- ChangeStreamRecordMapper - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.mapper
- 
This class is responsible for transforming a  Struct to a  List of  ChangeStreamRecord models. 
- changeStreamRecordMapper() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.mapper.MapperFactory
- 
Creates and returns a singleton instance of a mapper class capable of transforming a  Struct into a  List of  ChangeStreamRecord subclasses. 
- ChangeStreamRecordMetadata - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.model
- 
- ChangeStreamRecordMetadata.Builder - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.model
-  
- ChangeStreamResultSet - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.dao
- 
Decorator class over a ResultSetthat provides telemetry for the streamed records.
 
- ChangeStreamResultSetMetadata - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.dao
- 
Represents telemetry metadata gathered during the consumption of a change stream query. 
- ChangeStreamsConstants - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams
- 
Single place for defining the constants used in the Spanner.readChangeStreams()connector.
 
- ChangeStreamsConstants() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamsConstants
-  
- channelNames - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- CHAR - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- CHAR_LENGTH - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- CHAR_LENGTH_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.BeamBuiltinMethods
-  
- characters() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
- 
- charLength(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- check(RelNode) - Method in interface org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall.JoinChecker
-  
- checkClientTrusted(X509Certificate[], String) - Method in class org.apache.beam.sdk.io.splunk.CustomX509TrustManager
-  
- checkConfiguration(ClientConfiguration, AwsOptions) - Method in interface org.apache.beam.sdk.io.aws2.common.ClientBuilderFactory
- 
Check if all necessary configuration is available to create clients. 
- checkConfiguration(ClientConfiguration, AwsOptions) - Method in class org.apache.beam.sdk.io.aws2.common.ClientBuilderFactory.DefaultClientBuilder
-  
- checkDone() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.ReadChangeStreamPartitionProgressTracker
- 
This is to signal to the runner that this restriction has completed. 
- checkDone() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRangeTracker
- 
Checks if the restriction has been processed successfully. 
- checkDone() - Method in class org.apache.beam.sdk.transforms.PeriodicSequence.OutputRangeTracker
-  
- checkDone() - Method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
-  
- checkDone() - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
-  
- checkDone() - Method in class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker
- 
Checks whether the restriction has been fully processed. 
- checkForAsyncFailure() - Method in class org.apache.beam.sdk.io.aws2.common.AsyncBatchWriteHandler
- 
Check if any failure happened async. 
- checkIfAnySubscriptionExists(String, Duration) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
- 
- CheckpointMarkImpl - Class in org.apache.beam.sdk.io.gcp.pubsublite.internal
-  
- checkServerTrusted(X509Certificate[], String) - Method in class org.apache.beam.sdk.io.splunk.CustomX509TrustManager
-  
- CheckStopReadingFn - Interface in org.apache.beam.sdk.io.kafka
-  
- CheckStopReadingFnWrapper - Class in org.apache.beam.sdk.io.kafka
-  
- checksum() - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata
- 
An optional checksum to identify the contents of a file. 
- ChildPartition - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.model
- 
A child partition represents a new partition that should be queried. 
- ChildPartition(String, HashSet<String>) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChildPartition
- 
Constructs a child partition, which will have its own token and the parents that it originated
 from. 
- ChildPartition(String, String) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChildPartition
- 
Constructs a child partition, which will have its own token and the parent that it originated
 from. 
- ChildPartitionsRecord - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.model
- 
Represents a ChildPartitionsRecord. 
- ChildPartitionsRecord(Timestamp, String, List<ChildPartition>, ChangeStreamRecordMetadata) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChildPartitionsRecord
- 
Constructs a child partitions record containing one or more child partitions. 
- childPartitionsRecordAction(PartitionMetadataDao, ChangeStreamMetrics) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.action.ActionFactory
- 
- ChildPartitionsRecordAction - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.action
- 
- CivilTimeEncoder - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
Encoder for TIME and DATETIME values, according to civil_time encoding. 
- classesToTranslators() - Method in interface org.apache.beam.runners.dataflow.util.CoderCloudObjectTranslatorRegistrar
- 
- classesToTranslators() - Method in class org.apache.beam.runners.dataflow.util.DefaultCoderCloudObjectTranslatorRegistrar
-  
- ClassLoaderFileSystem - Class in org.apache.beam.sdk.io
- 
A read-only  FileSystem implementation looking up resources using a ClassLoader. 
- ClassLoaderFileSystem.ClassLoaderFileSystemRegistrar - Class in org.apache.beam.sdk.io
- 
- ClassLoaderFileSystem.ClassLoaderResourceId - Class in org.apache.beam.sdk.io
-  
- ClassLoaderFileSystemRegistrar() - Constructor for class org.apache.beam.sdk.io.ClassLoaderFileSystem.ClassLoaderFileSystemRegistrar
-  
- classNamesToTranslators() - Method in interface org.apache.beam.runners.dataflow.util.CoderCloudObjectTranslatorRegistrar
- 
- classNamesToTranslators() - Method in class org.apache.beam.runners.dataflow.util.DefaultCoderCloudObjectTranslatorRegistrar
-  
- ClassWithSchema() - Constructor for class org.apache.beam.sdk.schemas.utils.ReflectUtils.ClassWithSchema
-  
- CleanTmpFilesFromGcsFn(ValueProvider<String>, String) - Constructor for class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read.CleanTmpFilesFromGcsFn
- 
Created object that will remove temp files from stage. 
- cleanup() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
-  
- cleanUpPrefix() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
- 
Delete all the metadata rows starting with the change stream name prefix, except for detect new
 partition row because it signals the existence of a pipeline with the change stream name. 
- CleanUpReadChangeStreamDoFn - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn
-  
- CleanUpReadChangeStreamDoFn(DaoFactory) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.CleanUpReadChangeStreamDoFn
-  
- clear(K, W) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.BagUserStateHandler
- 
Clears the bag user state for the given key and window. 
- clear() - Static method in class org.apache.beam.runners.spark.metrics.MetricsAccumulator
-  
- clear() - Static method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsAccumulator
-  
- clear() - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
-  
- clear() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
-  
- clear() - Method in interface org.apache.beam.sdk.state.State
- 
Clear out the state location. 
- clear() - Method in interface org.apache.beam.sdk.state.Timer
- 
Clears a timer. 
- clearCache() - Static method in class org.apache.beam.runners.spark.io.MicrobatchSource
-  
- clearOutputElements() - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- clearOutputElements(TupleTag<T>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- clearRange(Instant, Instant) - Method in interface org.apache.beam.sdk.state.OrderedListState
- 
Clear a timestamp-limited subrange of the list. 
- clearState(ReduceFn<K, T, Iterable<T>, W>.Context) - Method in class org.apache.beam.runners.twister2.translators.functions.internal.SystemReduceFnBuffering
-  
- clearWarnings() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- ClickHouseIO - Class in org.apache.beam.sdk.io.clickhouse
- 
An IO to write to ClickHouse. 
- ClickHouseIO() - Constructor for class org.apache.beam.sdk.io.clickhouse.ClickHouseIO
-  
- ClickHouseIO.Write<T> - Class in org.apache.beam.sdk.io.clickhouse
- 
- ClickHouseWriter - Class in org.apache.beam.sdk.io.clickhouse
- 
Writes Rows and field values using ClickHousePipedOutputStream.
 
- ClickHouseWriter() - Constructor for class org.apache.beam.sdk.io.clickhouse.ClickHouseWriter
-  
- clientBuffered(ExecutorService) - Static method in class org.apache.beam.sdk.fn.stream.OutboundObserverFactory
- 
Create a buffering  OutboundObserverFactory for client-side RPCs with the specified
  ExecutorService and the default buffer size. 
- clientBuffered(ExecutorService, int) - Static method in class org.apache.beam.sdk.fn.stream.OutboundObserverFactory
- 
Create a buffering  OutboundObserverFactory for client-side RPCs with the specified
  ExecutorService and buffer size. 
- ClientBuilderFactory - Interface in org.apache.beam.sdk.io.aws2.common
- 
- ClientBuilderFactory.DefaultClientBuilder - Class in org.apache.beam.sdk.io.aws2.common
- 
- ClientBuilderFactory.SkipCertificateVerificationTrustManagerProvider - Class in org.apache.beam.sdk.io.aws2.common
- 
Trust provider to skip certificate verification. 
- ClientConfiguration - Class in org.apache.beam.sdk.io.aws2.common
- 
AWS client configuration. 
- ClientConfiguration() - Constructor for class org.apache.beam.sdk.io.aws2.common.ClientConfiguration
-  
- ClientConfiguration.Builder - Class in org.apache.beam.sdk.io.aws2.common
-  
- clientDirect() - Static method in class org.apache.beam.sdk.fn.stream.OutboundObserverFactory
- 
- Clock - Interface in org.apache.beam.runners.direct
- 
Access to the current time. 
- clone() - Method in class org.apache.beam.runners.dataflow.util.CloudObject
-  
- clone() - Method in class org.apache.beam.runners.flink.metrics.MetricsAccumulator
-  
- clonesOf(OutT) - Static method in class org.apache.beam.sdk.transforms.SerializableFunctions
-  
- close() - Method in class org.apache.beam.runners.dataflow.util.PackageUtil
-  
- close() - Method in class org.apache.beam.runners.flink.metrics.FileReporter
-  
- close() - Method in class org.apache.beam.runners.fnexecution.artifact.ArtifactRetrievalService
-  
- close() - Method in class org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService
-  
- close() - Method in class org.apache.beam.runners.fnexecution.control.DefaultExecutableStageContext
-  
- close() - Method in class org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory
-  
- close() - Method in class org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory.WrappedSdkHarnessClient
-  
- close() - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClient
-  
- close() - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService
-  
- close() - Method in interface org.apache.beam.runners.fnexecution.control.RemoteBundle
- 
Closes this bundle. 
- close() - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor.ActiveBundle
- 
Blocks until bundle processing is finished. 
- close() - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient
-  
- close() - Method in class org.apache.beam.runners.fnexecution.control.SingleEnvironmentInstanceJobBundleFactory
- 
Deprecated. 
- close() - Method in class org.apache.beam.runners.fnexecution.data.GrpcDataService
-  
- close() - Method in class org.apache.beam.runners.fnexecution.environment.ProcessEnvironment
-  
- close() - Method in interface org.apache.beam.runners.fnexecution.environment.RemoteEnvironment
- 
. 
- close() - Method in class org.apache.beam.runners.fnexecution.environment.StaticRemoteEnvironment
-  
- close() - Method in class org.apache.beam.runners.fnexecution.logging.GrpcLoggingService
-  
- close() - Method in class org.apache.beam.runners.fnexecution.provisioning.StaticGrpcProvisionService
-  
- close() - Method in class org.apache.beam.runners.fnexecution.state.GrpcStateService
-  
- close() - Method in class org.apache.beam.runners.fnexecution.status.BeamWorkerStatusGrpcService
-  
- close() - Method in class org.apache.beam.runners.jet.processors.BoundedSourceP
-  
- close() - Method in class org.apache.beam.runners.jet.processors.UnboundedSourceP
-  
- close() - Method in class org.apache.beam.runners.jobsubmission.InMemoryJobService
-  
- close() - Method in class org.apache.beam.runners.portability.CloseableResource
- 
Closes the underlying resource. 
- close(T) - Method in interface org.apache.beam.runners.portability.CloseableResource.Closer
-  
- close() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-  
- close() - Method in class org.apache.beam.runners.twister2.translators.functions.DoFnFunction
-  
- close() - Method in class org.apache.beam.runners.twister2.translators.functions.Twister2SinkFunction
-  
- close() - Method in class org.apache.beam.sdk.expansion.service.ExpansionServer
-  
- close() - Method in class org.apache.beam.sdk.expansion.service.ExpansionService
-  
- close() - Method in class org.apache.beam.sdk.extensions.arrow.ArrowConversion.RecordBatchRowIterator
-  
- close() - Method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingReadableByteChannel
-  
- close() - Method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingSeekableByteChannel
-  
- close() - Method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingWritableByteChannel
-  
- close() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- close() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
-  
- close() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer
-  
- close() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
-  
- close() - Method in interface org.apache.beam.sdk.fn.data.CloseableFnDataReceiver
- 
. 
- close() - Method in interface org.apache.beam.sdk.fn.server.FnService
- 
. 
- close() - Method in class org.apache.beam.sdk.fn.server.GrpcFnServer
-  
- close() - Method in class org.apache.beam.sdk.fn.stream.DataStreams.ElementDelimitedOutputStream
-  
- close() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
- 
Closes the channel and returns the bundle result. 
- close() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
- 
Closes any ReadableByteChannelcreated for the current reader.
 
- close() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StorageClient
- 
Close the client object. 
- close() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.DatasetServiceImpl
-  
- close() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.WriteStreamServiceImpl
-  
- close() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.BigtableChangeStreamAccessor
-  
- close() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.DaoFactory
-  
- close() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
- 
Gracefully close the underlying netty channel. 
- close() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
-  
- close() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- close() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.BlockingCommitterImpl
-  
- close() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.ManagedFactoryImpl
-  
- close() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.UnboundedReaderImpl
-  
- close() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.ChangeStreamResultSet
- 
Closes the current change stream ResultSet.
 
- close() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerAccessor
-  
- close() - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- close() - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
-  
- close() - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantDeserializer
-  
- close() - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantSerializer
-  
- close() - Method in class org.apache.beam.sdk.io.solace.broker.JcsmpSessionService
-  
- close() - Method in interface org.apache.beam.sdk.io.solace.broker.MessageProducer
- 
Closes the message producer. 
- close() - Method in interface org.apache.beam.sdk.io.solace.broker.MessageReceiver
- 
Closes the message receiver. 
- close() - Method in class org.apache.beam.sdk.io.solace.broker.SessionService
- 
Gracefully closes the connection to the service. 
- close() - Method in class org.apache.beam.sdk.io.solace.broker.SolaceMessageProducer
-  
- close() - Method in class org.apache.beam.sdk.io.solace.broker.SolaceMessageReceiver
-  
- close() - Method in class org.apache.beam.sdk.io.Source.Reader
- 
Closes the reader. 
- close() - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.ThriftWriter
-  
- close() - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- close() - Method in class org.apache.beam.sdk.transforms.errorhandling.ErrorHandler.DefaultErrorHandler
-  
- close() - Method in class org.apache.beam.sdk.transforms.errorhandling.ErrorHandler.PTransformErrorHandler
-  
- CloseableFnDataReceiver<T> - Interface in org.apache.beam.sdk.fn.data
- 
A receiver of streamed data that can be closed. 
- CloseableResource<T> - Class in org.apache.beam.runners.portability
- 
An AutoCloseablethat wraps a resource that needs to be cleaned up but does not implementAutoCloseableitself.
 
- CloseableResource.CloseException - Exception in org.apache.beam.runners.portability
- 
An exception that wraps errors thrown while a resource is being closed. 
- CloseableResource.Closer<T> - Interface in org.apache.beam.runners.portability
- 
A function that knows how to clean up after a resource. 
- CloseableThrowingConsumer<ExceptionT extends java.lang.Exception,T> - Interface in org.apache.beam.sdk.function
- 
- CLOSESTREAM_COUNT - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
Counter for the total number of heartbeats identified during the execution of the Connector. 
- closeTo(double, double) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- CloudObject - Class in org.apache.beam.runners.dataflow.util
- 
A representation of an arbitrary Java object to be instantiated by Dataflow workers. 
- cloudObjectClassName() - Method in interface org.apache.beam.runners.dataflow.util.CloudObjectTranslator
- 
- cloudObjectClassName() - Method in class org.apache.beam.runners.dataflow.util.RowCoderCloudObjectTranslator
-  
- cloudObjectClassName() - Method in class org.apache.beam.runners.dataflow.util.SchemaCoderCloudObjectTranslator
-  
- CloudObjects - Class in org.apache.beam.runners.dataflow.util
- 
- CloudObjectTranslator<T> - Interface in org.apache.beam.runners.dataflow.util
- 
A translator that takes an object and creates a  CloudObject which can be converted back
 to the original object. 
- CloudPubsubTransforms - Class in org.apache.beam.sdk.io.gcp.pubsublite
- 
A class providing transforms between Cloud Pub/Sub and Pub/Sub Lite message types. 
- CloudResourceManagerOptions - Interface in org.apache.beam.sdk.extensions.gcp.options
- 
Properties needed when using Google CloudResourceManager with the Apache Beam SDK. 
- CloudVision - Class in org.apache.beam.sdk.extensions.ml
- 
Factory class for implementations of AnnotateImages.
 
- CloudVision() - Constructor for class org.apache.beam.sdk.extensions.ml.CloudVision
-  
- CloudVision.AnnotateImagesFromBytes - Class in org.apache.beam.sdk.extensions.ml
- 
Accepts  ByteString (encoded image contents) with optional  DoFn.SideInput with a  Map of  ImageContext to
 the image. 
- CloudVision.AnnotateImagesFromBytesWithContext - Class in org.apache.beam.sdk.extensions.ml
- 
Accepts  KVs of  ByteString (encoded image contents) and  ImageContext. 
- CloudVision.AnnotateImagesFromGcsUri - Class in org.apache.beam.sdk.extensions.ml
- 
Accepts  String (image URI on GCS) with optional  DoFn.SideInput with a  Map of  ImageContext to
 the image. 
- CloudVision.AnnotateImagesFromGcsUriWithContext - Class in org.apache.beam.sdk.extensions.ml
- 
Accepts  KVs of  String (GCS URI to the image) and  ImageContext. 
- CO_GBK_RESULT_SCHEMA - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- CodahaleCsvSink - Class in org.apache.beam.runners.spark.structuredstreaming.metrics.sink
- 
- CodahaleCsvSink(Properties, MetricRegistry, SecurityManager) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.metrics.sink.CodahaleCsvSink
- 
Constructor for Spark 3.1.x and earlier. 
- CodahaleCsvSink(Properties, MetricRegistry) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.metrics.sink.CodahaleCsvSink
- 
Constructor for Spark 3.2.x and later. 
- CodahaleGraphiteSink - Class in org.apache.beam.runners.spark.structuredstreaming.metrics.sink
- 
- CodahaleGraphiteSink(Properties, MetricRegistry, SecurityManager) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.metrics.sink.CodahaleGraphiteSink
- 
Constructor for Spark 3.1.x and earlier. 
- CodahaleGraphiteSink(Properties, MetricRegistry) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.metrics.sink.CodahaleGraphiteSink
- 
Constructor for Spark 3.2.x and later. 
- coder() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.TimerSpec
-  
- Coder<T> - Class in org.apache.beam.sdk.coders
- 
A  Coder<T> defines how to encode and decode values of type  T into
 byte streams. 
- Coder() - Constructor for class org.apache.beam.sdk.coders.Coder
-  
- CODER - Static variable in class org.apache.beam.sdk.io.iceberg.SnapshotInfo
-  
- Coder() - Constructor for class org.apache.beam.sdk.io.range.OffsetRange.Coder
-  
- coder - Variable in class org.apache.beam.sdk.io.ReadAllViaFileBasedSourceTransform
-  
- Coder.Context - Class in org.apache.beam.sdk.coders
- 
- Coder.NonDeterministicException - Exception in org.apache.beam.sdk.coders
- 
Exception thrown by  Coder.verifyDeterministic() if the encoding is not deterministic,
 including details of why the encoding is not deterministic. 
- CoderCloudObjectTranslatorRegistrar - Interface in org.apache.beam.runners.dataflow.util
- 
Coder authors have the ability to automatically have their  Coder registered with
 the Dataflow Runner by creating a  ServiceLoader entry and a concrete implementation of
 this interface. 
- coderConsistentWithEquals(Coder<T>, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<T>and values of typeT, the values are equal
 if and only if the encoded bytes are equal.
 
- coderConsistentWithEqualsInContext(Coder<T>, Coder.Context, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<T>,Coder.Context, and values of typeT, the values are equal if and only if the encoded bytes are equal, in anyCoder.Context.
 
- coderDecodeEncodeContentsEqual(Coder<CollectionT>, CollectionT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<Collection<T>>, and value of typeCollection<T>, encoding followed by decoding yields an equal value of typeCollection<T>, in anyCoder.Context.
 
- coderDecodeEncodeContentsEqualInContext(Coder<CollectionT>, Coder.Context, CollectionT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<Collection<T>>, and value of typeCollection<T>, encoding followed by decoding yields an equal value of typeCollection<T>, in the givenCoder.Context.
 
- coderDecodeEncodeContentsInSameOrder(Coder<IterableT>, IterableT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<Collection<T>>, and value of typeCollection<T>, encoding followed by decoding yields an equal value of typeCollection<T>, in anyCoder.Context.
 
- coderDecodeEncodeContentsInSameOrderInContext(Coder<IterableT>, Coder.Context, IterableT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<Iterable<T>>, and value of typeIterable<T>,
 encoding followed by decoding yields an equal value of typeCollection<T>, in the givenCoder.Context.
 
- coderDecodeEncodeEqual(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<T>, and value of typeT, encoding followed by
 decoding yields an equal value of typeT, in anyCoder.Context.
 
- coderDecodeEncodeEqualInContext(Coder<T>, Coder.Context, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<T>,Coder.Context, and value of typeT, encoding followed by decoding yields an equal value of typeT.
 
- coderDecodeEncodeInContext(Coder<T>, Coder.Context, T, Matcher<T>) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<T>,Coder.Context, and value of typeT, encoding followed by decoding yields a value of typeTand tests that the matcher
 succeeds on the values.
 
- coderDecodesBase64(Coder<T>, String, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-  
- coderDecodesBase64(Coder<T>, List<String>, List<T>) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-  
- coderDecodesBase64ContentsEqual(Coder<IterableT>, String, IterableT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-  
- coderDecodesBase64ContentsEqual(Coder<IterableT>, List<String>, List<IterableT>) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-  
- coderDeterministic(Coder<T>, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<T>, and values of typeT, if the values are
 equal then the encoded bytes are equal, in anyCoder.Context.
 
- coderDeterministicInContext(Coder<T>, Coder.Context, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<T>,Coder.Context, and values of typeT, if the values are equal then the encoded bytes are equal.
 
- coderEncodesBase64(Coder<T>, T, String) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-  
- coderEncodesBase64(Coder<T>, List<T>, List<String>) - Static method in class org.apache.beam.sdk.testing.CoderProperties
-  
- CoderException - Exception in org.apache.beam.sdk.coders
- 
An Exceptionthrown if there is a problem encoding or decoding a value.
 
- CoderException(String) - Constructor for exception org.apache.beam.sdk.coders.CoderException
-  
- CoderException(String, Throwable) - Constructor for exception org.apache.beam.sdk.coders.CoderException
-  
- CoderException(Throwable) - Constructor for exception org.apache.beam.sdk.coders.CoderException
-  
- coderFor(TypeDescriptor<T>, List<? extends Coder<?>>) - Method in class org.apache.beam.sdk.coders.CoderProvider
- 
Returns a Coder<T>to use for values of a particular type, given the Coders for each of
 the type's generic parameter types.
 
- coderFor(TypeDescriptor<T>, List<? extends Coder<?>>) - Method in class org.apache.beam.sdk.coders.DefaultCoder.DefaultCoderProviderRegistrar.DefaultCoderProvider
- 
- coderForFieldType(Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.SchemaCoder
-  
- coderFromCloudObject(CloudObject) - Static method in class org.apache.beam.runners.dataflow.util.CloudObjects
-  
- CoderHelpers - Class in org.apache.beam.runners.spark.coders
- 
Serialization utility class. 
- CoderHelpers - Class in org.apache.beam.runners.spark.structuredstreaming.translation.helpers
- 
Serialization utility class. 
- CoderHelpers.FromByteFunction<K,V> - Class in org.apache.beam.runners.spark.coders
- 
A function for converting a byte array pair to a key-value pair. 
- CoderProperties - Class in org.apache.beam.sdk.testing
- 
Properties for use in  Coder tests. 
- CoderProperties() - Constructor for class org.apache.beam.sdk.testing.CoderProperties
-  
- CoderProperties.TestElementByteSizeObserver - Class in org.apache.beam.sdk.testing
- 
An ElementByteSizeObserverthat records the observed element sizes for testing
 purposes.
 
- CoderProvider - Class in org.apache.beam.sdk.coders
- 
- CoderProvider() - Constructor for class org.apache.beam.sdk.coders.CoderProvider
-  
- CoderProviderRegistrar - Interface in org.apache.beam.sdk.coders
- 
Coder creators have the ability to automatically have their  coders
 registered with this SDK by creating a  ServiceLoader entry and a concrete implementation
 of this interface. 
- CoderProviders - Class in org.apache.beam.sdk.coders
- 
Static utility methods for creating and working with  CoderProviders. 
- CoderRegistry - Class in org.apache.beam.sdk.coders
- 
- coderSerializable(Coder<T>) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that the given Coder<T>can be correctly serialized and deserialized.
 
- CoderSizeEstimator<T> - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.estimator
- 
This class is used to estimate the size in bytes of a given element. 
- CoderSizeEstimator(Coder<T>) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.estimator.CoderSizeEstimator
-  
- CoGbkResult - Class in org.apache.beam.sdk.transforms.join
- 
- CoGbkResult(CoGbkResultSchema, Iterable<RawUnionValue>) - Constructor for class org.apache.beam.sdk.transforms.join.CoGbkResult
- 
- CoGbkResult(CoGbkResultSchema, Iterable<RawUnionValue>, int, int) - Constructor for class org.apache.beam.sdk.transforms.join.CoGbkResult
-  
- CoGbkResult.CoGbkResultCoder - Class in org.apache.beam.sdk.transforms.join
- 
- CoGbkResultSchema - Class in org.apache.beam.sdk.transforms.join
- 
- CoGbkResultSchema(TupleTagList) - Constructor for class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
- 
Builds a schema from a tuple of TupleTag<?>s.
 
- CoGroup - Class in org.apache.beam.sdk.schemas.transforms
- 
A transform that performs equijoins across multiple schema  PCollections. 
- CoGroup() - Constructor for class org.apache.beam.sdk.schemas.transforms.CoGroup
-  
- CoGroup.By - Class in org.apache.beam.sdk.schemas.transforms
- 
Defines the set of fields to extract for the join key, as well as other per-input join options. 
- CoGroup.ExpandCrossProduct - Class in org.apache.beam.sdk.schemas.transforms
- 
A  PTransform that calculates the cross-product join. 
- CoGroup.Impl - Class in org.apache.beam.sdk.schemas.transforms
- 
The implementing PTransform. 
- CoGroup.Result - Class in org.apache.beam.sdk.schemas.transforms
-  
- CoGroupByKey<K> - Class in org.apache.beam.sdk.transforms.join
- 
- collect(String, Dataset<T>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.EvaluationContext
- 
The purpose of this utility is to mark the evaluation of Spark actions, both during Pipeline
 translation, when evaluation is required, and when finally evaluating the pipeline. 
- COLLECTION_TYPES - Static variable in enum org.apache.beam.sdk.schemas.Schema.TypeName
-  
- CollectionCoder<T> - Class in org.apache.beam.sdk.coders
- 
- CollectionCoder(Coder<T>) - Constructor for class org.apache.beam.sdk.coders.CollectionCoder
-  
- collectionEncoder(Encoder<T>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers
- 
Creates a Spark EncoderofArrayTypefor JavaCollections with nullable
 elements.
 
- collectionEncoder(Encoder<T>, boolean) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers
- 
Creates a Spark EncoderofArrayTypefor JavaCollections.
 
- column(SqlParserPos, SqlIdentifier, SqlDataTypeSpec, SqlNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlDdlNodes
- 
Creates a column declaration. 
- Column() - Constructor for class org.apache.beam.sdk.io.clickhouse.TableSchema.Column
-  
- Column() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerSchema.Column
-  
- COLUMN_CREATED_AT - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataAdminDao
- 
Metadata table column name for the timestamp at which the partition row was first created. 
- COLUMN_END_TIMESTAMP - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataAdminDao
- 
Metadata table column name for the timestamp to end the change stream query of the partition. 
- COLUMN_FAMILIES - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
-  
- COLUMN_FINISHED_AT - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataAdminDao
- 
- COLUMN_HEARTBEAT_MILLIS - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataAdminDao
- 
Metadata table column name for the change stream query heartbeat interval in millis. 
- COLUMN_PARENT_TOKENS - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataAdminDao
- 
Metadata table column name for parent partition tokens. 
- COLUMN_PARTITION_TOKEN - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataAdminDao
- 
Metadata table column name for the partition token. 
- COLUMN_RUNNING_AT - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataAdminDao
- 
- COLUMN_SCHEDULED_AT - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataAdminDao
- 
Metadata table column name for the timestamp at which the partition was scheduled by the  DetectNewPartitionsDoFn SDF. 
- COLUMN_START_TIMESTAMP - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataAdminDao
- 
Metadata table column name for the timestamp to start the change stream query of the partition. 
- COLUMN_STATE - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataAdminDao
- 
Metadata table column name for the state that the partition is currently in. 
- COLUMN_WATERMARK - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataAdminDao
- 
Metadata table column name for the current watermark of the partition. 
- columns() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema
-  
- COLUMNS - Static variable in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiCDC
-  
- COLUMNS_MAPPING - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.RowUtils
-  
- columnType() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.Column
-  
- ColumnType() - Constructor for class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- ColumnType - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.model
- 
Defines a column type from a Cloud Spanner table with the following information: column name,
 column type, flag indicating if column is primary key and column position in the table. 
- ColumnType(String, TypeCode, boolean, long) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ColumnType
-  
- Combine - Class in org.apache.beam.sdk.transforms
- 
PTransforms for combiningPCollectionelements globally and per-key.
 
- combine(Iterable<? extends Instant>) - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
- 
- combine(Instant...) - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
- 
- Combine.AccumulatingCombineFn<InputT,AccumT extends Combine.AccumulatingCombineFn.Accumulator<InputT,AccumT,OutputT>,OutputT> - Class in org.apache.beam.sdk.transforms
- 
- Combine.AccumulatingCombineFn.Accumulator<InputT,AccumT,OutputT> - Interface in org.apache.beam.sdk.transforms
- 
The type of mutable accumulator values used by this AccumulatingCombineFn.
 
- Combine.BinaryCombineDoubleFn - Class in org.apache.beam.sdk.transforms
- 
An abstract subclass of  Combine.CombineFn for implementing combiners that are more easily and
 efficiently expressed as binary operations on  doubles. 
- Combine.BinaryCombineFn<V> - Class in org.apache.beam.sdk.transforms
- 
An abstract subclass of  Combine.CombineFn for implementing combiners that are more easily
 expressed as binary operations. 
- Combine.BinaryCombineIntegerFn - Class in org.apache.beam.sdk.transforms
- 
An abstract subclass of  Combine.CombineFn for implementing combiners that are more easily and
 efficiently expressed as binary operations on  ints 
- Combine.BinaryCombineLongFn - Class in org.apache.beam.sdk.transforms
- 
An abstract subclass of  Combine.CombineFn for implementing combiners that are more easily and
 efficiently expressed as binary operations on  longs. 
- Combine.CombineFn<InputT,AccumT,OutputT> - Class in org.apache.beam.sdk.transforms
- 
A CombineFn<InputT, AccumT, OutputT>specifies how to combine a collection of input
 values of typeInputTinto a single output value of typeOutputT.
 
- Combine.Globally<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
- 
Combine.Globally<InputT, OutputT> takes a  PCollection<InputT> and returns a
  PCollection<OutputT> whose elements are the result of combining all the elements in
 each window of the input  PCollection, using a specified  CombineFn<InputT, AccumT, OutputT>. 
- Combine.GloballyAsSingletonView<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
- 
Combine.GloballyAsSingletonView<InputT, OutputT> takes a  PCollection<InputT>
 and returns a  PCollectionView<OutputT> whose elements are the result of combining all
 the elements in each window of the input  PCollection, using a specified  CombineFn<InputT, AccumT, OutputT>. 
- Combine.GroupedValues<K,InputT,OutputT> - Class in org.apache.beam.sdk.transforms
- 
GroupedValues<K, InputT, OutputT> takes a  PCollection<KV<K, Iterable<InputT>>>,
 such as the result of  GroupByKey, applies a specified  CombineFn<InputT, AccumT, OutputT> to each of the input  KV<K, Iterable<InputT>>
 elements to produce a combined output  KV<K, OutputT> element, and returns a  PCollection<KV<K, OutputT>> containing all the combined output elements. 
- Combine.Holder<V> - Class in org.apache.beam.sdk.transforms
- 
Holds a single value value of type Vwhich may or may not be present.
 
- Combine.IterableCombineFn<V> - Class in org.apache.beam.sdk.transforms
- 
- Combine.PerKey<K,InputT,OutputT> - Class in org.apache.beam.sdk.transforms
- 
PerKey<K, InputT, OutputT>takes aPCollection<KV<K, InputT>>, groups it by
 key, applies a combining function to theInputTvalues associated with each key to
 produce a combinedOutputTvalue, and returns aPCollection<KV<K, OutputT>>representing a map from each distinct key of the inputPCollectionto the corresponding
 combined value.
 
- Combine.PerKeyWithHotKeyFanout<K,InputT,OutputT> - Class in org.apache.beam.sdk.transforms
- 
- Combine.SimpleCombineFn<V> - Class in org.apache.beam.sdk.transforms
- 
Deprecated. 
- CombineAsIterable<T> - Class in org.apache.beam.sdk.io.gcp.bigquery
-  
- CombineAsIterable() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.CombineAsIterable
-  
- CombineFieldsByFields() - Constructor for class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
-  
- combineFn() - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CountIf
-  
- CombineFn() - Constructor for class org.apache.beam.sdk.transforms.Combine.CombineFn
-  
- combineFn() - Static method in class org.apache.beam.sdk.transforms.Count
- 
- combineFn() - Static method in class org.apache.beam.sdk.transforms.Latest
- 
- combineFn(int) - Static method in class org.apache.beam.sdk.transforms.Sample
- 
Returns a  Combine.CombineFn that computes a fixed-sized uniform sample of its inputs. 
- CombineFnBase - Class in org.apache.beam.sdk.transforms
- 
For internal use only; no backwards-compatibility guarantees. 
- CombineFnBase() - Constructor for class org.apache.beam.sdk.transforms.CombineFnBase
-  
- CombineFnBase.GlobalCombineFn<InputT,AccumT,OutputT> - Interface in org.apache.beam.sdk.transforms
- 
For internal use only; no backwards-compatibility guarantees. 
- CombineFns - Class in org.apache.beam.sdk.transforms
- 
Static utility methods that create combine function instances. 
- CombineFns() - Constructor for class org.apache.beam.sdk.transforms.CombineFns
-  
- CombineFns.CoCombineResult - Class in org.apache.beam.sdk.transforms
- 
A tuple of outputs produced by a composed combine functions. 
- CombineFns.ComposeCombineFnBuilder - Class in org.apache.beam.sdk.transforms
- 
- CombineFns.ComposedCombineFn<DataT> - Class in org.apache.beam.sdk.transforms
- 
- CombineFns.ComposedCombineFnWithContext<DataT> - Class in org.apache.beam.sdk.transforms
- 
- CombineFnTester - Class in org.apache.beam.sdk.testing
- 
- CombineFnTester() - Constructor for class org.apache.beam.sdk.testing.CombineFnTester
-  
- CombineFnWithContext() - Constructor for class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
-  
- CombineWithContext - Class in org.apache.beam.sdk.transforms
- 
This class contains combine functions that have access to PipelineOptionsand side inputs
 throughCombineWithContext.Context.
 
- CombineWithContext() - Constructor for class org.apache.beam.sdk.transforms.CombineWithContext
-  
- CombineWithContext.CombineFnWithContext<InputT,AccumT,OutputT> - Class in org.apache.beam.sdk.transforms
- 
A combine function that has access to PipelineOptionsand side inputs throughCombineWithContext.Context.
 
- CombineWithContext.Context - Class in org.apache.beam.sdk.transforms
- 
Information accessible to all methods in CombineFnWithContextandKeyedCombineFnWithContext.
 
- CombineWithContext.RequiresContextInternal - Interface in org.apache.beam.sdk.transforms
- 
An internal interface for signaling that a GloballyCombineFnor aPerKeyCombineFnneeds to accessCombineWithContext.Context.
 
- combining(Combine.CombineFn<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
- combining(CombineWithContext.CombineFnWithContext<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
For internal use only; no backwards compatibility guarantees 
- combining(Coder<AccumT>, Combine.CombineFn<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
Identical to #combining(CombineFn), but with an accumulator coder explicitly supplied.
 
- combining(Coder<AccumT>, CombineWithContext.CombineFnWithContext<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
For internal use only; no backwards compatibility guarantees 
- combiningFromInputInternal(Coder<InputT>, Combine.CombineFn<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
For internal use only; no backwards-compatibility guarantees. 
- CombiningState<InputT,AccumT,OutputT> - Interface in org.apache.beam.sdk.state
- 
A  ReadableState cell defined by a  Combine.CombineFn, accepting multiple input values,
 combining them as specified into accumulators, and producing a single output value. 
- comment(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.Table.Builder
-  
- commit() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- commitOffset(Offset) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.BlockingCommitterImpl
-  
- commitOffsets() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Enable committing record offset. 
- commitOffsetsInFinalize() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Finalized offsets are committed to Kafka. 
- commitWriteStreams(String, Iterable<String>) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.WriteStreamService
- 
Commit write streams of type PENDING. 
- commitWriteStreams(String, Iterable<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.WriteStreamServiceImpl
-  
- commitWriteStreams(String, Iterable<String>) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- commonPrefixLength(RandomAccessData, RandomAccessData) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.UnsignedLexicographicalComparator
- 
Compute the length of the common prefix of the two provided sets of bytes. 
- compact(AccumT) - Method in class org.apache.beam.sdk.extensions.sql.TypedCombineFnDelegate
-  
- compact(AccumT) - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
- 
Returns an accumulator that represents the same logical value as the input accumulator, but
 may have a more compact representation. 
- compact(List<V>) - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
-  
- compact(Object[]) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
-  
- compact(Object[], CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
-  
- compact(AccumT, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
- 
Returns an accumulator that represents the same logical value as the input accumulator, but
 may have a more compact representation. 
- compare(JobMessage, JobMessage) - Method in class org.apache.beam.runners.dataflow.util.MonitoringUtil.TimeStampComparator
-  
- compare(RandomAccessData, RandomAccessData) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.UnsignedLexicographicalComparator
-  
- compare(RandomAccessData, RandomAccessData, int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.UnsignedLexicographicalComparator
- 
Compare the two sets of bytes starting at the given offset. 
- compare(Row, Row) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSortRel.BeamSqlRowComparator
-  
- compare(T, T) - Method in class org.apache.beam.sdk.transforms.Top.Largest
- 
Deprecated. 
- compare(T, T) - Method in class org.apache.beam.sdk.transforms.Top.Natural
-  
- compare(T, T) - Method in class org.apache.beam.sdk.transforms.Top.Reversed
-  
- compare(T, T) - Method in class org.apache.beam.sdk.transforms.Top.Smallest
- 
Deprecated. 
- compare(KV<K, V>, KV<K, V>) - Method in class org.apache.beam.sdk.values.KV.OrderByKey
-  
- compare(KV<K, V>, KV<K, V>) - Method in class org.apache.beam.sdk.values.KV.OrderByValue
-  
- compareSchemaField(Schema.Field, Schema.Field) - Static method in class org.apache.beam.sdk.io.jdbc.SchemaUtil
- 
compares two fields. 
- compareTo(ByteArray) - Method in class org.apache.beam.runners.spark.util.ByteArray
-  
- compareTo(Path) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- compareTo(ContiguousSequenceRange) - Method in class org.apache.beam.sdk.extensions.ordered.ContiguousSequenceRange
-  
- compareTo(Object) - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- compareTo(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKey
- 
ByteKey implements  Comparable<ByteKey> by comparing the arrays
 in lexicographic order. 
- compareTo(RedisCursor) - Method in class org.apache.beam.sdk.io.redis.RedisCursor
- 
RedisCursor implements  Comparable<RedisCursor> by transforming
 the cursors to an index of the Redis table. 
- compareTo(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
-  
- comparing(SerializableFunction<? super T, ? extends V>) - Static method in interface org.apache.beam.sdk.transforms.SerializableComparator
- 
- comparingNullFirst(Function<? super T, ? extends K>) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
-  
- CompatibilityError() - Constructor for class org.apache.beam.sdk.schemas.transforms.Cast.CompatibilityError
-  
- compile(List<CEPPattern>, Schema) - Static method in class org.apache.beam.sdk.extensions.sql.impl.nfa.NFA
-  
- CompileException(DiagnosticCollector<?>) - Constructor for exception org.apache.beam.sdk.schemas.transforms.providers.StringCompiler.CompileException
-  
- complete() - Method in class org.apache.beam.runners.jet.processors.BoundedSourceP
-  
- complete() - Method in class org.apache.beam.runners.jet.processors.ImpulseP
-  
- complete() - Method in class org.apache.beam.runners.jet.processors.StatefulParDoP
-  
- complete() - Method in class org.apache.beam.runners.jet.processors.UnboundedSourceP
-  
- complete() - Method in class org.apache.beam.runners.jet.processors.ViewP
-  
- complete() - Method in class org.apache.beam.runners.jet.processors.WindowGroupP
-  
- complete(List<TimestampedValue<OutputT>>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
- 
Constructs a  Watch.Growth.PollResult with the given outputs and declares that there will be no
 new outputs for the current input. 
- complete(Instant, List<OutputT>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
- 
- completed() - Method in interface org.apache.beam.runners.local.PipelineMessageReceiver
- 
Report that the pipeline has successfully completed. 
- complexityFactor - Variable in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator
-  
- COMPONENT_ENCODINGS - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- compose() - Static method in class org.apache.beam.sdk.transforms.CombineFns
- 
- compose(SerializableFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.PTransform
- 
For a SerializableFunction<InputT, OutputT>fn, returns aPTransformgiven by applyingfn.apply(v)to the inputPCollection<InputT>.
 
- compose(String, SerializableFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.PTransform
- 
- ComposeCombineFnBuilder() - Constructor for class org.apache.beam.sdk.transforms.CombineFns.ComposeCombineFnBuilder
-  
- COMPOSITE_TYPES - Static variable in enum org.apache.beam.sdk.schemas.Schema.TypeName
-  
- CompressedReader(CompressedSource<T>, FileBasedSource.FileBasedReader<T>) - Constructor for class org.apache.beam.sdk.io.CompressedSource.CompressedReader
- 
Create a CompressedReaderfrom aCompressedSourceand delegate reader.
 
- CompressedSource<T> - Class in org.apache.beam.sdk.io
- 
A Source that reads from compressed files. 
- CompressedSource.CompressedReader<T> - Class in org.apache.beam.sdk.io
- 
- CompressedSource.CompressionMode - Enum in org.apache.beam.sdk.io
- 
- CompressedSource.DecompressingChannelFactory - Interface in org.apache.beam.sdk.io
- 
Factory interface for creating channels that decompress the content of an underlying channel. 
- Compression - Enum in org.apache.beam.sdk.io
- 
Various compression types for reading/writing files. 
- compute(Partition, TaskContext) - Method in class org.apache.beam.runners.spark.io.SourceRDD.Bounded
-  
- compute(Partition, TaskContext) - Method in class org.apache.beam.runners.spark.io.SourceRDD.Unbounded
-  
- compute(Iterator<WindowedValue<T>>, RecordCollector<WindowedValue<T>>) - Method in class org.apache.beam.runners.twister2.translators.functions.AssignWindowsFunction
-  
- compute(Iterator<WindowedValue<InputT>>, RecordCollector<RawUnionValue>) - Method in class org.apache.beam.runners.twister2.translators.functions.DoFnFunction
-  
- compute(Iterator<RawUnionValue>, RecordCollector<WindowedValue<OutputT>>) - Method in class org.apache.beam.runners.twister2.translators.functions.OutputTagFilter
-  
- computeIfAbsent(K, Function<? super K, ? extends V>) - Method in interface org.apache.beam.sdk.state.MapState
- 
A deferred read-followed-by-write. 
- computeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter
-  
- computeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
-  
- concat(List<T>, List<T>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.utils.ScalaInterop
-  
- concat(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- concat(String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- concat(String, String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- concat(String, String, String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- concat(String, String, String, String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- CONCAT - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- concat(Iterable<T>...) - Static method in class org.apache.beam.sdk.fn.stream.PrefetchableIterables
- 
Concatentates the Iterables.
 
- concat(Iterator<T>...) - Static method in class org.apache.beam.sdk.fn.stream.PrefetchableIterators
- 
Concatentates the Iterators.
 
- CONCAT_FIELD_NAMES - Static variable in class org.apache.beam.sdk.schemas.utils.SelectHelpers
- 
This policy keeps all levels of a name. 
- CONCAT_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.BeamBuiltinMethods
-  
- Concatenate() - Constructor for class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Concatenate
-  
- ConcatenateAsIterable() - Constructor for class org.apache.beam.runners.flink.FlinkStreamingAggregationsTranslators.ConcatenateAsIterable
-  
- concatFieldNames() - Method in class org.apache.beam.sdk.schemas.transforms.Select.Flattened
- 
For nested fields, concatenate all the names separated by a _ character in the flattened
 schema. 
- concatIterators(Iterator<Iterator<T>>) - Static method in class org.apache.beam.sdk.fn.stream.PrefetchableIterators
-  
- CONCRETE_CLASS - Static variable in class org.apache.beam.sdk.io.WriteFiles
- 
For internal use by runners. 
- config() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- config() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
-  
- Config() - Constructor for class org.apache.beam.sdk.managed.testing.TestSchemaTransformProvider.Config
-  
- Configuration() - Constructor for class org.apache.beam.io.debezium.DebeziumTransformRegistrar.ReadBuilder.Configuration
-  
- configuration - Variable in class org.apache.beam.runners.jobsubmission.JobServerDriver
-  
- Configuration() - Constructor for class org.apache.beam.sdk.expansion.service.WindowIntoTransformProvider.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.extensions.schemaio.expansion.ExternalSchemaIOTransformRegistrar.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.extensions.sql.expansion.ExternalSqlTransformRegistrar.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.ReadDataBuilder.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.WriteBuilder.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.ReadBuilder.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.Write.External.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.io.snowflake.crosslanguage.ReadBuilder.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.io.snowflake.crosslanguage.WriteBuilder.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.schemas.transforms.providers.FlattenTransformProvider.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.schemas.transforms.providers.JavaExplodeTransformProvider.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.schemas.transforms.providers.JavaFilterTransformProvider.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.schemas.transforms.providers.JavaRowUdf.Configuration
-  
- Configuration() - Constructor for class org.apache.beam.sdk.schemas.transforms.providers.LoggingTransformProvider.Configuration
-  
- configurationClass() - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.expansion.service.WindowIntoTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.csv.providers.CsvWriteTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformProvider
- 
- configurationClass() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryExportReadSchemaTransformProvider
- 
Returns the expected class of the configuration. 
- configurationClass() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.json.providers.JsonWriteTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadProvider
- 
Returns the expected class of the configuration. 
- configurationClass() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteProvider
- 
Returns the expected class of the configuration. 
- configurationClass() - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.schemas.transforms.providers.FlattenTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaExplodeTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaFilterTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.schemas.transforms.providers.LoggingTransformProvider
-  
- configurationClass() - Method in class org.apache.beam.sdk.schemas.transforms.TypedSchemaTransformProvider
-  
- ConfigurationLocator() - Constructor for class org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptions.ConfigurationLocator
-  
- configurationSchema() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSchemaIOProvider
- 
Returns the expected schema of the configuration object. 
- configurationSchema() - Method in class org.apache.beam.sdk.extensions.sql.expansion.SqlTransformSchemaTransformProvider
-  
- configurationSchema() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySchemaIOProvider
- 
Returns the expected schema of the configuration object. 
- configurationSchema() - Method in class org.apache.beam.sdk.io.gcp.datastore.DataStoreV1SchemaIOProvider
- 
Returns the expected schema of the configuration object. 
- configurationSchema() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubSchemaIOProvider
- 
Returns the expected schema of the configuration object. 
- configurationSchema() - Method in class org.apache.beam.sdk.io.jdbc.JdbcSchemaIOProvider
- 
Returns the expected schema of the configuration object. 
- configurationSchema() - Method in interface org.apache.beam.sdk.schemas.io.SchemaIOProvider
- 
Returns the expected schema of the configuration object. 
- configurationSchema() - Method in interface org.apache.beam.sdk.schemas.transforms.SchemaTransformProvider
- 
Returns the expected schema of the configuration object. 
- configurationSchema() - Method in class org.apache.beam.sdk.schemas.transforms.TypedSchemaTransformProvider
-  
- configure(Map<String, ?>, boolean) - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantDeserializer
-  
- configure(Map<String, ?>, boolean) - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantSerializer
-  
- configure() - Static method in class org.apache.beam.sdk.transforms.windowing.Window
- 
Returns a new builder for a  Window transform for setting windowing parameters other
 than the windowing function. 
- ConfigWrapper<T extends PluginConfig> - Class in org.apache.beam.sdk.io.cdap
- 
Class for building  PluginConfig object of the specific class . 
- ConfigWrapper(Class<T>) - Constructor for class org.apache.beam.sdk.io.cdap.ConfigWrapper
-  
- ConfluentSchemaRegistryDeserializerProvider<T> - Class in org.apache.beam.sdk.io.kafka
- 
- connect(String, Properties) - Method in class org.apache.beam.sdk.extensions.sql.impl.JdbcDriver
- 
Configures Beam-specific options and opens a JDBC connection to Calcite. 
- connect(TableProvider, PipelineOptions) - Static method in class org.apache.beam.sdk.extensions.sql.impl.JdbcDriver
- 
- connect() - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
- 
Connect to the Redis instance. 
- connect() - Method in class org.apache.beam.sdk.io.solace.broker.JcsmpSessionService
-  
- connect() - Method in class org.apache.beam.sdk.io.solace.broker.SessionService
- 
Establishes a connection to the service. 
- CONNECT_STRING_PREFIX - Static variable in class org.apache.beam.sdk.extensions.sql.impl.JdbcDriver
-  
- connection() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- connectionAcquisitionTimeout(Integer) - Method in class org.apache.beam.sdk.io.aws2.common.HttpClientConfiguration.Builder
- 
Milliseconds to wait when acquiring a connection from the pool before giving up and timing
 out. 
- connectionAcquisitionTimeout() - Method in class org.apache.beam.sdk.io.aws2.common.HttpClientConfiguration
- 
Milliseconds to wait when acquiring a connection from the pool before giving up and timing out. 
- ConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- ConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.ConnectionConfiguration
-  
- ConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
-  
- ConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.solr.SolrIO.ConnectionConfiguration
-  
- ConnectionManager - Class in org.apache.beam.sdk.io.cassandra
-  
- ConnectionManager() - Constructor for class org.apache.beam.sdk.io.cassandra.ConnectionManager
-  
- connectionMaxIdleTime(Integer) - Method in class org.apache.beam.sdk.io.aws2.common.HttpClientConfiguration.Builder
- 
Maximum milliseconds a connection should be allowed to remain open while idle. 
- connectionMaxIdleTime() - Method in class org.apache.beam.sdk.io.aws2.common.HttpClientConfiguration
- 
Maximum milliseconds a connection should be allowed to remain open while idle. 
- connectionTimeout(Integer) - Method in class org.apache.beam.sdk.io.aws2.common.HttpClientConfiguration.Builder
- 
Milliseconds to wait when initially establishing a connection before giving up and timing
 out. 
- connectionTimeout() - Method in class org.apache.beam.sdk.io.aws2.common.HttpClientConfiguration
- 
Milliseconds to wait when initially establishing a connection before giving up and timing out. 
- connectionTimeToLive(Integer) - Method in class org.apache.beam.sdk.io.aws2.common.HttpClientConfiguration.Builder
- 
Maximum milliseconds a connection should be allowed to remain open, regardless of usage
 frequency. 
- connectionTimeToLive() - Method in class org.apache.beam.sdk.io.aws2.common.HttpClientConfiguration
- 
Maximum milliseconds a connection should be allowed to remain open, regardless of usage
 frequency. 
- ConnectorConfiguration() - Constructor for class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
-  
- Connectors - Enum in org.apache.beam.io.debezium
- 
Enumeration of debezium connectors. 
- consistentWithEquals() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.runners.fnexecution.wire.ByteStringCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BigEndianShortCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BitSetCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BooleanCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.ByteCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.Coder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.DequeCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.DoubleCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.DurationCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.FloatCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.InstantCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.KvCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
- 
LengthPrefixCoderis consistent with equals if the nestedCoderis.
 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.ListCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.MapCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.NullableCoder
- 
NullableCoderis consistent with equals if the nestedCoderis.
 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.SerializableCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.SortedMapCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.TimestampPrefixingWindowCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.VarIntCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.VarLongCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.coders.ZstdCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
- 
Returns  true if this  Coder is injective with respect to  Object.equals(java.lang.Object). 
- consistentWithEquals() - Method in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow.Coder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.io.fs.MetadataCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.io.fs.MetadataCoderV2
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.io.fs.ResourceIdCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.io.range.OffsetRange.Coder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
-  
- consistentWithEquals() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
-  
- ConsoleIO - Class in org.apache.beam.runners.spark.io
- 
Print to console. 
- ConsoleIO.Write - Class in org.apache.beam.runners.spark.io
- 
Write to console. 
- ConsoleIO.Write.Unbound<T> - Class in org.apache.beam.runners.spark.io
- 
- constant(FileBasedSink.FilenamePolicy, SerializableFunction<UserT, OutputT>) - Static method in class org.apache.beam.sdk.io.DynamicFileDestinations
- 
- constant(FileBasedSink.FilenamePolicy) - Static method in class org.apache.beam.sdk.io.DynamicFileDestinations
- 
A specialization of #constant(FilenamePolicy, SerializableFunction)for the case where
 UserT and OutputT are the same type and the format function is the identity.
 
- constant(OutT) - Static method in class org.apache.beam.sdk.transforms.SerializableFunctions
-  
- CONSTANT_WINDOW_SIZE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
-  
- constantDestinations(FileBasedSink.FilenamePolicy, Schema, Map<String, Object>, CodecFactory, SerializableFunction<UserT, OutputT>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- constantDestinations(FileBasedSink.FilenamePolicy, Schema, Map<String, Object>, CodecFactory, SerializableFunction<UserT, OutputT>, AvroSink.DatumWriterFactory<OutputT>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- constructFilter(List<RexNode>) - Method in class org.apache.beam.sdk.extensions.sql.meta.BaseBeamTable
-  
- constructFilter(List<RexNode>) - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTable
- 
Generate an IO implementation of BeamSqlTableFilterfor predicate push-down.
 
- constructFilter(List<RexNode>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigtable.BigtableTable
-  
- constructFilter(List<RexNode>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable
-  
- constructName(ResourceId, String, String, int, int, String, String) - Static method in class org.apache.beam.sdk.io.DefaultFilenamePolicy
- 
Constructs a fully qualified name from components. 
- consumesProjection() - Method in interface org.apache.beam.sdk.schemas.ProjectionConsumer
- 
- contains(PCollectionView<T>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.CachedSideInputReader
-  
- contains(PCollectionView<T>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.SparkSideInputReader
-  
- contains(PCollectionView<T>) - Method in class org.apache.beam.runners.spark.util.CachedSideInputReader
-  
- contains(PCollectionView<T>) - Method in class org.apache.beam.runners.spark.util.SparkSideInputReader
-  
- contains(PCollectionView<T>) - Method in class org.apache.beam.runners.twister2.utils.Twister2SideInputReader
-  
- contains(Descriptors.Descriptor) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
-  
- contains(T) - Method in interface org.apache.beam.sdk.state.SetState
- 
- contains(T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- contains(Coder<T>, T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- contains(SerializableMatcher<? super T>...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- contains(List<SerializableMatcher<? super T>>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- contains(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
- 
Returns whether this window contains the given window. 
- containsInAnyOrder(T...) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
- 
Asserts that the iterable in question contains the provided elements. 
- containsInAnyOrder(SerializableMatcher<? super T>...) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
- 
Asserts that the iterable in question matches the provided elements. 
- containsInAnyOrder() - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
- 
- containsInAnyOrder(Iterable<T>) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
- 
Asserts that the iterable in question contains the provided elements. 
- containsInAnyOrder(T...) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
- 
Checks that the Iterablecontains the expected elements, in any order.
 
- containsInAnyOrder(Iterable<T>) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
- 
Checks that the Iterablecontains the expected elements, in any order.
 
- containsInAnyOrder(SerializableMatcher<? super T>...) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
- 
Checks that the Iterablecontains elements that match the provided matchers, in any
 order.
 
- containsInAnyOrder() - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-  
- containsInAnyOrder(T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- containsInAnyOrder(Coder<T>, T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- containsInAnyOrder(SerializableMatcher<? super T>...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- containsInAnyOrder(Collection<SerializableMatcher<? super T>>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- containsKey(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
- 
Returns  true if the specified  ByteKey is contained within this range. 
- containsKey(Object) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
-  
- containsKey(K) - Method in interface org.apache.beam.sdk.state.MultimapState
- 
- containsSeekableInput(RelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
- 
This method returns whether any of the children of the relNode are Seekable. 
- containsString(String) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- containsValue(Object) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
-  
- Context(boolean) - Constructor for class org.apache.beam.sdk.coders.Coder.Context
- 
Deprecated. 
- Context(TableDataInsertAllResponse.InsertErrors) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy.Context
-  
- Context() - Constructor for class org.apache.beam.sdk.schemas.utils.SchemaZipFold.Context
-  
- Context() - Constructor for class org.apache.beam.sdk.transforms.CombineWithContext.Context
-  
- Context() - Constructor for class org.apache.beam.sdk.transforms.Contextful.Fn.Context
-  
- Contextful<ClosureT> - Class in org.apache.beam.sdk.transforms
- 
Pair of a bit of user code (a "closure") and the  Requirements needed to run it. 
- Contextful.Fn<InputT,OutputT> - Interface in org.apache.beam.sdk.transforms
- 
A function from an input to an output that may additionally access  Contextful.Fn.Context when
 computing the result. 
- Contextful.Fn.Context - Class in org.apache.beam.sdk.transforms
- 
- ContextualTextIO - Class in org.apache.beam.sdk.io.contextualtextio
- 
PTransforms that read text files and collect contextual information of the elements in
 the input. 
- ContextualTextIO.Read - Class in org.apache.beam.sdk.io.contextualtextio
- 
- ContextualTextIO.ReadFiles - Class in org.apache.beam.sdk.io.contextualtextio
- 
- ContiguousSequenceRange - Class in org.apache.beam.sdk.extensions.ordered
- 
A range of contiguous event sequences and the latest timestamp of the events in the range. 
- ContiguousSequenceRange() - Constructor for class org.apache.beam.sdk.extensions.ordered.ContiguousSequenceRange
-  
- continuously(Duration, Watch.Growth.TerminationCondition<String, ?>, boolean) - Method in class org.apache.beam.sdk.io.FileIO.Match
- 
- continuously(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.FileIO.Match
- 
- continuously(Duration, Watch.Growth.TerminationCondition<String, ?>, boolean) - Method in class org.apache.beam.sdk.io.FileIO.MatchAll
- 
Like Match#continuously(Duration, TerminationCondition, boolean).
 
- continuously(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.FileIO.MatchAll
- 
Like Match#continuously(Duration, TerminationCondition).
 
- continuously(Duration, Watch.Growth.TerminationCondition<String, ?>, boolean) - Method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
- 
Continuously watches for new files at the given interval until the given termination
 condition is reached, where the input to the condition is the filepattern. 
- continuously(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
- 
Continuously watches for new files at the given interval until the given termination
 condition is reached, where the input to the condition is the filepattern. 
- control(StreamObserver<BeamFnApi.InstructionRequest>) - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService
- 
Called by gRPC for each incoming connection from an SDK harness, and enqueue an available SDK
 harness client. 
- ControlClientPool - Interface in org.apache.beam.runners.fnexecution.control
- 
- ControlClientPool.Sink - Interface in org.apache.beam.runners.fnexecution.control
- 
- ControlClientPool.Source - Interface in org.apache.beam.runners.fnexecution.control
- 
- ConversionContext - Class in org.apache.beam.sdk.extensions.sql.zetasql.translation
- 
Conversion context, some rules need this data to convert the nodes. 
- ConversionOptions() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.ConversionOptions
-  
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamCalcRule
-  
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamEnumerableConverterRule
-  
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIntersectRule
-  
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIOSinkRule
-  
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamMatchRule
-  
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamMinusRule
-  
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSideInputLookupJoinRule
-  
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSortRule
-  
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamTableFunctionScanRule
-  
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamUncollectRule
-  
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamUnionRule
-  
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamValuesRule
-  
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamWindowRule
-  
- convert() - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable.RowToDocument
-  
- convert(ResolvedNodes.ResolvedQueryStmt, List<RelNode>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter
-  
- convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.unnest.BeamZetaSqlUncollectRule
-  
- Convert - Class in org.apache.beam.sdk.schemas.transforms
- 
A set of utilities for converting between different objects supporting schemas. 
- Convert() - Constructor for class org.apache.beam.sdk.schemas.transforms.Convert
-  
- convert(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
-  
- convertArray(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
-  
- convertArray(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
-  
- convertArray(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
-  
- convertArray(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
-  
- convertAvroFieldStrict(Object, Schema, Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Strict conversion from AVRO to Beam, strict because it doesn't do widening or narrowing during
 conversion. 
- convertAvroFormat(Schema.FieldType, Object, BigQueryUtils.ConversionOptions) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
Tries to convert an Avro decoded value to a Beam field value based on the target type of the
 Beam field. 
- convertByteBuffer(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
-  
- convertByteBuffer(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
-  
- convertByteBuffer(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
-  
- convertByteBuffer(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
-  
- convertCharSequence(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
-  
- convertCharSequence(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
-  
- convertCharSequence(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
-  
- convertCharSequence(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
-  
- convertCollection(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
-  
- convertCollection(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
-  
- convertCollection(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
-  
- convertCollection(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
-  
- convertDateTime(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
-  
- convertDateTime(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
-  
- convertDateTime(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
-  
- convertDateTime(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
-  
- convertDefault(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.AvroConvertType
-  
- convertDefault(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.AvroConvertValueForGetter
-  
- convertDefault(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.AvroConvertValueForSetter
-  
- convertDefault(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
-  
- convertDefault(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
-  
- convertDefault(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
-  
- convertDefault(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
-  
- ConvertedSchemaInformation(SchemaCoder<T>, Schema.FieldType) - Constructor for class org.apache.beam.sdk.schemas.utils.ConvertHelpers.ConvertedSchemaInformation
-  
- convertEnum(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
-  
- convertEnum(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
-  
- convertEnum(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
-  
- convertEnum(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
-  
- convertGenericRecordToTableRow(GenericRecord, TableSchema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
- convertGenericRecordToTableRow(GenericRecord) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
Convert generic record to Bq TableRow. 
- ConvertHelpers - Class in org.apache.beam.sdk.schemas.utils
- 
Helper functions for converting between equivalent schema types. 
- ConvertHelpers() - Constructor for class org.apache.beam.sdk.schemas.utils.ConvertHelpers
-  
- ConvertHelpers.ConvertedSchemaInformation<T> - Class in org.apache.beam.sdk.schemas.utils
- 
Return value after converting a schema. 
- convertIterable(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
-  
- convertIterable(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
-  
- convertIterable(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
-  
- convertIterable(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
-  
- convertList(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
-  
- convertList(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
-  
- convertList(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
-  
- convertList(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
-  
- convertMap(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
-  
- convertMap(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
-  
- convertMap(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
-  
- convertMap(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
-  
- convertNewPartitionRowKeyToPartition(ByteString) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
- 
Convert new partition row key to partition to process metadata read from Bigtable. 
- convertNode2Map(JsonNode) - Static method in class org.apache.beam.sdk.extensions.sql.TableUtils
-  
- convertNumbers(TableRow) - Static method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
-  
- convertPartitionToNewPartitionRowKey(Range.ByteStringRange) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
- 
Convert partition to a New Partition row key to query for partitions ready to be streamed as
 the result of splits and merges. 
- convertPartitionToStreamPartitionRowKey(Range.ByteStringRange) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
- 
Convert partition to a Stream Partition row key to query for metadata of partitions that are
 currently being streamed. 
- convertPrimitive(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
-  
- convertPrimitive(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
-  
- convertPrimitive(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
-  
- convertPrimitive(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
-  
- convertRelNodeToRexRangeRef(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
-  
- convertRelOptCost(RelOptCost) - Static method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
-  
- convertResolvedLiteral(ResolvedNodes.ResolvedLiteral) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
- 
Convert a resolved literal to a RexNode. 
- convertRexNodeFromResolvedExpr(ResolvedNodes.ResolvedExpr, List<ResolvedColumn>, List<RelDataTypeField>, Map<String, RexNode>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
- 
Create a RexNode for a corresponding resolved expression node. 
- convertRexNodeFromResolvedExpr(ResolvedNodes.ResolvedExpr) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
- 
Create a RexNode for a corresponding resolved expression. 
- convertRootQuery(ConversionContext, ResolvedNodes.ResolvedQueryStmt) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter
-  
- convertStreamPartitionRowKeyToPartition(ByteString) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
- 
Convert stream partition row key to partition to process metadata read from Bigtable. 
- convertTableValuedFunction(RelNode, TableValuedFunction, List<ResolvedNodes.ResolvedFunctionArgument>, List<ResolvedColumn>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
- 
Convert a TableValuedFunction in ZetaSQL to a RexCall in Calcite. 
- convertToBagSpecInternal(StateSpec<CombiningState<InputT, AccumT, OutputT>>) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
For internal use only; no backwards-compatibility guarantees. 
- convertToBeamRel(String, QueryPlanner.QueryParameters) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner
- 
It parses and validate the input query, then convert into a  BeamRelNode tree. 
- convertToBeamRel(String, QueryPlanner.QueryParameters) - Method in interface org.apache.beam.sdk.extensions.sql.impl.QueryPlanner
- 
It parses and validate the input query, then convert into a  BeamRelNode tree. 
- convertToBeamRel(String) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
-  
- convertToBeamRel(String, Map<String, Value>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
-  
- convertToBeamRel(String, List<Value>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
-  
- convertToBeamRel(String, QueryPlanner.QueryParameters) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
-  
- convertToFileResourceIfPossible(String) - Static method in class org.apache.beam.sdk.io.FileBasedSink
- 
This is a helper function for turning a user-provided output filename prefix and converting it
 into a  ResourceId for writing output files. 
- convertToJcsmpDestination(Solace.Destination) - Static method in class org.apache.beam.sdk.io.solace.SolaceIO
- 
- convertToMapSpecInternal(StateSpec<SetState<KeyT>>) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
For internal use only; no backwards-compatibility guarantees. 
- convertToMultimapSpecInternal(StateSpec<MapState<KeyT, ValueT>>) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
For internal use only; no backwards-compatibility guarantees. 
- ConvertType(boolean) - Constructor for class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
-  
- ConvertValueForGetter(StackManipulation) - Constructor for class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
-  
- ConvertValueForSetter(StackManipulation) - Constructor for class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
-  
- copy() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
- 
Returns a copy of this RandomAccessData. 
- copy() - Method in class org.apache.beam.runners.spark.metrics.MetricsContainerStepMapAccumulator
-  
- copy() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsAccumulator
-  
- copy(Iterable<String>, Iterable<String>) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
-  
- copy(RelTraitSet, RelNode, ImmutableBitSet, List<ImmutableBitSet>, List<AggregateCall>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamAggregationRel
-  
- copy(RelTraitSet, RelNode, RexProgram) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel
-  
- copy(RelTraitSet, RexNode, RelNode, RelNode, JoinRelType, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCoGBKJoinRel
-  
- copy(RelTraitSet, List<RelNode>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter
-  
- copy(RelTraitSet, List<RelNode>, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIntersectRel
-  
- copy(RelTraitSet, List<RelNode>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
-  
- copy(RelNode, RelDataType, RexNode, boolean, boolean, Map<String, RexNode>, Map<String, RexNode>, RexNode, Map<String, ? extends SortedSet<String>>, boolean, ImmutableBitSet, RelCollation, RexNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamMatchRel
-  
- copy(RelTraitSet, List<RelNode>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamMatchRel
-  
- copy(RelTraitSet, List<RelNode>, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamMinusRel
-  
- copy(RelTraitSet, RexNode, RelNode, RelNode, JoinRelType, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSideInputJoinRel
-  
- copy(RelTraitSet, RexNode, RelNode, RelNode, JoinRelType, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSideInputLookupJoinRel
-  
- copy(RelTraitSet, RelNode, RelCollation, RexNode, RexNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSortRel
-  
- copy(RelTraitSet, List<RelNode>, RexNode, Type, RelDataType, Set<RelColumnMapping>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamTableFunctionScanRel
-  
- copy(RelTraitSet, RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUncollectRel
-  
- copy(RelTraitSet, List<RelNode>, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnionRel
-  
- copy(RelTraitSet, RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnnestRel
-  
- copy(RelTraitSet, List<RelNode>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamWindowRel
-  
- copy(RelTraitSet, RelNode, List<RexLiteral>, RelDataType, List<Window.Group>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamWindowRel
-  
- copy(RelTraitSet, RelNode, RexProgram) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcRel
-  
- copy(RelTraitSet, RelNode) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.unnest.BeamZetaSqlUncollectRel
-  
- copy(RelTraitSet, RelNode) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.unnest.BeamZetaSqlUnnestRel
-  
- copy(RelTraitSet, List<RelNode>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.unnest.ZetaSqlUnnest
-  
- copy(RelTraitSet, RelNode) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.unnest.ZetaSqlUnnest
-  
- copy(List<ClassLoaderFileSystem.ClassLoaderResourceId>, List<ClassLoaderFileSystem.ClassLoaderResourceId>) - Method in class org.apache.beam.sdk.io.ClassLoaderFileSystem
-  
- copy(List<ResourceIdT>, List<ResourceIdT>) - Method in class org.apache.beam.sdk.io.FileSystem
- 
Copies a Listof file-like resources from one location to another.
 
- copy(List<ResourceId>, List<ResourceId>, MoveOptions...) - Static method in class org.apache.beam.sdk.io.FileSystems
- 
Copies a Listof file-like resources from one location to another.
 
- copyFrom(ByteBuffer) - Static method in class org.apache.beam.sdk.io.range.ByteKey
- 
Creates a new  ByteKey backed by a copy of the data remaining in the specified  ByteBuffer. 
- copyFrom(byte[]) - Static method in class org.apache.beam.sdk.io.range.ByteKey
- 
Creates a new  ByteKey backed by a copy of the specified  byte[]. 
- copyFrom(FieldSpecifierNotationParser.DotExpressionComponentContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionComponentContext
-  
- copyFrom(FieldSpecifierNotationParser.QualifierListContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifierListContext
-  
- copyResourcesFromJar(JarFile) - Method in class org.apache.beam.runners.jobsubmission.PortablePipelineJarCreator
- 
- copyToList(ArrayData, DataType) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.Utils
-  
- coreName() - Method in class org.apache.beam.sdk.io.solr.SolrIO.ReplicaInfo
-  
- coreUrl() - Method in class org.apache.beam.sdk.io.solr.SolrIO.ReplicaInfo
-  
- CorrelationKey() - Constructor for class org.apache.beam.sdk.io.solace.data.Solace.CorrelationKey
-  
- cosh(Double) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinTrigonometricFunctions
- 
COSH(X) 
- CosmosClientBuilderFactory() - Constructor for class org.apache.beam.sdk.io.azure.cosmos.CosmosOptions.CosmosClientBuilderFactory
-  
- CosmosIO - Class in org.apache.beam.sdk.io.azure.cosmos
-  
- CosmosIO.BoundedCosmosBDSource<T> - Class in org.apache.beam.sdk.io.azure.cosmos
- 
- CosmosIO.Read<T> - Class in org.apache.beam.sdk.io.azure.cosmos
-  
- CosmosOptions - Interface in org.apache.beam.sdk.io.azure.cosmos
-  
- CosmosOptions.CosmosClientBuilderFactory - Class in org.apache.beam.sdk.io.azure.cosmos
- 
Create a cosmos client from the pipeline options. 
- Count - Class in org.apache.beam.sdk.transforms
- 
- countAsserts(Pipeline) - Static method in class org.apache.beam.sdk.testing.PAssert
-  
- COUNTER - Static variable in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTable
-  
- Counter - Interface in org.apache.beam.sdk.metrics
- 
A metric that reports a single long value and can be incremented or decremented. 
- counter(String, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
- 
Create a metric that can be incremented and decremented, and is aggregated by taking the sum. 
- counter(Class<?>, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
- 
Create a metric that can be incremented and decremented, and is aggregated by taking the sum. 
- CounterImpl - Class in org.apache.beam.runners.jet.metrics
- 
- CounterMark(long, Instant) - Constructor for class org.apache.beam.sdk.io.CountingSource.CounterMark
- 
Creates a checkpoint mark reflecting the last emitted value. 
- CounterMarkCoder() - Constructor for class org.apache.beam.sdk.io.CountingSource.CounterMarkCoder
-  
- CountErrors(Counter) - Constructor for class org.apache.beam.sdk.transforms.errorhandling.ErrorHandler.PTransformErrorHandler.WriteErrorMetrics.CountErrors
-  
- CountIf - Class in org.apache.beam.sdk.extensions.sql.impl.transform.agg
- 
Returns the count of TRUE values for expression. 
- COUNTIF - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- CountIf.CountIfFn - Class in org.apache.beam.sdk.extensions.sql.impl.transform.agg
-  
- CountIfFn() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CountIf.CountIfFn
-  
- CountingReadableByteChannel - Class in org.apache.beam.sdk.extensions.gcp.util.channels
-  
- CountingReadableByteChannel(ReadableByteChannel, Consumer<Integer>) - Constructor for class org.apache.beam.sdk.extensions.gcp.util.channels.CountingReadableByteChannel
-  
- CountingSeekableByteChannel - Class in org.apache.beam.sdk.extensions.gcp.util.channels
-  
- CountingSeekableByteChannel(SeekableByteChannel, Consumer<Integer>, Consumer<Integer>) - Constructor for class org.apache.beam.sdk.extensions.gcp.util.channels.CountingSeekableByteChannel
-  
- CountingSource - Class in org.apache.beam.sdk.io
- 
- CountingSource.CounterMark - Class in org.apache.beam.sdk.io
- 
The checkpoint for an unbounded  CountingSource is simply the last value produced. 
- CountingSource.CounterMarkCoder - Class in org.apache.beam.sdk.io
- 
A custom coder for CounterMark.
 
- CountingWritableByteChannel - Class in org.apache.beam.sdk.extensions.gcp.util.channels
-  
- CountingWritableByteChannel(WritableByteChannel, Consumer<Integer>) - Constructor for class org.apache.beam.sdk.extensions.gcp.util.channels.CountingWritableByteChannel
-  
- countPartitionsCreatedAfter(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao
- 
- CountWords() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.CountWords
-  
- CovarianceFn<T extends java.lang.Number> - Class in org.apache.beam.sdk.extensions.sql.impl.transform.agg
- 
Combine.CombineFnfor Covariance onNumbertypes.
 
- coverSameKeySpace(List<Range.ByteStringRange>, Range.ByteStringRange) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ByteStringRangeHelper
- 
Returns true if parentPartitions form a proper superset of childPartition. 
- CrashingRunner - Class in org.apache.beam.sdk.testing
- 
- CrashingRunner() - Constructor for class org.apache.beam.sdk.testing.CrashingRunner
-  
- create() - Static method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Creates a ConnectorConfiguration. 
- create(DataflowPipelineOptions) - Static method in class org.apache.beam.runners.dataflow.DataflowClient
-  
- create(PCollectionView<?>, Coder<T>) - Static method in class org.apache.beam.runners.dataflow.DataflowRunner.StreamingPCollectionViewWriterFn
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions.DataflowClientFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions.StagerFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions.UnboundedReaderMaxReadTimeFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowPipelineOptions.StagingLocationFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions.EnableWindmillServiceDirectPathFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions.GlobalConfigRefreshPeriodFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions.HarnessUpdateReportingPeriodFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions.LocalWindmillHostportFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions.MaxStackTraceDepthToReportFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions.PeriodicStatusPageDirectoryFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions.WindmillServiceStreamingRpcBatchLimitFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DefaultGcpRegionFactory
-  
- create() - Static method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.PayloadTranslator
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.direct.DirectOptions.AvailableParallelismFactory
-  
- create(FlinkJobServerDriver.FlinkServerConfiguration) - Static method in class org.apache.beam.runners.flink.FlinkJobInvoker
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.flink.FlinkPipelineOptions.MaxBundleSizeFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.flink.FlinkPipelineOptions.MaxBundleTimeFactory
-  
- create(boolean) - Static method in class org.apache.beam.runners.flink.TestFlinkRunner
-  
- create(String, ByteString, OutputStream) - Static method in class org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService.ArtifactDestination
-  
- create(JobInfo) - Static method in class org.apache.beam.runners.fnexecution.control.DefaultExecutableStageContext
-  
- create(JobInfo) - Static method in class org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory
-  
- create(JobInfo, Map<String, EnvironmentFactory.Provider>) - Static method in class org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory
-  
- create() - Static method in class org.apache.beam.runners.fnexecution.control.MapControlClientPool
- 
- create(String) - Method in interface org.apache.beam.runners.fnexecution.control.OutputReceiverFactory
- 
- create(ReferenceCountingExecutableStageContextFactory.Creator, SerializableFunction<Object, Boolean>) - Static method in class org.apache.beam.runners.fnexecution.control.ReferenceCountingExecutableStageContextFactory
-  
- create(EnvironmentFactory, GrpcFnServer<GrpcDataService>, GrpcFnServer<GrpcStateService>, IdGenerator) - Static method in class org.apache.beam.runners.fnexecution.control.SingleEnvironmentInstanceJobBundleFactory
- 
Deprecated. 
- create(String, String) - Method in class org.apache.beam.runners.fnexecution.control.TimerReceiverFactory
-  
- create(PipelineOptions, ExecutorService, OutboundObserverFactory) - Static method in class org.apache.beam.runners.fnexecution.data.GrpcDataService
-  
- create(PipelineOptions, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<FnApiControlClientPoolService>, ControlClientPool.Source) - Static method in class org.apache.beam.runners.fnexecution.environment.EmbeddedEnvironmentFactory
-  
- create(GrpcFnServer<FnApiControlClientPoolService>, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<ArtifactRetrievalService>, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool.Source, IdGenerator) - Static method in class org.apache.beam.runners.fnexecution.environment.ExternalEnvironmentFactory
-  
- create(ProcessManager, RunnerApi.Environment, String, InstructionRequestHandler) - Static method in class org.apache.beam.runners.fnexecution.environment.ProcessEnvironment
-  
- create(ProcessManager, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool.Source, IdGenerator, PipelineOptions) - Static method in class org.apache.beam.runners.fnexecution.environment.ProcessEnvironmentFactory
-  
- create() - Static method in class org.apache.beam.runners.fnexecution.environment.ProcessManager
-  
- create(String, String, String, Struct) - Static method in class org.apache.beam.runners.fnexecution.provisioning.JobInfo
-  
- create(ProvisionApi.ProvisionInfo, HeaderAccessor) - Static method in class org.apache.beam.runners.fnexecution.provisioning.StaticGrpcProvisionService
-  
- create() - Static method in class org.apache.beam.runners.fnexecution.state.GrpcStateService
- 
- create(Endpoints.ApiServiceDescriptor, HeaderAccessor) - Static method in class org.apache.beam.runners.fnexecution.status.BeamWorkerStatusGrpcService
- 
- create(GrpcFnServer<ArtifactStagingService>, Function<String, String>, ThrowingConsumer<Exception, String>, JobInvoker) - Static method in class org.apache.beam.runners.jobsubmission.InMemoryJobService
- 
Creates an InMemoryJobService. 
- create(GrpcFnServer<ArtifactStagingService>, Function<String, String>, ThrowingConsumer<Exception, String>, JobInvoker, int) - Static method in class org.apache.beam.runners.jobsubmission.InMemoryJobService
- 
Creates an InMemoryJobService. 
- create() - Method in interface org.apache.beam.runners.jobsubmission.JobServerDriver.JobInvokerFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.portability.testing.TestPortablePipelineOptions.DefaultJobServerConfigFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.spark.SparkCommonPipelineOptions.StorageLevelFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.spark.SparkCommonPipelineOptions.TmpCheckpointDirFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.runners.spark.SparkContextOptions.EmptyListenersList
-  
- create(SparkJobServerDriver.SparkServerConfiguration) - Static method in class org.apache.beam.runners.spark.SparkJobInvoker
-  
- create() - Static method in class org.apache.beam.runners.spark.SparkRunner
- 
Creates and returns a new SparkRunner with default options. 
- create(SparkPipelineOptions) - Static method in class org.apache.beam.runners.spark.SparkRunner
- 
Creates and returns a new SparkRunner with specified options. 
- create() - Static method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingRunner
- 
Creates and returns a new SparkStructuredStreamingRunner with default options. 
- create(SparkStructuredStreamingPipelineOptions) - Static method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingRunner
- 
Creates and returns a new SparkStructuredStreamingRunner with specified options. 
- create(Map<String, Broadcast<SideInputValues<?>>>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.SparkSideInputReader
- 
Creates a  SideInputReader for Spark from a map of PCollectionView  tag ids and the corresponding broadcasted  SideInputValues. 
- create(PipelineOptions) - Method in class org.apache.beam.runners.spark.TestSparkPipelineOptions.DefaultStopPipelineWatermarkFactory
-  
- create(byte[], Coder<T>) - Static method in class org.apache.beam.runners.spark.util.SideInputBroadcast
-  
- create(ExpansionService, String, int) - Static method in class org.apache.beam.sdk.expansion.service.ExpansionServer
- 
Create a  ExpansionServer for the provided ExpansionService running on an arbitrary
 port. 
- create(List<String>, Map<String, List<Dependency>>) - Static method in class org.apache.beam.sdk.expansion.service.ExpansionServiceConfig
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.expansion.service.ExpansionServiceOptions.ExpansionServiceConfigFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.expansion.service.ExpansionServiceOptions.JavaClassLookupAllowListFactory
-  
- create(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.TypeWithNullability
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.DefaultProjectFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.EnableStreamingEngineFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpOAuthScopesFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpTempLocationFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpUserCredentialsFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcsOptions.ExecutorServiceFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcsOptions.PathValidatorFactory
-  
- create(GcsPath, String) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
- 
- create(GcsPath, String, Integer) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
- 
- create(GcsPath, GcsUtil.CreateOptions) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
- 
Creates an object in GCS and prepares for uploading its contents. 
- create(String, String) - Static method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.GcsCountersOptions
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.GcsReadOptionsFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.GcsUtilFactory
- 
- create(PipelineOptions, Storage, HttpRequestInitializer, ExecutorService, Credentials, Integer, GcsUtil.GcsCountersOptions, GoogleCloudStorageReadOptions) - Static method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.GcsUtilFactory
- 
Returns an instance of  GcsUtil based on the given parameters. 
- create(StorageObject) - Static method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.StorageObjectOrIOException
-  
- create(IOException) - Static method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.StorageObjectOrIOException
-  
- create(OrderedProcessingHandler<EventTypeT, EventKeyTypeT, StateTypeT, ResultTypeT>) - Static method in class org.apache.beam.sdk.extensions.ordered.OrderedEventProcessor
- 
Create the transform. 
- create(Long, long, Long, Long, long, long, long, boolean, ContiguousSequenceRange) - Static method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus
-  
- create(EventT, UnprocessedEvent.Reason) - Static method in class org.apache.beam.sdk.extensions.ordered.UnprocessedEvent
- 
Create new unprocessed event. 
- create(EventT, Exception) - Static method in class org.apache.beam.sdk.extensions.ordered.UnprocessedEvent
- 
Create new unprocessed event which failed due to an exception thrown. 
- create(Coder<InputT>) - Static method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.ApproximateDistinctFn
- 
- create(Coder<InputT>) - Static method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.CountMinSketchFn
- 
- create(double) - Static method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.TDigestQuantilesFn
- 
- create(BufferedExternalSorter.Options) - Static method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter
-  
- create(ExternalSorter.Options) - Static method in class org.apache.beam.sdk.extensions.sorter.ExternalSorter
- 
- create(BufferedExternalSorter.Options) - Static method in class org.apache.beam.sdk.extensions.sorter.SortValues
- 
Returns a  SortValues<PrimaryKeyT, SecondaryKeyT, ValueT>PTransform. 
- create(double, double, double) - Static method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
-  
- create(double) - Static method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
- 
It creates an instance with rate=0 and window=rowCount for bounded sources. 
- create(Method) - Static method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
- 
Creates Functionfrom
 given method.
 
- create(Method, String) - Static method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
- 
Creates org.apache.beam.vendor.calcite.v1_20_0.org.apache.calcite.schema.Functionfrom
 given method.
 
- create(List<String>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.TableName
- 
Full table name with path. 
- create(List<String>, String) - Static method in class org.apache.beam.sdk.extensions.sql.impl.TableName
- 
Table name plus the path up to but not including table name. 
- create(String, Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamBuiltinAggregations
-  
- create(String, Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamBuiltinAnalyticFunctions
-  
- create(DataCatalogPipelineOptions) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
-  
- create(Method, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.UserFunctionDefinitions.JavaScalarFunction
-  
- create(Class<?>, String, String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ZetaSqlScalarFunctionImpl
- 
Creates Functionfrom
 given class.
 
- create(Method, String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ZetaSqlScalarFunctionImpl
- 
Creates Functionfrom
 given method.
 
- create(RelTraitSet, RelNode, boolean) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.unnest.ZetaSqlUnnest
- 
Creates an Uncollect. 
- create(String) - Static method in class org.apache.beam.sdk.fn.channel.AddHarnessIdInterceptor
-  
- create(String, Coder<T>, FnDataReceiver<T>) - Static method in class org.apache.beam.sdk.fn.data.DataEndpoint
-  
- create(Coder<T>, FnDataReceiver<T>) - Static method in class org.apache.beam.sdk.fn.data.DecodingFnDataReceiver
-  
- create(String, String, Coder<T>, FnDataReceiver<T>) - Static method in class org.apache.beam.sdk.fn.data.TimerEndpoint
-  
- create(List<? extends FnService>, Endpoints.ApiServiceDescriptor, ServerFactory) - Static method in class org.apache.beam.sdk.fn.server.GrpcFnServer
- 
- create(ServiceT, Endpoints.ApiServiceDescriptor, ServerFactory) - Static method in class org.apache.beam.sdk.fn.server.GrpcFnServer
- 
Create a  GrpcFnServer for the provided  FnService which will run at the endpoint
 specified in the  Endpoints.ApiServiceDescriptor. 
- create(ServiceT, Endpoints.ApiServiceDescriptor) - Static method in class org.apache.beam.sdk.fn.server.GrpcFnServer
- 
- create() - Static method in class org.apache.beam.sdk.fn.server.InProcessServerFactory
-  
- create(List<BindableService>, Endpoints.ApiServiceDescriptor) - Method in class org.apache.beam.sdk.fn.server.InProcessServerFactory
-  
- create(List<BindableService>, Endpoints.ApiServiceDescriptor) - Method in class org.apache.beam.sdk.fn.server.ServerFactory
- 
Creates an instance of this server at the address specified by the given service descriptor and
 bound to multiple services. 
- create(List<BindableService>, Endpoints.ApiServiceDescriptor) - Method in class org.apache.beam.sdk.fn.server.ServerFactory.InetSocketAddressServerFactory
-  
- create(StreamObserver<ReqT>, Runnable) - Static method in class org.apache.beam.sdk.fn.stream.ForwardingClientResponseObserver
-  
- create(StreamObserver<ReqT>, Runnable, Runnable) - Static method in class org.apache.beam.sdk.fn.stream.ForwardingClientResponseObserver
-  
- create(BuilderT, AwsOptions) - Method in interface org.apache.beam.sdk.io.aws2.common.ClientBuilderFactory
- 
Configure a client builder  BuilderT using the global defaults in  AwsOptions. 
- create(BuilderT, ClientConfiguration, AwsOptions) - Method in interface org.apache.beam.sdk.io.aws2.common.ClientBuilderFactory
- 
- create(BuilderT, ClientConfiguration, AwsOptions) - Method in class org.apache.beam.sdk.io.aws2.common.ClientBuilderFactory.DefaultClientBuilder
-  
- create(AwsCredentialsProvider, Region, URI) - Static method in class org.apache.beam.sdk.io.aws2.common.ClientConfiguration
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIOOptions.MapFactory
-  
- create() - Static method in class org.apache.beam.sdk.io.aws2.kinesis.WatermarkParameters
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.io.aws2.options.AwsOptions.AwsRegionFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.io.aws2.options.AwsOptions.AwsUserCredentialsFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.io.aws2.options.S3Options.S3UploadBufferSizeBytesFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.io.aws2.options.S3Options.SSECustomerKeyFactory
-  
- create(String, String, String, long, long) - Static method in class org.apache.beam.sdk.io.aws2.sqs.SqsMessage
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.io.azure.cosmos.CosmosOptions.CosmosClientBuilderFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.io.azure.options.AzureOptions.AzureUserCredentialsFactory
-  
- create(ClassLoaderFileSystem.ClassLoaderResourceId, CreateOptions) - Method in class org.apache.beam.sdk.io.ClassLoaderFileSystem
-  
- create(String[], String, String) - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
Creates a new Elasticsearch connection configuration. 
- create(String[], String) - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
Creates a new Elasticsearch connection configuration with no default type. 
- create(String[]) - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
Creates a new Elasticsearch connection configuration with no default index nor type. 
- create() - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Document
-  
- create(int, Duration) - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.RetryConfiguration
- 
Creates RetryConfiguration for  ElasticsearchIO with provided maxAttempts,
 maxDurations and exponential backoff based retries. 
- create(WritableByteChannel) - Method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
- 
Deprecated. 
- create(WritableByteChannel) - Method in interface org.apache.beam.sdk.io.FileBasedSink.WritableByteChannelFactory
-  
- create(EmptyMatchTreatment) - Static method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
- 
- create(ResourceIdT, CreateOptions) - Method in class org.apache.beam.sdk.io.FileSystem
- 
Returns a write channel for the given ResourceIdT.
 
- create(ResourceId, String) - Static method in class org.apache.beam.sdk.io.FileSystems
- 
- create(ResourceId, CreateOptions) - Static method in class org.apache.beam.sdk.io.FileSystems
- 
- create(MatchResult.Status, List<MatchResult.Metadata>) - Static method in class org.apache.beam.sdk.io.fs.MatchResult
- 
- create(MatchResult.Status, IOException) - Static method in class org.apache.beam.sdk.io.fs.MatchResult
- 
- create(ValueProvider<TableReference>, DataFormat, ValueProvider<List<String>>, ValueProvider<String>, SerializableFunction<SchemaAndRecord, T>, Coder<T>, BigQueryServices, boolean) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageTableSource
-  
- create(ValueProvider<TableReference>, ValueProvider<List<String>>, ValueProvider<String>, SerializableFunction<SchemaAndRecord, T>, Coder<T>, BigQueryServices) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageTableSource
-  
- create() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.StreamingInsertsMetrics.StreamingInsertsMetricsImpl
-  
- create(Schema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery
- 
Creates an instance of this rule. 
- create(long) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteResult
-  
- create(Schema, String) - Static method in class org.apache.beam.sdk.io.gcp.datastore.EntityToRow
- 
Create a PTransform instance. 
- create(String, String) - Static method in class org.apache.beam.sdk.io.gcp.datastore.RowToEntity
- 
Create a PTransform instance. 
- create() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
- 
- create() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal
- 
Creates an instance of this rule. 
- create(SubscriptionPartition) - Method in interface org.apache.beam.sdk.io.gcp.pubsublite.internal.ManagedFactory
-  
- create(SubscriptionPartition) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.ManagedFactoryImpl
-  
- create(SpannerConfig, String, String) - Static method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.MetadataSpannerConfigFactory
- 
Generates a SpannerConfig that can be used to access the change stream metadata database by
 copying only the necessary fields from the given primary database SpannerConfig and setting the
 instance ID and database ID to the supplied metadata values. 
- create(Mutation, Mutation...) - Static method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
- 
Creates a new group. 
- create(Mutation, Iterable<Mutation>) - Static method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
-  
- create() - Static method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- create() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- create(BatchTransactionId) - Static method in class org.apache.beam.sdk.io.gcp.spanner.Transaction
-  
- create(String, String, String, Boolean) - Static method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher.TableAndQuery
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsOptions.GoogleAdsCredentialsFactory
-  
- create(Map<String, String>) - Static method in class org.apache.beam.sdk.io.hcatalog.HCatalogBeamSchema
- 
Create the schema adapter. 
- create(PipelineOptions) - Method in class org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptions.ConfigurationLocator
-  
- create(ValueProvider<String>, ValueProvider<String>, ValueProvider<String>) - Static method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.DataSourceConfiguration
-  
- create(DataSource) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
-  
- create(String, String) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
-  
- create(ValueProvider<String>, ValueProvider<String>) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
-  
- create(int, Duration, Duration) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO.RetryConfiguration
-  
- create() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcWriteResult
-  
- create() - Static method in class org.apache.beam.sdk.io.jms.RetryConfiguration
-  
- create(int) - Static method in class org.apache.beam.sdk.io.jms.RetryConfiguration
-  
- create(int, Duration, Duration) - Static method in class org.apache.beam.sdk.io.jms.RetryConfiguration
-  
- create() - Static method in class org.apache.beam.sdk.io.kafka.KafkaMetrics.KafkaMetricsImpl
-  
- create() - Static method in class org.apache.beam.sdk.io.mongodb.AggregationQuery
-  
- create() - Static method in class org.apache.beam.sdk.io.mongodb.FindQuery
-  
- create() - Static method in class org.apache.beam.sdk.io.mongodb.UpdateConfiguration
-  
- create(String, String) - Static method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
- 
Describe a connection configuration to the MQTT broker. 
- create(String) - Static method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
-  
- create() - Static method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverConfiguration
-  
- create(String, String, String) - Static method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverConfiguration
-  
- create() - Static method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
-  
- create(String, int) - Static method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
-  
- create(ValueProvider<String>, ValueProvider<Integer>) - Static method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
-  
- create(String) - Static method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.DataSourceConfiguration
-  
- create() - Static method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- create(DataSource) - Static method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
- create() - Method in class org.apache.beam.sdk.io.solace.broker.BasicAuthJcsmpSessionServiceFactory
-  
- create() - Method in class org.apache.beam.sdk.io.solace.broker.BasicAuthSempClientFactory
-  
- create() - Method in class org.apache.beam.sdk.io.solace.broker.GCPSecretSessionServiceFactory
-  
- create(JCSMPProperties, Queue) - Static method in class org.apache.beam.sdk.io.solace.broker.JcsmpSessionService
-  
- create() - Method in interface org.apache.beam.sdk.io.solace.broker.SempClientFactory
- 
This method is the core of the factory interface. 
- create() - Method in class org.apache.beam.sdk.io.solace.broker.SessionServiceFactory
- 
This is the core method that subclasses must implement. 
- create() - Static method in class org.apache.beam.sdk.io.solace.RetryCallableManager
- 
- create(String) - Static method in class org.apache.beam.sdk.io.solr.SolrIO.ConnectionConfiguration
- 
Creates a new Solr connection configuration. 
- create(int, Duration) - Static method in class org.apache.beam.sdk.io.solr.SolrIO.RetryConfiguration
-  
- create() - Method in class org.apache.beam.sdk.io.splunk.SplunkEvent.Builder
- 
- create() - Method in class org.apache.beam.sdk.io.splunk.SplunkWriteError.Builder
- 
- create(Set<List<String>>) - Static method in class org.apache.beam.sdk.metrics.BoundedTrieResult
- 
- create(long, long, long, long) - Static method in class org.apache.beam.sdk.metrics.DistributionResult
-  
- create(long, Instant) - Static method in class org.apache.beam.sdk.metrics.GaugeResult
-  
- create(String, Map<String, String>) - Static method in class org.apache.beam.sdk.metrics.LabeledMetricNameUtils.ParsedMetricName
-  
- create(String) - Static method in class org.apache.beam.sdk.metrics.LabeledMetricNameUtils.ParsedMetricName
-  
- create(String, MetricName) - Static method in class org.apache.beam.sdk.metrics.MetricKey
-  
- create(Iterable<MetricResult<Long>>, Iterable<MetricResult<DistributionResult>>, Iterable<MetricResult<GaugeResult>>, Iterable<MetricResult<StringSetResult>>, Iterable<MetricResult<BoundedTrieResult>>) - Static method in class org.apache.beam.sdk.metrics.MetricQueryResults
-  
- create(MetricKey, Boolean, T) - Static method in class org.apache.beam.sdk.metrics.MetricResult
-  
- create(MetricKey, T, T) - Static method in class org.apache.beam.sdk.metrics.MetricResult
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.metrics.MetricsOptions.NoOpMetricsSink
-  
- create(Set<String>) - Static method in class org.apache.beam.sdk.metrics.StringSetResult
- 
- create(PipelineOptions) - Method in interface org.apache.beam.sdk.options.DefaultValueFactory
- 
- create(PipelineOptions) - Method in class org.apache.beam.sdk.options.ExecutorOptions.ScheduledExecutorServiceFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.options.PipelineOptions.AtomicLongFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.options.PipelineOptions.DirectRunner
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.options.PipelineOptions.JobNameFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.options.PipelineOptions.UserAgentFactory
-  
- create() - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
- 
Creates and returns an object that implements  PipelineOptions using the values
 configured on this builder during construction. 
- create() - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
- 
- create(PipelineOptions) - Method in class org.apache.beam.sdk.options.SdkHarnessOptions.BundleProcessorCacheTimeoutFactory
-  
- create(PipelineOptions) - Method in class org.apache.beam.sdk.options.SdkHarnessOptions.DefaultMaxCacheMemoryUsageMbFactory
-  
- create() - Static method in class org.apache.beam.sdk.Pipeline
- 
- create(PipelineOptions) - Static method in class org.apache.beam.sdk.Pipeline
- 
- create() - Static method in class org.apache.beam.sdk.PipelineRunner
- 
- create(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.schemas.CachingFactory
-  
- create(TypeDescriptor<?>, Schema) - Method in interface org.apache.beam.sdk.schemas.Factory
-  
- create() - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
- create(Map<String, Integer>) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
- 
Create an enumeration type over a set of String->Integer values. 
- create(List<String>) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
- 
Create an enumeration type from a fixed set of String values; integer values will be
 automatically chosen. 
- create(String...) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
- 
Create an enumeration type from a fixed set of String values; integer values will be
 automatically chosen. 
- create(Schema.Field...) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
- 
- create(List<Schema.Field>) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
- 
- create(List<Schema.Field>, Map<String, Integer>) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
- 
- create(Object...) - Method in interface org.apache.beam.sdk.schemas.SchemaUserTypeCreator
-  
- create() - Static method in class org.apache.beam.sdk.schemas.transforms.AddFields
-  
- create(List<String>, String) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast.CompatibilityError
-  
- create() - Static method in class org.apache.beam.sdk.schemas.transforms.Filter
-  
- create() - Static method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
- 
Returns a transform that does a global combine using an aggregation built up by calls to
 aggregateField and aggregateFields. 
- create() - Static method in class org.apache.beam.sdk.schemas.transforms.RenameFields
- 
Create an instance of this transform. 
- create() - Static method in class org.apache.beam.sdk.schemas.transforms.Select
-  
- create(Class<?>, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils.ClassWithSchema
-  
- create(TypeDescriptor<T>, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils.TypeDescriptorWithSchema
-  
- create(List<String>, Optional<Schema.TypeName>) - Static method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold.Context
-  
- create() - Static method in class org.apache.beam.sdk.testing.TestPipeline
- 
Creates and returns a new test pipeline. 
- create(PipelineOptions) - Method in class org.apache.beam.sdk.testing.TestPipelineOptions.AlwaysPassMatcherFactory
-  
- create(Coder<T>) - Static method in class org.apache.beam.sdk.testing.TestStream
- 
- create(Schema) - Static method in class org.apache.beam.sdk.testing.TestStream
-  
- create(Schema, TypeDescriptor<T>, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Static method in class org.apache.beam.sdk.testing.TestStream
-  
- create(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
- 
Returns an approximate quantiles combiner with the given compareFnand desired number
 of quantiles.
 
- create(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
- 
- create(int, ComparatorT, long, double) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
- 
Creates an approximate quantiles combiner with the given compareFnand desired number
 of quantiles.
 
- Create<T> - Class in org.apache.beam.sdk.transforms
- 
Create<T>takes a collection of elements of typeTknown when the pipeline is
 constructed and returns aPCollection<T>containing the elements.
 
- Create() - Constructor for class org.apache.beam.sdk.transforms.Create
-  
- create() - Static method in class org.apache.beam.sdk.transforms.Distinct
- 
Returns a Distinct<T>PTransform.
 
- create() - Static method in class org.apache.beam.sdk.transforms.DoFnSchemaInformation
- 
Create an instance. 
- create() - Static method in class org.apache.beam.sdk.transforms.GroupByKey
- 
Returns a GroupByKey<K, V>PTransform.
 
- create(long, long, SerializableFunction<InputT, Long>, Duration) - Static method in class org.apache.beam.sdk.transforms.GroupIntoBatches.BatchingParams
-  
- create() - Static method in class org.apache.beam.sdk.transforms.Impulse
- 
- create() - Static method in class org.apache.beam.sdk.transforms.join.CoGroupByKey
- 
Returns a CoGroupByKey<K>PTransform.
 
- create(JsonToRow.JsonToRowWithErrFn) - Static method in class org.apache.beam.sdk.transforms.JsonToRow.JsonToRowWithErrFn.ParseWithError
-  
- create() - Static method in class org.apache.beam.sdk.transforms.Keys
- 
Returns a Keys<K>PTransform.
 
- create() - Static method in class org.apache.beam.sdk.transforms.KvSwap
- 
Returns a KvSwap<K, V>PTransform.
 
- create() - Static method in class org.apache.beam.sdk.transforms.PeriodicImpulse
-  
- create() - Static method in class org.apache.beam.sdk.transforms.PeriodicSequence
-  
- create() - Static method in class org.apache.beam.sdk.transforms.resourcehints.ResourceHints
- 
- create(PipelineOptions) - Method in class org.apache.beam.sdk.transforms.resourcehints.ResourceHintsOptions.EmptyListDefault
-  
- create() - Static method in class org.apache.beam.sdk.transforms.Values
- 
Returns a Values<V>PTransform.
 
- create(T) - Static method in class org.apache.beam.sdk.values.PCollectionViews.ValueOrMetadata
-  
- create(Coder<T>, Coder<MetaT>) - Static method in class org.apache.beam.sdk.values.PCollectionViews.ValueOrMetadataCoder
-  
- Create.OfValueProvider<T> - Class in org.apache.beam.sdk.transforms
- 
- Create.TimestampedValues<T> - Class in org.apache.beam.sdk.transforms
- 
A PTransformthat creates aPCollectionwhose elements have associated
 timestamps.
 
- Create.Values<T> - Class in org.apache.beam.sdk.transforms
- 
A PTransformthat creates aPCollectionfrom a set of in-memory objects.
 
- Create.WindowedValues<T> - Class in org.apache.beam.sdk.transforms
- 
A PTransformthat creates aPCollectionwhose elements have associated
 windowing metadata.
 
- createAccumulator() - Method in class org.apache.beam.runners.flink.FlinkStreamingAggregationsTranslators.ConcatenateAsIterable
-  
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.ordered.combiner.DefaultSequenceCombiner
-  
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.ApproximateDistinctFn
-  
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.CountMinSketchFn
-  
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.TDigestQuantilesFn
-  
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.sql.impl.LazyAggregateCombineFn
-  
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CountIf.CountIfFn
-  
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CovarianceFn
-  
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.VarianceFn
-  
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamBuiltinAggregations.BitXOr
-  
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.sql.impl.udaf.ArrayAgg.ArrayAggArray
-  
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.sql.impl.udaf.StringAgg.StringAggByte
-  
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.sql.impl.udaf.StringAgg.StringAggString
-  
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.sql.provider.UdfTestProvider.Sum
-  
- createAccumulator() - Method in class org.apache.beam.sdk.extensions.sql.TypedCombineFnDelegate
-  
- createAccumulator() - Method in interface org.apache.beam.sdk.extensions.sql.udf.AggregateFn
- 
Returns a new, mutable accumulator value, representing the accumulation of zero input values. 
- createAccumulator() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Concatenate
-  
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
-  
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn
- 
Deprecated. 
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
-  
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
-  
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
-  
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
-  
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
- 
Returns a new, mutable accumulator value, representing the accumulation of zero input values. 
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
-  
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
-  
- createAccumulator(CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
-  
- createAccumulator(CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
- 
Returns a new, mutable accumulator value, representing the accumulation of zero input values. 
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
-  
- createAccumulator() - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
-  
- createAll(Class<?>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
- 
Creates Functionfor
 each method in a given class.
 
- createArrayOf(String, Object[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- createArtifactServerFactory(JobServerDriver.ServerConfiguration) - Static method in class org.apache.beam.runners.jobsubmission.JobServerDriver
-  
- createBacklogGauge(String, int) - Static method in class org.apache.beam.sdk.io.kafka.KafkaSinkMetrics
- 
Creates a  Gauge metric to record per partition backlog with the name 
- createBacklogGauge(MetricName) - Static method in class org.apache.beam.sdk.io.kafka.KafkaSinkMetrics
- 
Creates a  Gauge metric to record per partition backlog with the name 
- createBatch(Class<?>, Class<?>, Class<?>) - Static method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Creates a batch plugin instance. 
- createBatchExecutionEnvironment(FlinkPipelineOptions) - Static method in class org.apache.beam.runners.flink.FlinkExecutionEnvironments
- 
If the submitted job is a batch processing job, this method creates the adequate Flink ExecutionEnvironmentdepending on the user-specified options.
 
- createBigQueryClientCustomErrors() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl
-  
- createBitXOr(Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamBuiltinAggregations
-  
- createBlob() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- createBlockGenerator(BlockGeneratorListener) - Method in class org.apache.beam.sdk.io.sparkreceiver.WrappedSupervisor
-  
- createBoundedTableStatistics(Double) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
-  
- createBucket(String, Bucket) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
- 
Creates a  Bucket under the specified project in Cloud Storage or propagates an
 exception. 
- createBuilder(S3Options) - Method in interface org.apache.beam.sdk.io.aws2.options.S3ClientBuilderFactory
-  
- createBuilder(S3Options) - Method in class org.apache.beam.sdk.io.aws2.s3.DefaultS3ClientBuilderFactory
-  
- createBuilder(BlobstoreOptions) - Method in class org.apache.beam.sdk.io.azure.blobstore.DefaultBlobstoreClientBuilderFactory
-  
- createBuilder(BlobstoreOptions) - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreClientBuilderFactory
-  
- createBytesXMLMessage(Solace.Record, boolean, DeliveryMode) - Static method in class org.apache.beam.sdk.io.solace.broker.MessageProducerUtils
- 
- createCatalogItems() - Static method in class org.apache.beam.sdk.extensions.ml.RecommendationAIIO
-  
- createClassLoader(List<String>) - Method in class org.apache.beam.sdk.extensions.sql.impl.JavaUdfLoader
-  
- createClob() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- createCombineFn(AggregateCall, Schema.Field, String) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.AggregationCombineFnAdapter
- 
- createCombineFnAnalyticsFunctions(AggregateCall, Schema.Field, String) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.AggregationCombineFnAdapter
- 
- createConstantCombineFn() - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.AggregationCombineFnAdapter
-  
- createConstructorCreator(Class<T>, Constructor<T>, Schema, List<FieldValueTypeInformation>, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
-  
- createConstructorCreator(Class<? super T>, Constructor<T>, Schema, List<FieldValueTypeInformation>, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
-  
- createContextual(DeserializationContext, BeanProperty) - Method in class org.apache.beam.sdk.options.ValueProvider.Deserializer
-  
- createDataCatalogClient(DataCatalogPipelineOptions) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
-  
- CreateDataflowView<ElemT,ViewT> - Class in org.apache.beam.runners.dataflow
- 
- createDataset(List<WindowedValue<T>>, Encoder<WindowedValue<T>>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- createDataset(String, String, DatasetProperties) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- createDataset(String, String, String, String, Long) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
- 
Create a  Dataset with the given  location,  description and default
 expiration time for tables in the dataset (if  null, tables don't expire). 
- createDataset(String, String, String, String, Long) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.DatasetServiceImpl
- 
Create a  Dataset with the given  location,  description and default
 expiration time for tables in the dataset (if  null, tables don't expire). 
- createDataset(String, String, String, String, Long) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- createDatasetFromRDD(SparkSession, BoundedSource<T>, Supplier<PipelineOptions>, Encoder<WindowedValue<T>>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.io.BoundedDatasetFactory
- 
- createDatasetFromRows(SparkSession, BoundedSource<T>, Supplier<PipelineOptions>, Encoder<WindowedValue<T>>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.io.BoundedDatasetFactory
- 
- createDecompressingChannel(ReadableByteChannel) - Method in enum org.apache.beam.sdk.io.CompressedSource.CompressionMode
- 
Deprecated. 
- createDecompressingChannel(ReadableByteChannel) - Method in interface org.apache.beam.sdk.io.CompressedSource.DecompressingChannelFactory
- 
Given a channel, create a channel that decompresses the content read from the channel. 
- createDefault() - Static method in class org.apache.beam.sdk.coders.CoderRegistry
- 
Creates a CoderRegistry containing registrations for all standard coders part of the core Java
 Apache Beam SDK and also any registrations provided by  coder
 registrars. 
- createDefault() - Static method in class org.apache.beam.sdk.fn.channel.ManagedChannelFactory
- 
Creates a ManagedChannelrelying on theManagedChannelBuilderto choose the
 channel type.
 
- createDefault() - Static method in class org.apache.beam.sdk.fn.server.ServerFactory
- 
- createDefault() - Static method in interface org.apache.beam.sdk.fn.server.ServerFactory.UrlFactory
-  
- createDefault() - Static method in class org.apache.beam.sdk.schemas.SchemaRegistry
-  
- createDefault() - Static method in class org.apache.beam.sdk.transforms.GroupIntoBatches.BatchingParams
-  
- createDicomStore(String, String) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Create a DicomStore. 
- createDicomStore(String, String, String) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Create a DicomStore with a PubSub listener. 
- createDicomStore(String, String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- createDicomStore(String, String, String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- CreateDisposition - Enum in org.apache.beam.sdk.io.snowflake.enums
- 
Enum containing all supported dispositions for table. 
- createEnvironment(RunnerApi.Environment, String) - Method in class org.apache.beam.runners.fnexecution.environment.DockerEnvironmentFactory
- 
- createEnvironment(RunnerApi.Environment, String) - Method in class org.apache.beam.runners.fnexecution.environment.EmbeddedEnvironmentFactory
-  
- createEnvironment(RunnerApi.Environment, String) - Method in interface org.apache.beam.runners.fnexecution.environment.EnvironmentFactory
- 
Creates an active RunnerApi.Environmentand returns a handle to it.
 
- createEnvironment(RunnerApi.Environment, String) - Method in class org.apache.beam.runners.fnexecution.environment.ExternalEnvironmentFactory
- 
- createEnvironment(RunnerApi.Environment, String) - Method in class org.apache.beam.runners.fnexecution.environment.ProcessEnvironmentFactory
- 
- createEnvironment(RunnerApi.Environment, String) - Method in class org.apache.beam.runners.fnexecution.environment.StaticRemoteEnvironmentFactory
-  
- createEnvironmentFactory(GrpcFnServer<FnApiControlClientPoolService>, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<ArtifactRetrievalService>, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool, IdGenerator) - Method in class org.apache.beam.runners.fnexecution.environment.DockerEnvironmentFactory.Provider
-  
- createEnvironmentFactory(GrpcFnServer<FnApiControlClientPoolService>, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<ArtifactRetrievalService>, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool, IdGenerator) - Method in class org.apache.beam.runners.fnexecution.environment.EmbeddedEnvironmentFactory.Provider
-  
- createEnvironmentFactory(GrpcFnServer<FnApiControlClientPoolService>, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<ArtifactRetrievalService>, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool, IdGenerator) - Method in interface org.apache.beam.runners.fnexecution.environment.EnvironmentFactory.Provider
- 
- createEnvironmentFactory(GrpcFnServer<FnApiControlClientPoolService>, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<ArtifactRetrievalService>, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool, IdGenerator) - Method in class org.apache.beam.runners.fnexecution.environment.ExternalEnvironmentFactory.Provider
-  
- createEnvironmentFactory(GrpcFnServer<FnApiControlClientPoolService>, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<ArtifactRetrievalService>, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool, IdGenerator) - Method in class org.apache.beam.runners.fnexecution.environment.ProcessEnvironmentFactory.Provider
-  
- createEnvironmentFactory(GrpcFnServer<FnApiControlClientPoolService>, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<ArtifactRetrievalService>, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool, IdGenerator) - Method in class org.apache.beam.runners.fnexecution.environment.StaticRemoteEnvironmentFactory.Provider
-  
- createEpoll() - Static method in class org.apache.beam.sdk.fn.channel.ManagedChannelFactory
- 
Creates a  ManagedChannelFactory backed by an  EpollDomainSocketChannel if the
 address is a  DomainSocketAddress. 
- createEpollDomainSocket() - Static method in class org.apache.beam.sdk.fn.server.ServerFactory
- 
Create a EpollDomainSocket.
 
- createEpollSocket() - Static method in class org.apache.beam.sdk.fn.server.ServerFactory
- 
Create a EpollSocket.
 
- createFactory() - Method in class org.apache.beam.sdk.extensions.sql.impl.JdbcDriver
-  
- createFactoryForCreateSubscription() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- createFactoryForGetSchema(PubsubClient.TopicPath, PubsubClient.SchemaPath, Schema) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- createFactoryForPublish(PubsubClient.TopicPath, Iterable<PubsubClient.OutgoingMessage>, Iterable<PubsubClient.OutgoingMessage>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
- 
Return a factory for testing publishers. 
- createFactoryForPull(Clock, PubsubClient.SubscriptionPath, int, Iterable<PubsubClient.IncomingMessage>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
- 
Return a factory for testing subscribers. 
- createFactoryForPullAndPublish(PubsubClient.SubscriptionPath, PubsubClient.TopicPath, Clock, int, Iterable<PubsubClient.IncomingMessage>, Iterable<PubsubClient.OutgoingMessage>, Iterable<PubsubClient.OutgoingMessage>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
- 
Returns a factory for a test that is expected to both publish and pull messages over the course
 of the test. 
- createFhirStore(String, String, String, String) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Create FHIR Store with a PubSub topic listener. 
- createFhirStore(String, String, String) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Create FHIR Store. 
- createFhirStore(String, String, String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- createFhirStore(String, String, String, String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- createFile() - Method in class org.apache.beam.sdk.jmh.io.TextSourceBenchmark.Data
- 
Generates a random file with NUM_LINESbetween 60 and 120 characters each.
 
- createForSubrangeOfFile(String, long, long) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
- 
- createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
-  
- createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.BlockBasedSource
- 
Creates a BlockBasedSourcefor the specified range in a single file.
 
- createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.CompressedSource
- 
Creates a CompressedSourcefor a subrange of a file.
 
- createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.FileBasedSource
- 
Creates and returns a new FileBasedSourceof the same type as the currentFileBasedSourcebacked by a given file and an offset range.
 
- createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.TextSource
-  
- createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.xml.XmlSource
-  
- createFrom(String) - Static method in class org.apache.beam.sdk.fn.channel.SocketAddressFactory
- 
Parse a SocketAddressfrom the given string.
 
- createGetter(FieldValueTypeInformation, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
-  
- createGetterConversions(StackManipulation) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.DefaultTypeConversionsFactory
-  
- createGetterConversions(StackManipulation) - Method in interface org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversionsFactory
-  
- createHL7v2Message(String, Message) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Creates an HL7v2 message. 
- createHL7v2Message(String, Message) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- createHL7v2Store(String, String) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Create hl 7 v 2 store hl 7 v 2 store. 
- createHL7v2Store(String, String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- createImplementor(Method) - Static method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
-  
- createInProcess() - Static method in class org.apache.beam.sdk.fn.channel.ManagedChannelFactory
- 
Creates a ManagedChannelusing an in-process channel.
 
- createInput(Pipeline, Map<String, PCollection<?>>) - Method in class org.apache.beam.sdk.expansion.service.ExpansionServiceSchemaTransformProvider
-  
- createInput(Pipeline, Map<String, PCollection<?>>) - Method in interface org.apache.beam.sdk.expansion.service.TransformProvider
-  
- createInputFormatInstance() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
- 
Creates instance of InputFormat class. 
- createInternal(WindowingStrategy) - Static method in class org.apache.beam.sdk.transforms.windowing.Window.Assign
-  
- createIterator() - Method in class org.apache.beam.sdk.fn.stream.PrefetchableIterables.Default
-  
- createJCSMPSendMultipleEntry(List<Solace.Record>, boolean, SerializableFunction<Solace.Record, Destination>, DeliveryMode) - Static method in class org.apache.beam.sdk.io.solace.broker.MessageProducerUtils
- 
- createJob(Job) - Method in class org.apache.beam.runners.dataflow.DataflowClient
- 
Creates the Dataflow  Job. 
- createJobInvocation(String, String, ListeningExecutorService, RunnerApi.Pipeline, FlinkPipelineOptions, PortablePipelineRunner) - Method in class org.apache.beam.runners.flink.FlinkJobInvoker
-  
- createJobServerFactory(JobServerDriver.ServerConfiguration) - Static method in class org.apache.beam.runners.jobsubmission.JobServerDriver
-  
- createJobService() - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver
-  
- createKafkaRead() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
-  
- createMessagesArray(Iterable<Solace.Record>, boolean) - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- createMetadata(MetaT) - Static method in class org.apache.beam.sdk.values.PCollectionViews.ValueOrMetadata
-  
- createMetadataTable() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
- 
Create the metadata table if it does not exist yet. 
- createNClob() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- createNewDataset(String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
- 
Creates a new dataset. 
- createNewDataset(String, String, Long) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
- 
Creates a new dataset with defaultTableExpirationMs. 
- createNewDataset(String, String, Long, String) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
- 
Creates a new dataset with defaultTableExpirationMs and in a specified location (GCP region). 
- createNewTable(String, String, Table) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
-  
- CreateOptions() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.CreateOptions
-  
- CreateOptions - Class in org.apache.beam.sdk.io.fs
- 
An abstract class that contains common configuration options for creating resources. 
- CreateOptions() - Constructor for class org.apache.beam.sdk.io.fs.CreateOptions
-  
- CreateOptions.Builder<BuilderT extends CreateOptions.Builder<BuilderT>> - Class in org.apache.beam.sdk.io.fs
- 
- CreateOptions.StandardCreateOptions - Class in org.apache.beam.sdk.io.fs
- 
A standard configuration options with builder. 
- CreateOptions.StandardCreateOptions.Builder - Class in org.apache.beam.sdk.io.fs
- 
- createOrUpdateReadChangeStreamMetadataTable(String, String, String) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO
- 
Utility method to create or update Read Change Stream metadata table. 
- createOutboundAggregator(Supplier<String>, boolean) - Method in interface org.apache.beam.runners.fnexecution.data.FnDataService
- 
- createOutboundAggregator(Supplier<String>, boolean) - Method in class org.apache.beam.runners.fnexecution.data.GrpcDataService
-  
- createOutputMap(Iterable<String>) - Static method in class org.apache.beam.runners.fnexecution.translation.PipelineTranslatorUtils
- 
Creates a mapping from PCollection id to output tag integer. 
- createPane(boolean, boolean, PaneInfo.Timing) - Static method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
-  
- createPane(boolean, boolean, PaneInfo.Timing, long, long) - Static method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
- 
Factory method to create a  PaneInfo with the specified parameters. 
- createPartitionMetadataTable() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataAdminDao
- 
Creates the metadata table in the given instance, database configuration, with the constructor
 specified table name. 
- createPipeline(PipelineOptions) - Method in class org.apache.beam.sdk.expansion.service.ExpansionService
-  
- createPipelineOptions(Map<String, String>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter
-  
- createPlanner(JdbcConnection, Collection<RuleSet>) - Method in interface org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.Factory
-  
- createPrepareContext() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- createPrimitiveOutputInternal(Pipeline, WindowingStrategy<?, ?>, PCollection.IsBounded, Coder<T>) - Static method in class org.apache.beam.sdk.values.PCollection
- 
For internal use only; no backwards-compatibility guarantees. 
- createPrimitiveOutputInternal(Pipeline, WindowingStrategy<?, ?>, PCollection.IsBounded, Coder<T>, TupleTag<?>) - Static method in class org.apache.beam.sdk.values.PCollection
- 
For internal use only; no backwards-compatibility guarantees. 
- createProcessContext(ValueInSingleWindow<InputT>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- createProperties() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- createPushDownRel(RelDataType, List<String>, BeamSqlTableFilter) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
-  
- createQuery(Expression, Class<T>) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- createQuery(Expression, Type) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- createQuery(String, String, String) - Static method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher
-  
- createQueryUsingStandardSql(String, String, String) - Static method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher
-  
- createQueueForTopic(String, String) - Method in class org.apache.beam.sdk.io.solace.broker.BasicAuthSempClient
-  
- createQueueForTopic(String, String) - Method in interface org.apache.beam.sdk.io.solace.broker.SempClient
- 
This is only called when a user requests to read data from a topic. 
- createRandomSubscription(PubsubClient.ProjectPath, PubsubClient.TopicPath, int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
- 
Create a random subscription for topic.
 
- createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.azure.cosmos.CosmosIO.BoundedCosmosBDSource
-  
- createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.BoundedSource
- 
- createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
-  
- createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
-  
- createReader(PipelineOptions, CheckpointMarkImpl) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.UnboundedSourceImpl
-  
- createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
-  
- createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
-  
- createReader(PipelineOptions, SolaceCheckpointMark) - Method in class org.apache.beam.sdk.io.solace.read.UnboundedSolaceSource
-  
- createReader(PipelineOptions, CheckpointMarkT) - Method in class org.apache.beam.sdk.io.UnboundedSource
- 
- createReadSession(CreateReadSessionRequest) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StorageClient
- 
Create a new read session against an existing table. 
- createRPCLatencyHistogram(KafkaSinkMetrics.RpcMethod, String) - Static method in class org.apache.beam.sdk.io.kafka.KafkaSinkMetrics
- 
Creates a  Histogram metric to record RPC latency with the name 
- createSchema(PubsubClient.SchemaPath, String, Schema.Type) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
- 
Create  Schema from Schema definition content. 
- createSchema(PubsubClient.SchemaPath, String, Schema.Type) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
- 
Create Schemafrom Schema definition content.
 
- createSchema(PubsubClient.SchemaPath, String, Schema.Type) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
- 
Create  Schema from Schema definition content. 
- createSchema(PubsubClient.SchemaPath, String, Schema.Type) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- createSessionToken(String) - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver
-  
- createSetter(FieldValueTypeInformation, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
-  
- createSetterConversions(StackManipulation) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.DefaultTypeConversionsFactory
-  
- createSetterConversions(StackManipulation) - Method in interface org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversionsFactory
-  
- createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
-  
- createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.BlockBasedSource
- 
Creates a BlockBasedReader.
 
- createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.CompressedSource
- 
Creates a FileBasedReaderto read a single file.
 
- createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
- 
Creates and returns an instance of a FileBasedReaderimplementation for the current
 source assuming the source represents a single file.
 
- createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.TextSource
-  
- createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.xml.XmlSource
-  
- createSingleMessage(Solace.Record, boolean) - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- createSource - Variable in class org.apache.beam.sdk.io.ReadAllViaFileBasedSourceTransform
-  
- createSourceForSubrange(long, long) - Method in class org.apache.beam.sdk.io.FileBasedSource
-  
- createSourceForSubrange(long, long) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
- 
- createSQLXML() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- createStateBackend(FlinkPipelineOptions) - Method in interface org.apache.beam.runners.flink.FlinkStateBackendFactory
-  
- createStatement() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- createStatement(int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- createStatement(int, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- createStateOnInitialEvent(EventT) - Method in interface org.apache.beam.sdk.extensions.ordered.EventExaminer
- 
If the event was the first event for a given key, create the state to hold the required data
 needed for processing. 
- createStaticCreator(Class<T>, Method, Schema, List<FieldValueTypeInformation>, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
-  
- createStaticCreator(Class<T>, Method, Schema, List<FieldValueTypeInformation>, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
-  
- CreateStream<T> - Class in org.apache.beam.runners.spark.io
- 
Create an input stream from Queue. 
- createStreamExecutionEnvironment(FlinkPipelineOptions, List<String>, String) - Static method in class org.apache.beam.runners.flink.FlinkExecutionEnvironments
- 
If the submitted job is a stream processing job, this method creates the adequate Flink StreamExecutionEnvironmentdepending on the
 user-specified options.
 
- createStreaming(Class<?>, SerializableFunction<V, Long>, Class<? extends Receiver<V>>, SerializableFunction<PluginConfig, Object[]>) - Static method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Creates a streaming plugin instance. 
- createStreaming(Class<?>, SerializableFunction<V, Long>, Class<? extends Receiver<V>>) - Static method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Creates a streaming plugin instance with default function for getting args for Receiver.
 
- createStringAggOperator(ResolvedNodes.ResolvedFunctionCallBase) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- createStruct(String, Object[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- createStructuralValues(Coder<T>, List<T>) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
- 
Testing utilities below depend on standard assertions and matchers to compare elements read by
 sources. 
- createSubscription(PubsubClient.TopicPath, PubsubClient.SubscriptionPath, int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
- 
Create subscriptiontotopic.
 
- createSubscription(PubsubClient.TopicPath, PubsubClient.SubscriptionPath, int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
-  
- createSubscription(PubsubClient.TopicPath, PubsubClient.SubscriptionPath, int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
-  
- createSubscription(PubsubClient.TopicPath, PubsubClient.SubscriptionPath, int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- createTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
-  
- createTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.InMemoryMetaTableProvider
-  
- createTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.ReadOnlyTableProvider
-  
- createTable(Table) - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.TableProvider
- 
Creates a table. 
- createTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
-  
- createTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.store.InMemoryMetaStore
-  
- createTable(Table) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
- 
Creates the specified table if it does not exist. 
- createTable(Table) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.DatasetServiceImpl
- 
Creates the specified table if it does not exist. 
- createTable(Table) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- CreateTableHelpers - Class in org.apache.beam.sdk.io.gcp.bigquery
-  
- CreateTableHelpers() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers
-  
- CreateTables<DestinationT,ElementT> - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
Creates any tables needed before performing streaming writes to the tables. 
- CreateTables(BigQueryIO.Write.CreateDisposition, DynamicDestinations<?, DestinationT>) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.CreateTables
- 
The list of tables created so far, so we don't try the creation each time. 
- createTest(String, String, String) - Static method in class org.apache.beam.sdk.io.gcp.datastore.RowToEntity
-  
- createTimestampPolicy(TopicPartition, Optional<Instant>) - Method in interface org.apache.beam.sdk.io.kafka.TimestampPolicyFactory
- 
Creates a TimestampPolicy for a partition. 
- createTopic(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
- 
Create topic.
 
- createTopic(PubsubClient.TopicPath, PubsubClient.SchemaPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
- 
- createTopic(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
-  
- createTopic(PubsubClient.TopicPath, PubsubClient.SchemaPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
-  
- createTopic(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
-  
- createTopic(PubsubClient.TopicPath, PubsubClient.SchemaPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
-  
- createTopic(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- createTopic(PubsubClient.TopicPath, PubsubClient.SchemaPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- createTransaction() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO
- 
Returns a transform that creates a batch transaction. 
- CreateTransaction() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
-  
- createTranslationContext(JobInfo, FlinkPipelineOptions, String, List<String>) - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator
- 
Creates a batch translation context. 
- createTranslationContext(JobInfo, FlinkPipelineOptions, ExecutionEnvironment) - Static method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator
-  
- createTranslationContext(JobInfo, FlinkPipelineOptions, String, List<String>) - Method in interface org.apache.beam.runners.flink.FlinkPortablePipelineTranslator
-  
- createTranslationContext(JobInfo, FlinkPipelineOptions, String, List<String>) - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator
- 
Creates a streaming translation context. 
- createTranslationContext(JobInfo, FlinkPipelineOptions, StreamExecutionEnvironment) - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator
- 
Creates a streaming translation context. 
- createTranslator() - Static method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator
- 
Creates a batch translator. 
- createTranslator(Map<String, FlinkBatchPortablePipelineTranslator.PTransformTranslator>) - Static method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator
- 
Creates a batch translator. 
- createTypeConversion(boolean) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.DefaultTypeConversionsFactory
-  
- createTypeConversion(boolean) - Method in interface org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversionsFactory
-  
- createUnboundedTableStatistics(Double) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
-  
- createUrl(String, int) - Method in interface org.apache.beam.sdk.fn.server.ServerFactory.UrlFactory
-  
- createValue(String, T) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
- 
Create a  OneOfType.Value specifying which field to set and the value to set. 
- createValue(int, T) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
- 
Create a  OneOfType.Value specifying which field to set and the value to set. 
- createValue(EnumerationType.Value, T) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
- 
Create a  OneOfType.Value specifying which field to set and the value to set. 
- createWatermarkPolicy() - Method in interface org.apache.beam.sdk.io.aws2.kinesis.WatermarkPolicyFactory
-  
- createWithBytesReadConsumer(SeekableByteChannel, Consumer<Integer>) - Static method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingSeekableByteChannel
-  
- createWithBytesWrittenConsumer(SeekableByteChannel, Consumer<Integer>) - Static method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingSeekableByteChannel
-  
- createWithNoOpConsumer(ReadableByteChannel) - Static method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingReadableByteChannel
-  
- createWithNoOpConsumer(SeekableByteChannel) - Static method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingSeekableByteChannel
-  
- createWithNoOpConsumer(WritableByteChannel) - Static method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingWritableByteChannel
-  
- createWithPortSupplier(Supplier<Integer>) - Static method in class org.apache.beam.sdk.fn.server.ServerFactory
- 
- createWithUrlFactory(ServerFactory.UrlFactory) - Static method in class org.apache.beam.sdk.fn.server.ServerFactory
- 
- createWithUrlFactoryAndPortSupplier(ServerFactory.UrlFactory, Supplier<Integer>) - Static method in class org.apache.beam.sdk.fn.server.ServerFactory
- 
- createWriteOperation() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSink
-  
- createWriteOperation() - Method in class org.apache.beam.sdk.io.FileBasedSink
- 
- createWriter() - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
- 
- createWriteStream(String, WriteStream.Type) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.WriteStreamService
- 
Create a Write Stream for use with the Storage Write API. 
- createWriteStream(String, WriteStream.Type) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.WriteStreamServiceImpl
-  
- createWriteStream(String, WriteStream.Type) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- createZetaSqlFunction(String, SqlTypeName) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
- 
Create a dummy SqlFunction of type OTHER_FUNCTION from given function name and return type. 
- CredentialFactory - Interface in org.apache.beam.sdk.extensions.gcp.auth
- 
Construct an oauth credential to be used by the SDK and the SDK workers. 
- credentialsProvider(AwsCredentialsProvider) - Method in class org.apache.beam.sdk.io.aws2.common.ClientConfiguration.Builder
- 
- credentialsProvider() - Method in class org.apache.beam.sdk.io.aws2.common.ClientConfiguration
- 
- CrossLanguageConfiguration() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.CrossLanguageConfiguration
-  
- CrossLanguageConfiguration - Class in org.apache.beam.sdk.io.snowflake.crosslanguage
- 
Parameters abstract class to expose the transforms to an external SDK. 
- CrossLanguageConfiguration() - Constructor for class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- crossProductJoin() - Method in class org.apache.beam.sdk.schemas.transforms.CoGroup.Impl
- 
Expand the join into individual rows, similar to SQL joins. 
- CsvConfiguration() - Constructor for class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.CsvConfiguration
-  
- csvConfigurationBuilder() - Static method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration
-  
- CsvIO - Class in org.apache.beam.sdk.io.csv
- 
- CsvIO() - Constructor for class org.apache.beam.sdk.io.csv.CsvIO
-  
- CsvIO.Write<T> - Class in org.apache.beam.sdk.io.csv
- 
- CsvIOParse<T> - Class in org.apache.beam.sdk.io.csv
- 
- CsvIOParse() - Constructor for class org.apache.beam.sdk.io.csv.CsvIOParse
-  
- CsvIOParseError - Class in org.apache.beam.sdk.io.csv
- 
CsvIOParseError is a data class to store errors from CSV record processing. 
- CsvIOParseError() - Constructor for class org.apache.beam.sdk.io.csv.CsvIOParseError
-  
- CsvIOParseResult<T> - Class in org.apache.beam.sdk.io.csv
- 
- csvLines2BeamRows(CSVFormat, String, Schema) - Static method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamTableUtils
- 
Decode zero or more CSV records from the given string, according to the specified  CSVFormat, and converts them to  Rows with the specified  Schema. 
- CsvSink - Class in org.apache.beam.runners.spark.metrics.sink
- 
- CsvSink(Properties, MetricRegistry, SecurityManager) - Constructor for class org.apache.beam.runners.spark.metrics.sink.CsvSink
- 
Constructor for Spark 3.1.x and earlier. 
- CsvSink(Properties, MetricRegistry) - Constructor for class org.apache.beam.runners.spark.metrics.sink.CsvSink
- 
Constructor for Spark 3.2.x and later. 
- CsvToRow(Schema, CSVFormat) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTableProvider.CsvToRow
-  
- CsvWriteConfiguration() - Constructor for class org.apache.beam.sdk.io.csv.providers.CsvWriteTransformProvider.CsvWriteConfiguration
-  
- CsvWriteSchemaTransformFormatProvider - Class in org.apache.beam.sdk.io.fileschematransform
- 
- CsvWriteSchemaTransformFormatProvider() - Constructor for class org.apache.beam.sdk.io.fileschematransform.CsvWriteSchemaTransformFormatProvider
-  
- CsvWriteTransformProvider - Class in org.apache.beam.sdk.io.csv.providers
- 
- CsvWriteTransformProvider() - Constructor for class org.apache.beam.sdk.io.csv.providers.CsvWriteTransformProvider
-  
- CsvWriteTransformProvider.CsvWriteConfiguration - Class in org.apache.beam.sdk.io.csv.providers
- 
Configuration for writing to BigQuery with Storage Write API. 
- CsvWriteTransformProvider.CsvWriteConfiguration.Builder - Class in org.apache.beam.sdk.io.csv.providers
- 
- CsvWriteTransformProvider.CsvWriteTransform - Class in org.apache.beam.sdk.io.csv.providers
- 
- ctxt - Variable in class org.apache.beam.runners.spark.SparkRunner.Evaluator
-  
- CURRENT_METADATA_TABLE_VERSION - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
-  
- currentEventTime() - Method in interface org.apache.beam.sdk.state.Timers
- 
Returns the current event time. 
- currentInputWatermarkTime() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-  
- currentOutputWatermarkTime() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-  
- currentProcessingTime() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-  
- currentProcessingTime() - Method in interface org.apache.beam.sdk.state.Timers
- 
Returns the current processing time. 
- currentRestriction() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.ReadChangeStreamPartitionProgressTracker
- 
Returns the streamProgress that was successfully claimed. 
- currentRestriction() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRangeTracker
-  
- currentRestriction() - Method in class org.apache.beam.sdk.transforms.PeriodicSequence.OutputRangeTracker
-  
- currentRestriction() - Method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
-  
- currentRestriction() - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
-  
- currentRestriction() - Method in class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker
- 
Returns a restriction accurately describing the full range of work the current DoFn.ProcessElementcall will do, including already completed work.
 
- currentSynchronizedProcessingTime() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-  
- currentSynchronizedProcessingTime() - Method in interface org.apache.beam.sdk.state.Timers
- 
Returns the current synchronized processing time or nullif unknown.
 
- currentWatermark - Variable in class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.LogAppendTimePolicy
-  
- currentWatermark() - Method in interface org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimator
- 
Return estimated output watermark. 
- currentWatermark() - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.Manual
-  
- currentWatermark() - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.MonotonicallyIncreasing
-  
- currentWatermark() - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.WallTime
-  
- custom() - Static method in class org.apache.beam.sdk.io.thrift.ThriftSchema
- 
Builds a schema provider that maps any thrift type to a Beam schema, allowing for custom thrift
 typedef entries (which cannot be resolved using the available metadata) to be manually
 registered with their corresponding beam types. 
- CUSTOM_SOURCE_FORMAT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- CustomCoder<T> - Class in org.apache.beam.sdk.coders
- 
- CustomCoder() - Constructor for class org.apache.beam.sdk.coders.CustomCoder
-  
- Customer - Class in org.apache.beam.sdk.extensions.sql.example.model
- 
Describes a customer. 
- Customer(int, String, String) - Constructor for class org.apache.beam.sdk.extensions.sql.example.model.Customer
-  
- Customer() - Constructor for class org.apache.beam.sdk.extensions.sql.example.model.Customer
-  
- CustomHttpErrors - Class in org.apache.beam.sdk.extensions.gcp.util
- 
An optional component to use with the RetryHttpRequestInitializerin order to provide
 custom errors for failing http calls.
 
- CustomHttpErrors.Builder - Class in org.apache.beam.sdk.extensions.gcp.util
- 
A Builder which allows building immutable CustomHttpErrors object. 
- CustomHttpErrors.MatcherAndError - Class in org.apache.beam.sdk.extensions.gcp.util
- 
A simple Tuple class for creating a list of HttpResponseMatcher and HttpResponseCustomError to
 print for the responses. 
- CustomTableResolver - Interface in org.apache.beam.sdk.extensions.sql.meta
- 
Interface that table providers can implement if they require custom table name resolution. 
- CustomTimestampPolicyWithLimitedDelay<K,V> - Class in org.apache.beam.sdk.io.kafka
- 
A policy for custom record timestamps where timestamps within a partition are expected to be
 roughly monotonically increasing with a cap on out of order event delays (say 1 minute). 
- CustomTimestampPolicyWithLimitedDelay(SerializableFunction<KafkaRecord<K, V>, Instant>, Duration, Optional<Instant>) - Constructor for class org.apache.beam.sdk.io.kafka.CustomTimestampPolicyWithLimitedDelay
- 
A policy for custom record timestamps where timestamps are expected to be roughly monotonically
 increasing with out of order event delays less than maxDelay.
 
- CustomX509TrustManager - Class in org.apache.beam.sdk.io.splunk
- 
A Custom X509TrustManager that trusts a user provided CA and default CA's. 
- CustomX509TrustManager(X509Certificate) - Constructor for class org.apache.beam.sdk.io.splunk.CustomX509TrustManager
-  
- factory - Variable in class org.apache.beam.sdk.extensions.sql.impl.CalciteFactoryWrapper
-  
- FACTORY - Static variable in class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner
-  
- FACTORY - Static variable in class org.apache.beam.sdk.extensions.sql.impl.parser.BeamSqlParser
- 
Parser factory. 
- FACTORY - Static variable in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
-  
- Factory() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel.Factory
-  
- FACTORY - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
-  
- FACTORY - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
- 
Factory for creating Pubsub clients using gRPC transport. 
- FACTORY - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
- 
Factory for creating Pubsub clients using Json transport. 
- Factory<T> - Interface in org.apache.beam.sdk.schemas
- 
A Factory interface for schema-related objects for a specific Java type. 
- failed(Exception) - Method in interface org.apache.beam.runners.local.PipelineMessageReceiver
- 
Report that a failure has occurred. 
- failed(Error) - Method in interface org.apache.beam.runners.local.PipelineMessageReceiver
- 
Report that a failure has occurred. 
- FAILED - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.Write
- 
The tag for the failed writes to HL7v2 store`. 
- FAILED_BODY - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write
- 
The tag for the failed writes to FHIR store. 
- FAILED_BUNDLES - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.ExecuteBundles
- 
The TupleTag used for bundles that failed to be executed for any reason. 
- FAILED_FILES - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write
- 
The tag for the files that failed to FHIR store. 
- FAILED_PUBLISH_TAG - Static variable in class org.apache.beam.sdk.io.solace.SolaceIO.Write
-  
- FAILED_WRITES - Static variable in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
-  
- failedRecords(List<RecT>, List<ResT>) - Method in class org.apache.beam.sdk.io.aws2.common.AsyncBatchWriteHandler
-  
- FailedRunningPipelineResults - Class in org.apache.beam.runners.jet
- 
Alternative implementation of  PipelineResult used to avoid throwing Exceptions in certain
 situations. 
- FailedRunningPipelineResults(RuntimeException) - Constructor for class org.apache.beam.runners.jet.FailedRunningPipelineResults
-  
- FailedWritesException(List<FirestoreV1.WriteFailure>) - Constructor for exception org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.FailedWritesException
-  
- failOnInsert(Map<TableRow, List<TableDataInsertAllResponse.InsertErrors>>) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
- 
Cause a given  TableRow object to fail when it's inserted. 
- FailsafeValueInSingleWindow<T,ErrorT> - Class in org.apache.beam.sdk.values
- 
An immutable tuple of value, timestamp, window, and pane. 
- FailsafeValueInSingleWindow() - Constructor for class org.apache.beam.sdk.values.FailsafeValueInSingleWindow
-  
- FailsafeValueInSingleWindow.Coder<T,ErrorT> - Class in org.apache.beam.sdk.values
- 
- failure(String, String, Metadata, Throwable) - Static method in class org.apache.beam.sdk.io.tika.ParseResult
-  
- Failure - Class in org.apache.beam.sdk.schemas.io
- 
A generic failure of an SQL transform. 
- Failure() - Constructor for class org.apache.beam.sdk.schemas.io.Failure
-  
- failure(PAssert.PAssertionSite, Throwable) - Static method in class org.apache.beam.sdk.testing.SuccessOrFailure
-  
- Failure() - Constructor for class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Failure
-  
- Failure.Builder - Class in org.apache.beam.sdk.schemas.io
-  
- FAILURE_COUNTER - Static variable in class org.apache.beam.sdk.testing.PAssert
-  
- FAILURE_TAG - Static variable in class org.apache.beam.sdk.extensions.ml.RecommendationAICreateCatalogItem
-  
- FAILURE_TAG - Static variable in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportCatalogItems
-  
- FAILURE_TAG - Static variable in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportUserEvents
-  
- FAILURE_TAG - Static variable in class org.apache.beam.sdk.extensions.ml.RecommendationAIPredict
-  
- FAILURE_TAG - Static variable in class org.apache.beam.sdk.extensions.ml.RecommendationAIWriteUserEvent
-  
- FailureCollectorWrapper - Class in org.apache.beam.sdk.io.cdap.context
- 
Class FailureCollectorWrapper is a class for collecting ValidationFailure. 
- FailureCollectorWrapper() - Constructor for class org.apache.beam.sdk.io.cdap.context.FailureCollectorWrapper
-  
- failures() - Method in class org.apache.beam.sdk.transforms.WithFailures.Result
-  
- failuresTo(List<PCollection<FailureElementT>>) - Method in class org.apache.beam.sdk.transforms.WithFailures.Result
- 
Adds the failure collection to the passed list and returns just the output collection. 
- FakeBigQueryServerStream(List<T>) - Constructor for class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices.FakeBigQueryServerStream
-  
- FakeBigQueryServices - Class in org.apache.beam.sdk.io.gcp.testing
- 
A fake implementation of BigQuery's query service.. 
- FakeBigQueryServices() - Constructor for class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
-  
- FakeBigQueryServices.FakeBigQueryServerStream<T> - Class in org.apache.beam.sdk.io.gcp.testing
- 
An implementation of BigQueryServerStreamwhich takes aListas theIterableto simulate a server stream.
 
- FakeDatasetService - Class in org.apache.beam.sdk.io.gcp.testing
- 
A fake dataset service that can be serialized, for use in testReadFromTable. 
- FakeDatasetService() - Constructor for class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- FakeJobService - Class in org.apache.beam.sdk.io.gcp.testing
- 
A fake implementation of BigQuery's job service. 
- FakeJobService() - Constructor for class org.apache.beam.sdk.io.gcp.testing.FakeJobService
-  
- FakeJobService(int) - Constructor for class org.apache.beam.sdk.io.gcp.testing.FakeJobService
-  
- Fanout() - Constructor for class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields.Fanout
-  
- features() - Method in class org.apache.beam.sdk.extensions.ml.AnnotateText
-  
- fetchDataflowJobId() - Static method in class org.apache.beam.sdk.extensions.gcp.util.GceMetadataUtil
-  
- fetchDataflowJobName() - Static method in class org.apache.beam.sdk.extensions.gcp.util.GceMetadataUtil
-  
- fetchDataflowWorkerId() - Static method in class org.apache.beam.sdk.extensions.gcp.util.GceMetadataUtil
-  
- FetchHL7v2Message() - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.HL7v2Read.FetchHL7v2Message
- 
Instantiates a new Fetch HL7v2 message DoFn. 
- FetchHL7v2Message() - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.Read.FetchHL7v2Message
- 
Instantiates a new Fetch HL7v2 message DoFn. 
- fewKeys(CombineFnBase.GlobalCombineFn<? super InputT, ?, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Combine
- 
- fewKeys() - Method in class org.apache.beam.sdk.transforms.GroupByKey
- 
Returns whether it groups just few keys. 
- FhirBundleParameter - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
FhirBundleParameter represents a FHIR bundle in JSON format to be executed on a FHIR store. 
- FhirBundleParameter() - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.FhirBundleParameter
-  
- FhirBundleResponse - Class in org.apache.beam.sdk.io.gcp.healthcare
-  
- FhirBundleResponse() - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.FhirBundleResponse
-  
- FhirIO - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
- FhirIO() - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.FhirIO
-  
- FhirIO.Deidentify - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
Deidentify FHIR resources from a FHIR store to a destination FHIR store. 
- FhirIO.Deidentify.DeidentifyFn - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
A function that schedules a deidentify operation and monitors the status. 
- FhirIO.ExecuteBundles - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
The type Execute bundles. 
- FhirIO.ExecuteBundlesResult - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
ExecuteBundlesResult contains both successfully executed bundles and information help debugging
 failed executions (eg metadata & error msgs). 
- FhirIO.Export - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
Export FHIR resources from a FHIR store to new line delimited json files on GCS or BigQuery. 
- FhirIO.Export.ExportResourcesFn - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
A function that schedules an export operation and monitors the status. 
- FhirIO.Import - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
Writes each bundle of elements to a new-line delimited JSON file on GCS and issues a
 fhirStores.import Request for that file. 
- FhirIO.Import.ContentStructure - Enum in org.apache.beam.sdk.io.gcp.healthcare
- 
The enum Content structure. 
- FhirIO.Read - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
The type Read. 
- FhirIO.Read.Result - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
The type Result. 
- FhirIO.Search<T> - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
The type Search. 
- FhirIO.Search.Result - Class in org.apache.beam.sdk.io.gcp.healthcare
-  
- FhirIO.Write - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
The type Write. 
- FhirIO.Write.AbstractResult - Class in org.apache.beam.sdk.io.gcp.healthcare
-  
- FhirIO.Write.Result - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
The type Result. 
- FhirIO.Write.WriteMethod - Enum in org.apache.beam.sdk.io.gcp.healthcare
- 
The enum Write method. 
- FhirIOPatientEverything - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
The type FhirIOPatientEverything for querying a FHIR Patient resource's compartment. 
- FhirIOPatientEverything() - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.FhirIOPatientEverything
-  
- FhirIOPatientEverything.PatientEverythingParameter - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
PatientEverythingParameter defines required attributes for a FHIR GetPatientEverything request
 in  FhirIOPatientEverything. 
- FhirIOPatientEverything.Result - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
- FhirResourcePagesIterator(HttpHealthcareApiClient.FhirResourcePagesIterator.FhirMethod, HealthcareApiClient, String, String, String, Map<String, Object>) - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient.FhirResourcePagesIterator
-  
- FhirSearchParameter<T> - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
FhirSearchParameter represents the query parameters for a FHIR search request, used as a
 parameter for  FhirIO.Search. 
- FhirSearchParameterCoder<T> - Class in org.apache.beam.sdk.io.gcp.healthcare
- 
- fhirStoresImport(String, String, String, FhirIO.Import.ContentStructure) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write
- 
Import method for batch writing resources. 
- fhirStoresImport(String, String, FhirIO.Import.ContentStructure) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write
-  
- fhirStoresImport(ValueProvider<String>, ValueProvider<String>, ValueProvider<String>, FhirIO.Import.ContentStructure) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write
-  
- field(Row, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedRow
-  
- Field() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.ArrayOfNestedStringBundle.Field
-  
- field - Variable in class org.apache.beam.sdk.jmh.schemas.RowBundles.ArrayOfNestedStringBundle.Field
-  
- Field() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.ArrayOfStringBundle.Field
-  
- field - Variable in class org.apache.beam.sdk.jmh.schemas.RowBundles.ArrayOfStringBundle.Field
-  
- Field() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.ByteBufferBundle.Field
-  
- field - Variable in class org.apache.beam.sdk.jmh.schemas.RowBundles.ByteBufferBundle.Field
-  
- Field() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.BytesBundle.Field
-  
- field - Variable in class org.apache.beam.sdk.jmh.schemas.RowBundles.BytesBundle.Field
-  
- Field() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.DateTimeBundle.Field
-  
- field - Variable in class org.apache.beam.sdk.jmh.schemas.RowBundles.DateTimeBundle.Field
-  
- Field() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.IntBundle.Field
-  
- field - Variable in class org.apache.beam.sdk.jmh.schemas.RowBundles.IntBundle.Field
-  
- Field() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.MapOfIntBundle.Field
-  
- field - Variable in class org.apache.beam.sdk.jmh.schemas.RowBundles.MapOfIntBundle.Field
-  
- Field() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.MapOfNestedIntBundle.Field
-  
- field - Variable in class org.apache.beam.sdk.jmh.schemas.RowBundles.MapOfNestedIntBundle.Field
-  
- Field() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.NestedBytesBundle.Field
-  
- field - Variable in class org.apache.beam.sdk.jmh.schemas.RowBundles.NestedBytesBundle.Field
-  
- Field() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.NestedIntBundle.Field
-  
- field - Variable in class org.apache.beam.sdk.jmh.schemas.RowBundles.NestedIntBundle.Field
-  
- Field() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.StringBuilderBundle.Field
-  
- field - Variable in class org.apache.beam.sdk.jmh.schemas.RowBundles.StringBuilderBundle.Field
-  
- Field() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.StringBundle.Field
-  
- field - Variable in class org.apache.beam.sdk.jmh.schemas.RowBundles.StringBundle.Field
-  
- Field() - Constructor for class org.apache.beam.sdk.schemas.Schema.Field
-  
- field(String, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.transforms.AddFields.Inner
- 
Add a new field of the specified type. 
- field(String, Schema.FieldType, Object) - Method in class org.apache.beam.sdk.schemas.transforms.AddFields.Inner
- 
Add a new field of the specified type. 
- fieldAccess(FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.transforms.Select
- 
- FieldAccessDescriptor - Class in org.apache.beam.sdk.schemas
- 
Used inside of a  DoFn to describe which fields in a schema
 type need to be accessed for processing. 
- FieldAccessDescriptor() - Constructor for class org.apache.beam.sdk.schemas.FieldAccessDescriptor
-  
- fieldAccessDescriptor(FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.transforms.CoGroup.By
- 
Join by the following field access descriptor. 
- FieldAccessDescriptor.FieldDescriptor - Class in org.apache.beam.sdk.schemas
- 
Description of a single field. 
- FieldAccessDescriptor.FieldDescriptor.Builder - Class in org.apache.beam.sdk.schemas
- 
Builder class. 
- FieldAccessDescriptor.FieldDescriptor.ListQualifier - Enum in org.apache.beam.sdk.schemas
- 
Qualifier for a list selector. 
- FieldAccessDescriptor.FieldDescriptor.MapQualifier - Enum in org.apache.beam.sdk.schemas
- 
Qualifier for a map selector. 
- FieldAccessDescriptor.FieldDescriptor.Qualifier - Class in org.apache.beam.sdk.schemas
- 
OneOf union for a collection selector. 
- FieldAccessDescriptor.FieldDescriptor.Qualifier.Kind - Enum in org.apache.beam.sdk.schemas
- 
The kind of qualifier. 
- FieldAccessDescriptorParser - Class in org.apache.beam.sdk.schemas.parser
- 
Parser for textual field-access selector. 
- FieldAccessDescriptorParser() - Constructor for class org.apache.beam.sdk.schemas.parser.FieldAccessDescriptorParser
-  
- FieldDescriptor() - Constructor for class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor
-  
- fieldFromType(TypeDescriptor, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.StaticSchemaInference
- 
Map a Java field type to a Beam Schema FieldType. 
- fieldIds(Integer...) - Static method in class org.apache.beam.sdk.schemas.transforms.CoGroup.By
- 
Join by the following field ids. 
- fieldIds(Integer...) - Static method in class org.apache.beam.sdk.schemas.transforms.Select
- 
Select a set of top-level field ids from the row. 
- fieldIdsAccessed() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Return the field ids accessed. 
- fieldNames(String...) - Static method in class org.apache.beam.sdk.schemas.transforms.CoGroup.By
- 
Join by the following field names. 
- fieldNames(String...) - Static method in class org.apache.beam.sdk.schemas.transforms.Select
- 
Select a set of top-level field names from the row. 
- fieldNamesAccessed() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Return the field names accessed. 
- fields() - Method in class org.apache.beam.sdk.io.splunk.SplunkEvent
-  
- fields(String...) - Static method in class org.apache.beam.sdk.schemas.transforms.DropFields
-  
- fields(Integer...) - Static method in class org.apache.beam.sdk.schemas.transforms.DropFields
-  
- fields(FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.transforms.DropFields
-  
- Fields() - Constructor for class org.apache.beam.sdk.schemas.transforms.Select.Fields
-  
- fields() - Static method in class org.apache.beam.sdk.state.StateKeySpec
-  
- FieldsEqual() - Constructor for class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual
-  
- fieldSpecifier() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- FieldSpecifierContext(ParserRuleContext, int) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.FieldSpecifierContext
-  
- FieldSpecifierNotationBaseListener - Class in org.apache.beam.sdk.schemas.parser.generated
- 
This class provides an empty implementation of  FieldSpecifierNotationListener,
 which can be extended to create a listener which only needs to handle a subset
 of the available methods. 
- FieldSpecifierNotationBaseListener() - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
-  
- FieldSpecifierNotationBaseVisitor<T> - Class in org.apache.beam.sdk.schemas.parser.generated
- 
This class provides an empty implementation of  FieldSpecifierNotationVisitor,
 which can be extended to create a visitor which only needs to handle a subset
 of the available methods. 
- FieldSpecifierNotationBaseVisitor() - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
-  
- FieldSpecifierNotationLexer - Class in org.apache.beam.sdk.schemas.parser.generated
-  
- FieldSpecifierNotationLexer(CharStream) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- FieldSpecifierNotationListener - Interface in org.apache.beam.sdk.schemas.parser.generated
- 
- FieldSpecifierNotationParser - Class in org.apache.beam.sdk.schemas.parser.generated
-  
- FieldSpecifierNotationParser(TokenStream) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- FieldSpecifierNotationParser.ArrayQualifierContext - Class in org.apache.beam.sdk.schemas.parser.generated
-  
- FieldSpecifierNotationParser.ArrayQualifierListContext - Class in org.apache.beam.sdk.schemas.parser.generated
-  
- FieldSpecifierNotationParser.DotExpressionComponentContext - Class in org.apache.beam.sdk.schemas.parser.generated
-  
- FieldSpecifierNotationParser.DotExpressionContext - Class in org.apache.beam.sdk.schemas.parser.generated
-  
- FieldSpecifierNotationParser.FieldSpecifierContext - Class in org.apache.beam.sdk.schemas.parser.generated
-  
- FieldSpecifierNotationParser.MapQualifierContext - Class in org.apache.beam.sdk.schemas.parser.generated
-  
- FieldSpecifierNotationParser.MapQualifierListContext - Class in org.apache.beam.sdk.schemas.parser.generated
-  
- FieldSpecifierNotationParser.QualifiedComponentContext - Class in org.apache.beam.sdk.schemas.parser.generated
-  
- FieldSpecifierNotationParser.QualifierListContext - Class in org.apache.beam.sdk.schemas.parser.generated
-  
- FieldSpecifierNotationParser.QualifyComponentContext - Class in org.apache.beam.sdk.schemas.parser.generated
-  
- FieldSpecifierNotationParser.SimpleIdentifierContext - Class in org.apache.beam.sdk.schemas.parser.generated
-  
- FieldSpecifierNotationParser.WildcardContext - Class in org.apache.beam.sdk.schemas.parser.generated
-  
- FieldSpecifierNotationVisitor<T> - Interface in org.apache.beam.sdk.schemas.parser.generated
- 
- FieldType() - Constructor for class org.apache.beam.sdk.schemas.Schema.FieldType
-  
- FieldTypeDescriptors - Class in org.apache.beam.sdk.schemas
- 
Utilities for converting between  Schema field types and  TypeDescriptors that
 define Java objects which can represent these field types. 
- FieldTypeDescriptors() - Constructor for class org.apache.beam.sdk.schemas.FieldTypeDescriptors
-  
- fieldTypeForJavaType(TypeDescriptor) - Static method in class org.apache.beam.sdk.schemas.FieldTypeDescriptors
- 
- fieldUpdate(String, String, String) - Static method in class org.apache.beam.sdk.io.mongodb.UpdateField
-  
- FieldValueGetter<ObjectT,ValueT> - Interface in org.apache.beam.sdk.schemas
- 
For internal use only; no backwards-compatibility guarantees. 
- fieldValueGetters(TypeDescriptor<T>, Schema) - Method in class org.apache.beam.sdk.extensions.avro.schemas.AvroRecordSchema
-  
- fieldValueGetters(TypeDescriptor<T>, Schema) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoMessageSchema
-  
- fieldValueGetters(TypeDescriptor<T>, Schema) - Method in class org.apache.beam.sdk.io.aws2.schemas.AwsSchemaProvider
-  
- fieldValueGetters(TypeDescriptor<T>, Schema) - Method in class org.apache.beam.sdk.io.thrift.ThriftSchema
-  
- fieldValueGetters(TypeDescriptor<T>, Schema) - Method in class org.apache.beam.sdk.schemas.AutoValueSchema
-  
- fieldValueGetters(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
- 
- fieldValueGetters(TypeDescriptor<T>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
- 
Deprecated. 
- fieldValueGetters(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProviderV2
-  
- fieldValueGetters(TypeDescriptor<T>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProviderV2
-  
- fieldValueGetters(TypeDescriptor<T>, Schema) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema
-  
- fieldValueGetters(TypeDescriptor<T>, Schema) - Method in class org.apache.beam.sdk.schemas.JavaFieldSchema
-  
- FieldValueSetter<ObjectT,ValueT> - Interface in org.apache.beam.sdk.schemas
- 
For internal use only; no backwards-compatibility guarantees. 
- FieldValueTypeInformation - Class in org.apache.beam.sdk.schemas
- 
Represents type information for a Java type that will be used to infer a Schema type. 
- FieldValueTypeInformation() - Constructor for class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- FieldValueTypeInformation.Builder - Class in org.apache.beam.sdk.schemas
-  
- fieldValueTypeInformations(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.extensions.avro.schemas.AvroRecordSchema
-  
- fieldValueTypeInformations(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoMessageSchema
-  
- fieldValueTypeInformations(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.io.aws2.schemas.AwsSchemaProvider
-  
- fieldValueTypeInformations(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.io.thrift.ThriftSchema
-  
- fieldValueTypeInformations(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.schemas.AutoValueSchema
-  
- fieldValueTypeInformations(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
- 
- fieldValueTypeInformations(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
- 
Deprecated. 
- fieldValueTypeInformations(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProviderV2
-  
- fieldValueTypeInformations(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProviderV2
-  
- fieldValueTypeInformations(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema
-  
- fieldValueTypeInformations(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.schemas.JavaFieldSchema
-  
- FieldValueTypeSupplier - Interface in org.apache.beam.sdk.schemas.utils
- 
A naming policy for schema fields. 
- FILE_ARTIFACT_URN - Static variable in class org.apache.beam.runners.fnexecution.artifact.ArtifactRetrievalService
-  
- FILE_NAME_FIELD - Static variable in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformProvider
-  
- FILE_TRIGGERING_BYTE_COUNT - Static variable in class org.apache.beam.sdk.io.WriteFiles
-  
- FILE_TRIGGERING_RECORD_BUFFERING_DURATION - Static variable in class org.apache.beam.sdk.io.WriteFiles
-  
- FILE_TRIGGERING_RECORD_COUNT - Static variable in class org.apache.beam.sdk.io.WriteFiles
-  
- FileBasedReader(FileBasedSource<T>) - Constructor for class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
- 
Subclasses should not perform IO operations at the constructor. 
- FileBasedSink<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io
- 
Abstract class for file-based output. 
- FileBasedSink(ValueProvider<ResourceId>, FileBasedSink.DynamicDestinations<?, DestinationT, OutputT>) - Constructor for class org.apache.beam.sdk.io.FileBasedSink
- 
Construct a  FileBasedSink with the given temp directory, producing uncompressed files. 
- FileBasedSink(ValueProvider<ResourceId>, FileBasedSink.DynamicDestinations<?, DestinationT, OutputT>, FileBasedSink.WritableByteChannelFactory) - Constructor for class org.apache.beam.sdk.io.FileBasedSink
- 
Construct a  FileBasedSink with the given temp directory and output channel type. 
- FileBasedSink(ValueProvider<ResourceId>, FileBasedSink.DynamicDestinations<?, DestinationT, OutputT>, Compression) - Constructor for class org.apache.beam.sdk.io.FileBasedSink
- 
Construct a  FileBasedSink with the given temp directory and output channel type. 
- FileBasedSink.CompressionType - Enum in org.apache.beam.sdk.io
- 
- FileBasedSink.DynamicDestinations<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io
- 
- FileBasedSink.FilenamePolicy - Class in org.apache.beam.sdk.io
- 
A naming policy for output files. 
- FileBasedSink.FileResult<DestinationT> - Class in org.apache.beam.sdk.io
- 
Result of a single bundle write. 
- FileBasedSink.FileResultCoder<DestinationT> - Class in org.apache.beam.sdk.io
- 
- FileBasedSink.OutputFileHints - Interface in org.apache.beam.sdk.io
- 
Provides hints about how to generate output files, such as a suggested filename suffix (e.g. 
- FileBasedSink.WritableByteChannelFactory - Interface in org.apache.beam.sdk.io
- 
- FileBasedSink.WriteOperation<DestinationT,OutputT> - Class in org.apache.beam.sdk.io
- 
Abstract operation that manages the process of writing to  FileBasedSink. 
- FileBasedSink.Writer<DestinationT,OutputT> - Class in org.apache.beam.sdk.io
- 
- FileBasedSource<T> - Class in org.apache.beam.sdk.io
- 
A common base class for all file-based  Sources. 
- FileBasedSource(ValueProvider<String>, EmptyMatchTreatment, long) - Constructor for class org.apache.beam.sdk.io.FileBasedSource
- 
Create a FileBaseSourcebased on a file or a file pattern specification, with the given
 strategy for treating filepatterns that do not match any files.
 
- FileBasedSource(ValueProvider<String>, long) - Constructor for class org.apache.beam.sdk.io.FileBasedSource
- 
- FileBasedSource(MatchResult.Metadata, long, long, long) - Constructor for class org.apache.beam.sdk.io.FileBasedSource
- 
Create a FileBasedSourcebased on a single file.
 
- FileBasedSource.FileBasedReader<T> - Class in org.apache.beam.sdk.io
- 
A  reader that implements code common to readers of  FileBasedSources. 
- FileBasedSource.Mode - Enum in org.apache.beam.sdk.io
- 
A given FileBasedSourcerepresents a file resource of one of these types.
 
- FileChecksumMatcher - Class in org.apache.beam.sdk.testing
- 
Matcher to verify checksum of the contents of an ShardedFilein E2E test.
 
- fileContentsHaveChecksum(String) - Static method in class org.apache.beam.sdk.testing.FileChecksumMatcher
-  
- FileIO - Class in org.apache.beam.sdk.io
- 
General-purpose transforms for working with files: listing files (matching), reading and writing. 
- FileIO() - Constructor for class org.apache.beam.sdk.io.FileIO
-  
- FileIO.Match - Class in org.apache.beam.sdk.io
- 
- FileIO.MatchAll - Class in org.apache.beam.sdk.io
- 
- FileIO.MatchConfiguration - Class in org.apache.beam.sdk.io
- 
Describes configuration for matching filepatterns, such as  EmptyMatchTreatment and
 continuous watching for matching files. 
- FileIO.ReadableFile - Class in org.apache.beam.sdk.io
- 
A utility class for accessing a potentially compressed file. 
- FileIO.ReadMatches - Class in org.apache.beam.sdk.io
- 
- FileIO.ReadMatches.DirectoryTreatment - Enum in org.apache.beam.sdk.io
- 
Enum to control how directories are handled. 
- FileIO.Sink<ElementT> - Interface in org.apache.beam.sdk.io
- 
- FileIO.Write<DestinationT,UserT> - Class in org.apache.beam.sdk.io
- 
- FileIO.Write.FileNaming - Interface in org.apache.beam.sdk.io
- 
A policy for generating names for shard files. 
- FilenamePolicy() - Constructor for class org.apache.beam.sdk.io.FileBasedSink.FilenamePolicy
-  
- filepattern(String) - Method in class org.apache.beam.sdk.io.FileIO.Match
- 
Matches the given filepattern. 
- filepattern(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.FileIO.Match
- 
- filepattern(String) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Parse
- 
Matches the given filepattern. 
- filepattern(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Parse
- 
- FileReadSchemaTransformConfiguration - Class in org.apache.beam.sdk.io.fileschematransform
-  
- FileReadSchemaTransformConfiguration() - Constructor for class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformConfiguration
-  
- FileReadSchemaTransformConfiguration.Builder - Class in org.apache.beam.sdk.io.fileschematransform
-  
- FileReadSchemaTransformFormatProvider - Interface in org.apache.beam.sdk.io.fileschematransform
- 
- FileReadSchemaTransformProvider - Class in org.apache.beam.sdk.io.fileschematransform
-  
- FileReadSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformProvider
-  
- FileReporter - Class in org.apache.beam.runners.flink.metrics
- 
Flink metrics reporterfor writing
 metrics to a file specified via the "metrics.reporter.file.path" config key (assuming an alias of
 "file" for this reporter in the "metrics.reporters" setting).
 
- FileReporter() - Constructor for class org.apache.beam.runners.flink.metrics.FileReporter
-  
- FileResult(ResourceId, int, BoundedWindow, PaneInfo, DestinationT) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.FileResult
-  
- FileResultCoder(Coder<BoundedWindow>, Coder<DestinationT>) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
-  
- fileSize(GcsPath) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
- 
Returns the file size from GCS or throws FileNotFoundExceptionif the resource does not
 exist.
 
- FileStagingOptions - Interface in org.apache.beam.sdk.options
- 
File staging related options. 
- FileSystem<ResourceIdT extends ResourceId> - Class in org.apache.beam.sdk.io
- 
File system interface in Beam. 
- FileSystem() - Constructor for class org.apache.beam.sdk.io.FileSystem
-  
- FileSystem.LineageLevel - Enum in org.apache.beam.sdk.io
-  
- FileSystemRegistrar - Interface in org.apache.beam.sdk.io
- 
- FileSystems - Class in org.apache.beam.sdk.io
- 
- FileSystems() - Constructor for class org.apache.beam.sdk.io.FileSystems
-  
- FileSystemUtils - Class in org.apache.beam.sdk.io
-  
- FileSystemUtils() - Constructor for class org.apache.beam.sdk.io.FileSystemUtils
-  
- FileWriteSchemaTransformConfiguration - Class in org.apache.beam.sdk.io.fileschematransform
- 
- FileWriteSchemaTransformConfiguration() - Constructor for class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration
-  
- FileWriteSchemaTransformConfiguration.Builder - Class in org.apache.beam.sdk.io.fileschematransform
-  
- FileWriteSchemaTransformConfiguration.CsvConfiguration - Class in org.apache.beam.sdk.io.fileschematransform
- 
Configures extra details related to writing CSV formatted files. 
- FileWriteSchemaTransformConfiguration.CsvConfiguration.Builder - Class in org.apache.beam.sdk.io.fileschematransform
-  
- FileWriteSchemaTransformConfiguration.ParquetConfiguration - Class in org.apache.beam.sdk.io.fileschematransform
- 
Configures extra details related to writing Parquet formatted files. 
- FileWriteSchemaTransformConfiguration.ParquetConfiguration.Builder - Class in org.apache.beam.sdk.io.fileschematransform
-  
- FileWriteSchemaTransformConfiguration.XmlConfiguration - Class in org.apache.beam.sdk.io.fileschematransform
- 
Configures extra details related to writing XML formatted files. 
- FileWriteSchemaTransformConfiguration.XmlConfiguration.Builder - Class in org.apache.beam.sdk.io.fileschematransform
-  
- FileWriteSchemaTransformFormatProvider - Interface in org.apache.beam.sdk.io.fileschematransform
- 
- FileWriteSchemaTransformFormatProviders - Class in org.apache.beam.sdk.io.fileschematransform
- 
- FileWriteSchemaTransformFormatProviders() - Constructor for class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformFormatProviders
-  
- FileWriteSchemaTransformProvider - Class in org.apache.beam.sdk.io.fileschematransform
- 
- FileWriteSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformProvider
-  
- FillGaps<ValueT> - Class in org.apache.beam.sdk.extensions.timeseries
- 
Fill gaps in timeseries. 
- FillGaps() - Constructor for class org.apache.beam.sdk.extensions.timeseries.FillGaps
-  
- FillGaps.FillGapsDoFn<ValueT> - Class in org.apache.beam.sdk.extensions.timeseries
-  
- FillGaps.InterpolateData<ValueT> - Class in org.apache.beam.sdk.extensions.timeseries
- 
Argument to withInterpolateFunction function. 
- Filter - Class in org.apache.beam.sdk.schemas.transforms
- 
A  PTransform for filtering a collection of schema types. 
- Filter() - Constructor for class org.apache.beam.sdk.schemas.transforms.Filter
-  
- Filter<T> - Class in org.apache.beam.sdk.transforms
- 
PTransforms for filtering from aPCollectionthe elements satisfying a predicate,
 or satisfying an inequality with a given value based on the elements' natural ordering.
 
- Filter.Inner<T> - Class in org.apache.beam.sdk.schemas.transforms
- 
Implementation of the filter. 
- filterCharacters(String) - Method in class org.apache.beam.runners.flink.metrics.FileReporter
-  
- FilterForMutationDoFn - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn
-  
- FilterForMutationDoFn() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.FilterForMutationDoFn
-  
- finalizeAllOutstandingBundles() - Method in class org.apache.beam.runners.fnexecution.control.BundleFinalizationHandlers.InMemoryFinalizer
- 
All finalization requests will be sent without waiting for the responses. 
- finalizeCheckpoint() - Method in class org.apache.beam.runners.spark.io.EmptyCheckpointMark
-  
- finalizeCheckpoint() - Method in class org.apache.beam.sdk.io.CountingSource.CounterMark
-  
- finalizeCheckpoint() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.CheckpointMarkImpl
-  
- finalizeCheckpoint() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark
-  
- finalizeCheckpoint() - Method in class org.apache.beam.sdk.io.solace.read.SolaceCheckpointMark
-  
- finalizeCheckpoint() - Method in interface org.apache.beam.sdk.io.UnboundedSource.CheckpointMark
- 
Called by the system to signal that this checkpoint mark has been committed along with all
 the records which have been read from the  UnboundedSource.UnboundedReader since the previous
 checkpoint was taken. 
- finalizeCheckpoint() - Method in class org.apache.beam.sdk.io.UnboundedSource.CheckpointMark.NoopCheckpointMark
-  
- finalizeDestination(DestinationT, BoundedWindow, Integer, Collection<FileBasedSink.FileResult<DestinationT>>) - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-  
- finalizeWriteStream(String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.WriteStreamService
- 
Finalize a write stream. 
- finalizeWriteStream(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.WriteStreamServiceImpl
-  
- finalizeWriteStream(String) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- find(String) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
- find(Pattern) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
- find(String, int) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
- find(Pattern, int) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
- find(String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
- find(Pattern, String) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
- Find(Pattern, int) - Constructor for class org.apache.beam.sdk.transforms.Regex.Find
-  
- findAll(String) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
- findAll(Pattern) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
- FindAll(Pattern) - Constructor for class org.apache.beam.sdk.transforms.Regex.FindAll
-  
- findAllTableIndexes() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao
- 
Finds all indexes for the metadata table. 
- findAvailablePort() - Static method in class org.apache.beam.sdk.extensions.python.PythonService
-  
- findDateTimePattern(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
-  
- findDateTimePattern(String, ImmutableMap<DateTimeUtils.TimestampPatterns, DateTimeFormatter>) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
-  
- findKV(String, int, int) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
- findKV(Pattern, int, int) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
- findKV(String, String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
- findKV(Pattern, String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
- FindKV(Pattern, int, int) - Constructor for class org.apache.beam.sdk.transforms.Regex.FindKV
-  
- FindName(Pattern, String) - Constructor for class org.apache.beam.sdk.transforms.Regex.FindName
-  
- FindNameKV(Pattern, String, String) - Constructor for class org.apache.beam.sdk.transforms.Regex.FindNameKV
-  
- FindQuery - Class in org.apache.beam.sdk.io.mongodb
- 
Builds a MongoDB FindQuery object. 
- FindQuery() - Constructor for class org.apache.beam.sdk.io.mongodb.FindQuery
-  
- finish(DoFn<SequencedMessage, Row>.FinishBundleContext) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.ErrorFn
-  
- finish() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.ErrorCounterFn
-  
- finish(DoFn<DataChangeRecord, Row>.FinishBundleContext) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.DataChangeRecordToRow
-  
- finish(DoFn<byte[], Row>.FinishBundleContext) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformProvider.ErrorFn
-  
- finishBundle() - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
-  
- finishBundle(DoFn<Row, Row>.FinishBundleContext, PipelineOptions) - Method in interface org.apache.beam.sdk.extensions.sql.BeamSqlSeekableTable
-  
- finishBundle(DoFn<Iterable<KV<DestinationT, WriteTables.Result>>, Iterable<KV<TableDestination, WriteTables.Result>>>.FinishBundleContext) - Method in class org.apache.beam.sdk.io.gcp.bigquery.UpdateSchemaDestination
-  
- finishBundle() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write.PubsubBoundedWriter
-  
- finishBundle() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.PubsubLiteSink
-  
- finishBundle(DoFn<KV<Integer, Solace.Record>, Solace.PublishResult>.FinishBundleContext) - Method in class org.apache.beam.sdk.io.solace.write.UnboundedBatchedSolaceWriter
-  
- finishBundle(DoFn<KV<Integer, Solace.Record>, Solace.PublishResult>.FinishBundleContext) - Method in class org.apache.beam.sdk.io.solace.write.UnboundedStreamingSolaceWriter
-  
- finishBundle() - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- finishBundle(DoFn<T, KV<Long, PCollectionViews.ValueOrMetadata<T, OffsetRange>>>.FinishBundleContext) - Method in class org.apache.beam.sdk.transforms.View.ToListViewDoFn
-  
- FinishBundleContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.FinishBundleContext
-  
- finishSpecifying() - Method in interface org.apache.beam.sdk.state.StateSpec
- 
For internal use only; no backwards-compatibility guarantees. 
- finishSpecifying(PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PCollection
- 
After building, finalizes this  PValue to make it ready for running. 
- finishSpecifying(PInput, PTransform<?, ?>) - Method in interface org.apache.beam.sdk.values.PValue
- 
After building, finalizes this  PValue to make it ready for being used as an input to a
  PTransform. 
- finishSpecifying(PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PValueBase
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.io.requestresponse.Result
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedEventProcessorResult
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Write.Result
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.WriteBatches.Result
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.csv.CsvIOParseResult
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.gcp.healthcare.DicomIO.ReadStudyMetadata.Result
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.ExecuteBundlesResult
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Read.Result
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Search.Result
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write.AbstractResult
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write.Result
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIOPatientEverything.Result
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.Write.Result
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteResult
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteResult
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.jms.WriteJmsResult
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.solace.write.SolaceOutput
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.WriteFilesResult
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.transforms.JsonToRow.ParseResult
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.transforms.WithFailures.Result
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PCollection
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PCollectionList
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PCollectionRowTuple
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
-  
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PDone
- 
Does nothing; there is nothing to finish specifying. 
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in interface org.apache.beam.sdk.values.POutput
- 
As part of applying the producing  PTransform, finalizes this output to make it ready
 for being used as an input and for running. 
- finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PValueBase
-  
- finishWrite() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
- 
- fireEligibleTimers(InMemoryTimerInternals, Map<KV<String, String>, FnDataReceiver<Timer>>, Object) - Static method in class org.apache.beam.runners.fnexecution.translation.PipelineTranslatorUtils
- 
Fires all timers which are ready to be fired. 
- FirestoreIO - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreOptions - Interface in org.apache.beam.sdk.io.gcp.firestore
-  
- FirestoreV1 - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.BatchGetDocuments - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.BatchGetDocuments.Builder - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.BatchWriteWithDeadLetterQueue - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.BatchWriteWithDeadLetterQueue.Builder - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.BatchWriteWithSummary - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.BatchWriteWithSummary.Builder - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.FailedWritesException - Exception in org.apache.beam.sdk.io.gcp.firestore
- 
Exception that is thrown if one or more Writes is unsuccessful
 with a non-retryable status code.
 
- FirestoreV1.ListCollectionIds - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.ListCollectionIds.Builder - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.ListDocuments - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.ListDocuments.Builder - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.PartitionQuery - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.PartitionQuery.Builder - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.Read - Class in org.apache.beam.sdk.io.gcp.firestore
- 
Type safe builder factory for read operations. 
- FirestoreV1.RunQuery - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.RunQuery.Builder - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- FirestoreV1.Write - Class in org.apache.beam.sdk.io.gcp.firestore
- 
Type safe builder factory for write operations. 
- FirestoreV1.WriteFailure - Class in org.apache.beam.sdk.io.gcp.firestore
- 
Failure details for an attempted Write.
 
- FirestoreV1.WriteSuccessSummary - Class in org.apache.beam.sdk.io.gcp.firestore
- 
Summary object produced when a number of writes are successfully written to Firestore in a
 single BatchWrite. 
- fireTimestamp() - Method in class org.apache.beam.sdk.transforms.DoFn.OnTimerContext
- 
Returns the firing timestamp of the current timer. 
- first - Variable in class org.apache.beam.sdk.transforms.PeriodicSequence.SequenceDefinition
-  
- fixDefaults() - Method in class org.apache.beam.sdk.values.WindowingStrategy
- 
Fixes all the defaults so that equals can be used to check that two strategies are the same,
 regardless of the state of "defaulted-ness". 
- FIXED_WINDOW_TVF - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.TVFStreamingUtils
-  
- FixedBytes - Class in org.apache.beam.sdk.schemas.logicaltypes
- 
A LogicalType representing a fixed-length byte array. 
- FixedPrecisionNumeric - Class in org.apache.beam.sdk.schemas.logicaltypes
- 
Fixed precision numeric types used to represent jdbc NUMERIC and DECIMAL types. 
- fixedSizeGlobally(int) - Static method in class org.apache.beam.sdk.transforms.Sample
- 
Returns a PTransformthat takes aPCollection<T>, selectssampleSizeelements, uniformly at random, and returns aPCollection<Iterable<T>>containing the
 selected elements.
 
- fixedSizePerKey(int) - Static method in class org.apache.beam.sdk.transforms.Sample
- 
Returns a PTransformthat takes an inputPCollection<KV<K, V>>and returns aPCollection<KV<K, Iterable<V>>>that contains an output element mapping each distinct
 key in the inputPCollectionto a sample ofsampleSizevalues associated with
 that key in the inputPCollection, taken uniformly at random.
 
- fixedString(int) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- FixedString - Class in org.apache.beam.sdk.schemas.logicaltypes
- 
A LogicalType representing a fixed-length string. 
- fixedStringSize() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- FixedWindows - Class in org.apache.beam.sdk.transforms.windowing
- 
A  WindowFn that windows values into fixed-size timestamp-based windows. 
- flatMap(KV<K, Iterable<WindowedValue<V>>>, RecordCollector<WindowedValue<KV<K, Iterable<V>>>>) - Method in class org.apache.beam.runners.twister2.translators.functions.GroupByWindowFunction
-  
- FlatMapElements<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
- 
PTransforms for mapping a simple function that returns iterables over the elements of a
  PCollection and merging the results. 
- FlatMapElements.FlatMapWithFailures<InputT,OutputT,FailureT> - Class in org.apache.beam.sdk.transforms
- 
- Flatten - Class in org.apache.beam.sdk.transforms
- 
Flatten<T>takes multiplePCollection<T>s bundled into aPCollectionList<T>and returns a singlePCollection<T>containing all the elements in
 all the inputPCollections.
 
- Flatten() - Constructor for class org.apache.beam.sdk.transforms.Flatten
-  
- Flatten.Iterables<T> - Class in org.apache.beam.sdk.transforms
- 
FlattenIterables<T>takes aPCollection<Iterable<T>>and returns aPCollection<T>that contains all the elements from each iterable.
 
- Flatten.PCollections<T> - Class in org.apache.beam.sdk.transforms
- 
- Flattened() - Constructor for class org.apache.beam.sdk.schemas.transforms.Select.Flattened
-  
- flattenedSchema() - Static method in class org.apache.beam.sdk.schemas.transforms.Select
- 
Selects every leaf-level field. 
- FlattenP - Class in org.apache.beam.runners.jet.processors
- 
Jet Processorimplementation for Beam's Flatten primitive.
 
- FlattenP.Supplier - Class in org.apache.beam.runners.jet.processors
- 
Jet  Processor supplier that will provide instances of  FlattenP. 
- flattenRel(RelStructuredTypeFlattener) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
-  
- FlattenTransformProvider - Class in org.apache.beam.sdk.schemas.transforms.providers
- 
- FlattenTransformProvider() - Constructor for class org.apache.beam.sdk.schemas.transforms.providers.FlattenTransformProvider
-  
- FlattenTransformProvider.Configuration - Class in org.apache.beam.sdk.schemas.transforms.providers
-  
- FlattenTransformProvider.Configuration.Builder - Class in org.apache.beam.sdk.schemas.transforms.providers
-  
- FlattenTranslatorBatch<T> - Class in org.apache.beam.runners.twister2.translators.batch
- 
Flatten translator. 
- FlattenTranslatorBatch() - Constructor for class org.apache.beam.runners.twister2.translators.batch.FlattenTranslatorBatch
-  
- FlattenWithHeterogeneousCoders - Interface in org.apache.beam.sdk.testing
- 
- FlinkBatchPortablePipelineTranslator - Class in org.apache.beam.runners.flink
- 
A translator that translates bounded portable pipelines into executable Flink pipelines. 
- FlinkBatchPortablePipelineTranslator(Map<String, FlinkBatchPortablePipelineTranslator.PTransformTranslator>) - Constructor for class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator
-  
- FlinkBatchPortablePipelineTranslator.BatchTranslationContext - Class in org.apache.beam.runners.flink
- 
Batch translation context. 
- FlinkBatchPortablePipelineTranslator.IsFlinkNativeTransform - Class in org.apache.beam.runners.flink
- 
Predicate to determine whether a URN is a Flink native transform. 
- FlinkBatchPortablePipelineTranslator.PTransformTranslator - Interface in org.apache.beam.runners.flink
- 
Transform translation interface. 
- FlinkDetachedRunnerResult - Class in org.apache.beam.runners.flink
- 
Result of a detached execution of a  Pipeline with Flink. 
- FlinkExecutionEnvironments - Class in org.apache.beam.runners.flink
- 
Utilities for Flink execution environments. 
- FlinkExecutionEnvironments() - Constructor for class org.apache.beam.runners.flink.FlinkExecutionEnvironments
-  
- FlinkJobInvoker - Class in org.apache.beam.runners.flink
- 
- FlinkJobInvoker(FlinkJobServerDriver.FlinkServerConfiguration) - Constructor for class org.apache.beam.runners.flink.FlinkJobInvoker
-  
- FlinkJobServerDriver - Class in org.apache.beam.runners.flink
- 
Driver program that starts a job server for the Flink runner. 
- FlinkJobServerDriver.FlinkServerConfiguration - Class in org.apache.beam.runners.flink
- 
Flink runner-specific Configuration for the jobServer. 
- FlinkKey - Class in org.apache.beam.runners.flink.adapter
-  
- FlinkKey() - Constructor for class org.apache.beam.runners.flink.adapter.FlinkKey
-  
- FlinkMetricContainer - Class in org.apache.beam.runners.flink.metrics
- 
Helper class for holding a MetricsContainerImpland forwarding Beam metrics to Flink
 accumulators and metrics.
 
- FlinkMetricContainer(RuntimeContext) - Constructor for class org.apache.beam.runners.flink.metrics.FlinkMetricContainer
-  
- FlinkMetricContainerWithoutAccumulator - Class in org.apache.beam.runners.flink.metrics
- 
The base helper class for holding a MetricsContainerImpland forwarding
 Beam metrics to Flink accumulators and metrics.
 
- FlinkMetricContainerWithoutAccumulator(MetricGroup) - Constructor for class org.apache.beam.runners.flink.metrics.FlinkMetricContainerWithoutAccumulator
-  
- FlinkMiniClusterEntryPoint - Class in org.apache.beam.runners.flink
- 
Entry point for starting an embedded Flink cluster. 
- FlinkMiniClusterEntryPoint() - Constructor for class org.apache.beam.runners.flink.FlinkMiniClusterEntryPoint
-  
- FlinkPipelineOptions - Interface in org.apache.beam.runners.flink
- 
Options which can be used to configure the Flink Runner. 
- FlinkPipelineOptions.MaxBundleSizeFactory - Class in org.apache.beam.runners.flink
- 
Maximum bundle size factory. 
- FlinkPipelineOptions.MaxBundleTimeFactory - Class in org.apache.beam.runners.flink
- 
Maximum bundle time factory. 
- FlinkPipelineRunner - Class in org.apache.beam.runners.flink
- 
- FlinkPipelineRunner(FlinkPipelineOptions, String, List<String>) - Constructor for class org.apache.beam.runners.flink.FlinkPipelineRunner
- 
Setup a flink pipeline runner. 
- FlinkPortableClientEntryPoint - Class in org.apache.beam.runners.flink
- 
Flink job entry point to launch a Beam pipeline by executing an external SDK driver program. 
- FlinkPortableClientEntryPoint(String) - Constructor for class org.apache.beam.runners.flink.FlinkPortableClientEntryPoint
-  
- FlinkPortablePipelineTranslator<T extends FlinkPortablePipelineTranslator.TranslationContext> - Interface in org.apache.beam.runners.flink
- 
Interface for portable Flink translators. 
- FlinkPortablePipelineTranslator.Executor - Interface in org.apache.beam.runners.flink
- 
A handle used to execute a translated pipeline. 
- FlinkPortablePipelineTranslator.TranslationContext - Interface in org.apache.beam.runners.flink
- 
The context used for pipeline translation. 
- FlinkPortableRunnerResult - Class in org.apache.beam.runners.flink
- 
Result of executing a portable  Pipeline with Flink. 
- FlinkRunner - Class in org.apache.beam.runners.flink
- 
A  PipelineRunner that executes the operations in the pipeline by first translating them
 to a Flink Plan and then executing them either locally or on a Flink cluster, depending on the
 configuration. 
- FlinkRunner(FlinkPipelineOptions) - Constructor for class org.apache.beam.runners.flink.FlinkRunner
-  
- FlinkRunnerRegistrar - Class in org.apache.beam.runners.flink
- 
AutoService registrar - will register FlinkRunner and FlinkOptions as possible pipeline runner
 services. 
- FlinkRunnerRegistrar.Options - Class in org.apache.beam.runners.flink
- 
Pipeline options registrar. 
- FlinkRunnerRegistrar.Runner - Class in org.apache.beam.runners.flink
- 
Pipeline runner registrar. 
- FlinkRunnerResult - Class in org.apache.beam.runners.flink
- 
Result of executing a  Pipeline with Flink. 
- FlinkServerConfiguration() - Constructor for class org.apache.beam.runners.flink.FlinkJobServerDriver.FlinkServerConfiguration
-  
- FlinkStateBackendFactory - Interface in org.apache.beam.runners.flink
- 
Constructs a StateBackend to use from flink pipeline options. 
- FlinkStreamingAggregationsTranslators - Class in org.apache.beam.runners.flink
-  
- FlinkStreamingAggregationsTranslators() - Constructor for class org.apache.beam.runners.flink.FlinkStreamingAggregationsTranslators
-  
- FlinkStreamingAggregationsTranslators.ConcatenateAsIterable<T> - Class in org.apache.beam.runners.flink
-  
- FlinkStreamingPortablePipelineTranslator - Class in org.apache.beam.runners.flink
- 
Translate an unbounded portable pipeline representation into a Flink pipeline representation. 
- FlinkStreamingPortablePipelineTranslator() - Constructor for class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator
-  
- FlinkStreamingPortablePipelineTranslator(Map<String, FlinkStreamingPortablePipelineTranslator.PTransformTranslator<FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext>>) - Constructor for class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator
-  
- FlinkStreamingPortablePipelineTranslator.IsFlinkNativeTransform - Class in org.apache.beam.runners.flink
- 
Predicate to determine whether a URN is a Flink native transform. 
- FlinkStreamingPortablePipelineTranslator.PTransformTranslator<T> - Interface in org.apache.beam.runners.flink
-  
- FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext - Class in org.apache.beam.runners.flink
- 
Streaming translation context. 
- FLOAT - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- FLOAT - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
The type of float fields. 
- FLOAT32 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- FLOAT64 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- FloatCoder - Class in org.apache.beam.sdk.coders
- 
A  FloatCoder encodes  Float values in 4 bytes using Java serialization. 
- floats() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
- 
- floatToByteArray(float) - Static method in class org.apache.beam.sdk.io.gcp.testing.BigtableUtils
-  
- flush(boolean) - Method in class org.apache.beam.runners.jet.metrics.JetMetricsContainer
-  
- flush() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Sink
-  
- flush() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
-  
- flush() - Method in interface org.apache.beam.sdk.fn.data.CloseableFnDataReceiver
- 
- flush() - Method in interface org.apache.beam.sdk.io.FileIO.Sink
- 
Flushes the buffered state (if any) before the channel is closed. 
- flush(String, long) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.WriteStreamService
- 
Flush a given stream up to the given offset. 
- flush(String, long) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.WriteStreamServiceImpl
-  
- flush(String, long) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- flush() - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Sink
-  
- flush() - Method in class org.apache.beam.sdk.io.TextIO.Sink
-  
- flush() - Method in class org.apache.beam.sdk.io.TFRecordIO.Sink
-  
- flush() - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.Sink
-  
- flush() - Method in class org.apache.beam.sdk.io.xml.XmlIO.Sink
-  
- flushBundle(DoFn<KV<Integer, Solace.Record>, Solace.PublishResult>.OnTimerContext) - Method in class org.apache.beam.sdk.io.solace.write.UnboundedBatchedSolaceWriter
-  
- fn(ProcessFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Contextful
- 
- fn(SerializableFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Contextful
- 
- fn(Contextful.Fn<InputT, OutputT>, Requirements) - Static method in class org.apache.beam.sdk.transforms.Contextful
- 
- FnApiControlClient - Class in org.apache.beam.runners.fnexecution.control
- 
A client for the control plane of an SDK harness, which can issue requests to it over the Fn API. 
- FnApiControlClientPoolService - Class in org.apache.beam.runners.fnexecution.control
- 
A Fn API control service which adds incoming SDK harness connections to a sink. 
- FnDataReceiver<T> - Interface in org.apache.beam.sdk.fn.data
- 
A receiver of streamed data. 
- FnDataService - Interface in org.apache.beam.runners.fnexecution.data
- 
The  FnDataService is able to forward inbound elements to a consumer and is also a
 consumer of outbound elements. 
- FnService - Interface in org.apache.beam.sdk.fn.server
- 
An interface sharing common behavior with services used during execution of user Fns. 
- forBagUserStateHandlerFactory(ProcessBundleDescriptors.ExecutableProcessBundleDescriptor, StateRequestHandlers.BagUserStateHandlerFactory) - Static method in class org.apache.beam.runners.fnexecution.state.StateRequestHandlers
- 
- forBatch(PCollectionView<ViewT>) - Static method in class org.apache.beam.runners.dataflow.CreateDataflowView
-  
- forBoolean(Boolean) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
- 
Constructs a CloudObjectrepresenting the given value.
 
- forBytes() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Init
- 
Returns a  HllCount.Init.Builder for a  HllCount.Init combining  PTransform that
 computes bytes-type HLL++ sketches. 
- forClass(Class<?>) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
- 
Constructs a CloudObjectto be used for serializing an instance of the supplied class
 for transport via the Dataflow API.
 
- forClassName(String) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
- 
Constructs a CloudObjectto be used for serializing data to be deserialized using the
 supplied class name the supplied class name for transport via the Dataflow API.
 
- forCoder(TypeDescriptor<?>, Coder<?>) - Static method in class org.apache.beam.sdk.coders.CoderProviders
- 
Creates a  CoderProvider that always returns the given coder for the specified type. 
- forConsumers(List<DataEndpoint<?>>, List<TimerEndpoint<?>>) - Static method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
- 
Creates a receiver that is able to consume elements multiplexing on to the provided set of
 endpoints. 
- forDescriptor(ProtoDomain, String) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoDynamicMessageSchema
- 
Create a new ProtoDynamicMessageSchema from a  ProtoDomain and for a message. 
- forDescriptor(ProtoDomain, Descriptors.Descriptor) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoDynamicMessageSchema
- 
Create a new ProtoDynamicMessageSchema from a  ProtoDomain and for a descriptor. 
- forDescriptor(Endpoints.ApiServiceDescriptor) - Method in class org.apache.beam.sdk.fn.channel.ManagedChannelFactory
-  
- forEncoding(ByteString) - Static method in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow
-  
- forever(Trigger) - Static method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
- 
Create a composite trigger that repeatedly executes the trigger repeated, firing each
 time it fires and ignoring any indications to finish.
 
- forField(TypeDescriptor<?>, Field, int) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- forFloat(Float) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
- 
Constructs a CloudObjectrepresenting the given value.
 
- forFloat(Double) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
- 
Constructs a CloudObjectrepresenting the given value.
 
- forGetter(Method, int) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- forGetter(TypeDescriptor<?>, Method, int) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- forHandler(RunnerApi.Environment, InstructionRequestHandler) - Static method in interface org.apache.beam.runners.fnexecution.environment.RemoteEnvironment
- 
- forInteger(Long) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
- 
Constructs a CloudObjectrepresenting the given value.
 
- forInteger(Integer) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
- 
Constructs a CloudObjectrepresenting the given value.
 
- forIntegers() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Init
- 
Returns a  HllCount.Init.Builder for a  HllCount.Init combining  PTransform that
 computes integer-type HLL++ sketches. 
- forIterableSideInput(String, String, Coder<V>, Coder<W>) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.SideInputHandlerFactory
- 
- forIterableSideInput(String, String, Coder<V>, Coder<W>) - Method in class org.apache.beam.runners.fnexecution.translation.BatchSideInputHandlerFactory
-  
- forIterableSideInput(String, String, Coder<V>, Coder<W>) - Method in class org.apache.beam.runners.fnexecution.translation.StreamingSideInputHandlerFactory
-  
- forKey(K) - Static method in class org.apache.beam.runners.spark.stateful.SparkStateInternals
-  
- forKeyAndState(K, Table<String, String, byte[]>) - Static method in class org.apache.beam.runners.spark.stateful.SparkStateInternals
-  
- forKnownType(Object) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
- 
Constructs a CloudObjectrepresenting the given value of a well-known cloud object
 type.
 
- forLongs() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Init
- 
Returns a  HllCount.Init.Builder for a  HllCount.Init combining  PTransform that
 computes long-type HLL++ sketches. 
- FORMAT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- FormatAsTextFn() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.FormatAsTextFn
-  
- formatByteStringRange(Range.ByteStringRange) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ByteStringRangeHelper
- 
Returns formatted string of a partition for debugging. 
- formatRecord(ElementT, Schema) - Method in interface org.apache.beam.sdk.extensions.avro.io.AvroIO.RecordFormatter
- 
Deprecated. 
- formatRecord(UserT) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
- 
Convert an input record type into the output type. 
- formatTimestamp(Instant) - Static method in class org.apache.beam.sdk.transforms.windowing.BoundedWindow
- 
- formatTimestampWithTimeZone(DateTime) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
-  
- forMultimapSideInput(String, String, KvCoder<K, V>, Coder<W>) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.SideInputHandlerFactory
- 
- forMultimapSideInput(String, String, KvCoder<K, V>, Coder<W>) - Method in class org.apache.beam.runners.fnexecution.translation.BatchSideInputHandlerFactory
-  
- forMultimapSideInput(String, String, KvCoder<K, V>, Coder<W>) - Method in class org.apache.beam.runners.fnexecution.translation.StreamingSideInputHandlerFactory
-  
- forNewInput(Instant, InputT) - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
- 
Called by the  Watch transform to create a new independent termination state for a
 newly arrived  InputT. 
- forOneOf(String, boolean, Map<String, FieldValueTypeInformation>) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- forOrdinal(int) - Static method in enum org.apache.beam.sdk.io.kafka.KafkaTimestampType
-  
- forProject(String, int, String) - Static method in class org.apache.beam.sdk.transformservice.launcher.TransformServiceLauncher
- 
Initializes a client for managing transform service instances. 
- forRegistry(MetricRegistry) - Static method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
-  
- forRegistry(MetricRegistry) - Static method in class org.apache.beam.runners.spark.structuredstreaming.metrics.WithMetricsSupport
-  
- forRequestObserver(String, StreamObserver<BeamFnApi.InstructionRequest>, ConcurrentMap<String, BeamFnApi.ProcessBundleDescriptor>) - Static method in class org.apache.beam.runners.fnexecution.control.FnApiControlClient
- 
- forService(InstructionRequestHandler) - Static method in class org.apache.beam.runners.fnexecution.environment.StaticRemoteEnvironmentFactory
-  
- forSetter(Method) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- forSetter(Method, String) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- forSetter(TypeDescriptor<?>, Method) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- forSetter(TypeDescriptor<?>, Method, String) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- forSideInputHandlerFactory(Map<String, Map<String, ProcessBundleDescriptors.SideInputSpec>>, StateRequestHandlers.SideInputHandlerFactory) - Static method in class org.apache.beam.runners.fnexecution.state.StateRequestHandlers
- 
- forSqlType(Schema.TypeName) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.BigDecimalConverter
-  
- forStage(ExecutableStage) - Method in class org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory
-  
- forStage(ExecutableStage) - Method in interface org.apache.beam.runners.fnexecution.control.JobBundleFactory
-  
- forStage(ExecutableStage) - Method in class org.apache.beam.runners.fnexecution.control.SingleEnvironmentInstanceJobBundleFactory
- 
Deprecated. 
- forStage(ExecutableStage, BatchSideInputHandlerFactory.SideInputGetter) - Static method in class org.apache.beam.runners.fnexecution.translation.BatchSideInputHandlerFactory
- 
Creates a new state handler for the given stage. 
- forStage(ExecutableStage, Map<RunnerApi.ExecutableStagePayload.SideInputId, PCollectionView<?>>, SideInputHandler) - Static method in class org.apache.beam.runners.fnexecution.translation.StreamingSideInputHandlerFactory
- 
Creates a new state handler for the given stage. 
- forStreamFromSources(List<Integer>, Map<Integer, GlobalWatermarkHolder.SparkWatermarks>) - Static method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
- 
Build the TimerInternalsaccording to the feeding streams.
 
- forStreaming(PCollectionView<ViewT>) - Static method in class org.apache.beam.runners.dataflow.CreateDataflowView
-  
- forString(String) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
- 
Constructs a CloudObjectrepresenting the given value.
 
- forStrings() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Init
- 
Returns a  HllCount.Init.Builder for a  HllCount.Init combining  PTransform that
 computes string-type HLL++ sketches. 
- forThrowable(Throwable) - Static method in class org.apache.beam.sdk.values.EncodableThrowable
- 
Wraps throwableand returns the result.
 
- forTransformHierarchy(TransformHierarchy, PipelineOptions) - Static method in class org.apache.beam.sdk.Pipeline
-  
- forTypeName(Schema.TypeName) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
-  
- forUserState(String, String, Coder<K>, Coder<V>, Coder<W>) - Method in class org.apache.beam.runners.fnexecution.state.InMemoryBagUserStateFactory
-  
- forUserState(String, String, Coder<K>, Coder<V>, Coder<W>) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.BagUserStateHandlerFactory
-  
- ForwardingClientResponseObserver<ReqT,RespT> - Class in org.apache.beam.sdk.fn.stream
- 
A ClientResponseObserverwhich delegates allStreamObservercalls.
 
- forWriter(LogWriter) - Static method in class org.apache.beam.runners.fnexecution.logging.GrpcLoggingService
-  
- freeze() - Method in class org.apache.beam.runners.jet.metrics.JetMetricResults
-  
- from(DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration) - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider
-  
- from(Map<String, String>) - Static method in class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
- 
Deprecated. Expects a map keyed by logger Names with values representingLevels.
 
- from(WindowIntoTransformProvider.Configuration) - Method in class org.apache.beam.sdk.expansion.service.WindowIntoTransformProvider
-  
- from(String) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Parse
- 
Reads from the given filename or filepattern. 
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Parse
- 
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Read
- 
Reads from the given filename or filepattern. 
- from(String) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Read
- 
- from(String, Row, Schema) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSchemaIOProvider
- 
Produce a SchemaIO given a String representing the data's location, the schema of the data that
 resides there, and some IO-specific configuration object. 
- from(ValueProvider<String>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
- 
Reads from the given file name or pattern ("glob"). 
- from(MatchResult.Metadata) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
-  
- from(String) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
- 
- from(String) - Static method in class org.apache.beam.sdk.extensions.python.PythonExternalTransform
- 
Instantiates a cross-language wrapper for a Python transform with a given transform name. 
- from(String, String) - Static method in class org.apache.beam.sdk.extensions.python.PythonExternalTransform
- 
Instantiates a cross-language wrapper for a Python transform with a given transform name. 
- from(Row) - Method in class org.apache.beam.sdk.extensions.sql.expansion.SqlTransformSchemaTransformProvider
-  
- from(ExecutorService) - Static method in class org.apache.beam.sdk.fn.test.TestExecutors
-  
- from(Supplier<ExecutorService>) - Static method in class org.apache.beam.sdk.fn.test.TestExecutors
-  
- from(FileBasedSource<T>) - Static method in class org.apache.beam.sdk.io.CompressedSource
- 
Creates a CompressedSourcefrom an underlyingFileBasedSource.
 
- from(String) - Method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.Read
- 
Reads text from the file(s) with the given filename or filename pattern. 
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.Read
- 
- from(CsvWriteTransformProvider.CsvWriteConfiguration) - Method in class org.apache.beam.sdk.io.csv.providers.CsvWriteTransformProvider
-  
- from(FileReadSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformProvider
-  
- from(FileWriteSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformProvider
- 
- from(BigQueryExportReadSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryExportReadSchemaTransformProvider
- 
- from(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
- 
Reads a BigQuery table specified as "[project_id]:[dataset_id].[table_id]"or"[dataset_id].[table_id]"for tables within the current project.
 
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
- 
- from(TableReference) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
- 
- from(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- from(TableReference) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- from(String, Row, Schema) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySchemaIOProvider
- 
Produces a SchemaIO given a String representing the data's location, the schema of the data
 that resides there, and some IO-specific configuration object. 
- from(BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider
-  
- from(BigQueryWriteConfiguration) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryFileLoadsSchemaTransformProvider
-  
- from(BigQueryWriteConfiguration) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProvider
-  
- from(BigQueryWriteConfiguration) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteSchemaTransformProvider
-  
- from(BigtableReadSchemaTransformProvider.BigtableReadSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProvider
-  
- from(BigtableWriteSchemaTransformProvider.BigtableWriteSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProvider
-  
- from(String, Row, Schema) - Method in class org.apache.beam.sdk.io.gcp.datastore.DataStoreV1SchemaIOProvider
- 
Produce a SchemaIO given a String representing the data's location, the schema of the data that
 resides there, and some IO-specific configuration object. 
- from(PubsubReadSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformProvider
-  
- from(String, Row, Schema) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubSchemaIOProvider
- 
Produce a SchemaIO given a String representing the data's location, the schema of the data that
 resides there, and some IO-specific configuration object. 
- from(PubsubWriteSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformProvider
-  
- from(PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider
-  
- from(PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider
-  
- from(Struct) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.mapper.PartitionMetadataMapper
- 
Transforms a  Struct representing a partition metadata row into a  PartitionMetadata model. 
- from(SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider
-  
- from(SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider
-  
- from(SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider
-  
- from(long) - Static method in class org.apache.beam.sdk.io.GenerateSequence
- 
Specifies the minimum number to generate (inclusive). 
- from(TableIdentifier) - Method in class org.apache.beam.sdk.io.iceberg.IcebergIO.ReadRows
-  
- from(SchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.iceberg.IcebergReadSchemaTransformProvider
-  
- from(IcebergWriteSchemaTransformProvider.Configuration) - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider
-  
- from(JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider
-  
- from(String, Row, Schema) - Method in class org.apache.beam.sdk.io.jdbc.JdbcSchemaIOProvider
- 
Produce a SchemaIO given a String representing the data's location, the schema of the data that
 resides there, and some IO-specific configuration object. 
- from(JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider
-  
- from(JsonWriteTransformProvider.JsonWriteConfiguration) - Method in class org.apache.beam.sdk.io.json.providers.JsonWriteTransformProvider
-  
- from(KafkaReadSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformProvider
-  
- from(KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration) - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider
-  
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Parse
-  
- from(String) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Parse
-  
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Read
- 
Reads from the given filename or filepattern. 
- from(String) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Read
- 
- from(BoundedSource<T>) - Method in class org.apache.beam.sdk.io.Read.Builder
- 
Returns a new Read.BoundedPTransformreading from the givenBoundedSource.
 
- from(UnboundedSource<T, ?>) - Method in class org.apache.beam.sdk.io.Read.Builder
- 
Returns a new Read.UnboundedPTransformreading from the givenUnboundedSource.
 
- from(BoundedSource<T>) - Static method in class org.apache.beam.sdk.io.Read
- 
Returns a new Read.BoundedPTransformreading from the givenBoundedSource.
 
- from(UnboundedSource<T, ?>) - Static method in class org.apache.beam.sdk.io.Read
- 
- from(SingleStoreSchemaTransformReadConfiguration) - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadProvider
- 
- from(SingleStoreSchemaTransformWriteConfiguration) - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteProvider
- 
- from(Solace.Queue) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Read
- 
Set the queue name to read from. 
- from(Solace.Topic) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Read
- 
Set the topic name to read from. 
- from(String) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Read
- 
Provide name of collection while reading from Solr. 
- from(String) - Method in class org.apache.beam.sdk.io.TextIO.Read
- 
Reads text files that reads from the file(s) with the given filename or filename pattern. 
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.TextIO.Read
- 
- from(String) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
- 
Returns a transform for reading TFRecord files that reads from the file(s) with the given
 filename or filename pattern. 
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
- 
- from(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
- 
Reads a single XML file or a set of XML files defined by a Java "glob" file pattern. 
- from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
- 
Reads a single XML file or a set of XML files defined by a Java "glob" file pattern. 
- from(ManagedSchemaTransformProvider.ManagedConfig) - Method in class org.apache.beam.sdk.managed.ManagedSchemaTransformProvider
-  
- from(TestSchemaTransformProvider.Config) - Method in class org.apache.beam.sdk.managed.testing.TestSchemaTransformProvider
-  
- from(Map<String, String>) - Static method in class org.apache.beam.sdk.options.SdkHarnessOptions.SdkHarnessLogLevelOverrides
- 
Expects a map keyed by logger Names with values representingLogLevels.
 
- from(GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration) - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider
-  
- from(String, Row, Schema) - Method in interface org.apache.beam.sdk.schemas.io.SchemaIOProvider
- 
Produce a SchemaIO given a String representing the data's location, the schema of the data that
 resides there, and some IO-specific configuration object. 
- from(FlattenTransformProvider.Configuration) - Method in class org.apache.beam.sdk.schemas.transforms.providers.FlattenTransformProvider
-  
- from(JavaExplodeTransformProvider.Configuration) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaExplodeTransformProvider
-  
- from(JavaFilterTransformProvider.Configuration) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaFilterTransformProvider
-  
- from(JavaMapToFieldsTransformProvider.Configuration) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider
-  
- from(LoggingTransformProvider.Configuration) - Method in class org.apache.beam.sdk.schemas.transforms.providers.LoggingTransformProvider
-  
- from(Row) - Method in interface org.apache.beam.sdk.schemas.transforms.SchemaTransformProvider
- 
Produce a  SchemaTransform from some transform-specific configuration object. 
- from(ConfigT) - Method in class org.apache.beam.sdk.schemas.transforms.TypedSchemaTransformProvider
- 
Produce a SchemaTransform from ConfigT. 
- from(Row) - Method in class org.apache.beam.sdk.schemas.transforms.TypedSchemaTransformProvider
- 
- from(HasDisplayData) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
- 
- from(double, double) - Static method in class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker.Progress
- 
A representation for the amount of known completed and remaining work. 
- fromArgs(String...) - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
- 
Sets the command line arguments to parse when constructing the  PipelineOptions. 
- fromArgs(String...) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
- 
Sets the command line arguments to parse when constructing the  PipelineOptions. 
- fromArray(T...) - Static method in class org.apache.beam.sdk.fn.stream.PrefetchableIterables
- 
- fromArray(T...) - Static method in class org.apache.beam.sdk.fn.stream.PrefetchableIterators
- 
- fromAvroType(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.FixedBytesField
- 
- fromBeamFieldType(Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.FixedBytesField
- 
- fromByteArray(byte[], Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
- 
Utility method for deserializing a byte array using the specified coder. 
- fromByteArray(byte[], Coder<T>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.CoderHelpers
- 
Utility method for deserializing a byte array using the specified coder. 
- fromByteArray(byte[], Coder<T>) - Static method in class org.apache.beam.runners.twister2.utils.TranslationUtils
- 
Utility method for deserializing a byte array using the specified coder. 
- fromByteArray(byte[], WindowedValue.WindowedValueCoder<T>) - Static method in class org.apache.beam.runners.twister2.utils.TranslationUtils
- 
Utility method for deserializing a byte array using the specified coder. 
- fromByteArrays(Collection<byte[]>, Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
- 
Utility method for deserializing a Iterable of byte arrays using the specified coder. 
- fromByteFunction(Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
- 
A function wrapper for converting a byte array to an object. 
- FromByteFunction(Coder<K>, Coder<V>) - Constructor for class org.apache.beam.runners.spark.coders.CoderHelpers.FromByteFunction
-  
- fromByteFunctionIterable(Coder<K>, Coder<V>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
- 
A function wrapper for converting a byte array pair to a key-value pair, where values are
 Iterable.
 
- fromCanonical(Compression) - Static method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
- 
Deprecated. 
- fromCloudDuration(String) - Static method in class org.apache.beam.runners.dataflow.util.TimeUtil
- 
Converts a Dataflow API duration string into a  Duration. 
- fromCloudObject(CloudObject) - Method in interface org.apache.beam.runners.dataflow.util.CloudObjectTranslator
- 
Converts back into the original object from a provided  CloudObject. 
- fromCloudObject(CloudObject) - Method in class org.apache.beam.runners.dataflow.util.RowCoderCloudObjectTranslator
- 
Convert from a cloud object. 
- fromCloudObject(CloudObject) - Method in class org.apache.beam.runners.dataflow.util.SchemaCoderCloudObjectTranslator
- 
Convert from a cloud object. 
- fromCloudPubsubMessages() - Static method in class org.apache.beam.sdk.io.gcp.pubsublite.CloudPubsubTransforms
- 
Transform messages publishable using PubsubIO to their equivalent Pub/Sub Lite publishable
 message. 
- fromCloudTime(String) - Static method in class org.apache.beam.runners.dataflow.util.TimeUtil
- 
Converts a time value received via the Dataflow API into the corresponding  Instant. 
- fromComponents(List<Coder<?>>, byte[], CoderTranslation.TranslationContext) - Method in class org.apache.beam.sdk.extensions.avro.AvroGenericCoderTranslator
-  
- fromComponents(String, String) - Static method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
- 
Creates a GcsPath from bucket and object components. 
- fromConfig(FlinkJobServerDriver.FlinkServerConfiguration) - Static method in class org.apache.beam.runners.flink.FlinkJobServerDriver
-  
- fromConfig(FlinkJobServerDriver.FlinkServerConfiguration, JobServerDriver.JobInvokerFactory) - Static method in class org.apache.beam.runners.flink.FlinkJobServerDriver
-  
- fromConfig(SparkJobServerDriver.SparkServerConfiguration) - Static method in class org.apache.beam.runners.spark.SparkJobServerDriver
-  
- fromConfigRow(Row, PipelineOptions) - Method in class org.apache.beam.sdk.schemas.transforms.SchemaTransformTranslation.SchemaTransformPayloadTranslator
-  
- fromExceptionInformation(RecordT, Coder<RecordT>, Exception, String) - Static method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord
-  
- fromExecutableStage(String, ExecutableStage, Endpoints.ApiServiceDescriptor, Endpoints.ApiServiceDescriptor) - Static method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors
- 
Note that the BeamFnApi.ProcessBundleDescriptoris constructed by:
 
   Adding gRPC read and write nodes wiring them to the specified data endpoint.
 
- fromExecutableStage(String, ExecutableStage, Endpoints.ApiServiceDescriptor) - Static method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors
-  
- fromExistingTable(String, String) - Static method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataTableNames
- 
Encapsulates a selected table name. 
- fromFeedRange(FeedRange) - Static method in class org.apache.beam.sdk.io.azure.cosmos.NormalizedRange
-  
- fromFile(String) - Static method in class org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService.ArtifactDestination
-  
- fromFile(String, OutputStream) - Static method in class org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService.ArtifactDestination
-  
- fromFile(File, boolean) - Static method in class org.apache.beam.sdk.io.LocalResources
-  
- fromHex(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
-  
- fromHttpResponse(HttpResponse) - Static method in class org.apache.beam.sdk.io.solace.broker.BrokerResponse
-  
- fromInt(Integer, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.DateConversion
-  
- fromInt(Integer, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimeMillisConversion
-  
- fromInt(Integer, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.DateConversion
-  
- fromInt(Integer, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimeConversion
-  
- fromIr(Ir, SbeSchema.IrOptions) - Static method in class org.apache.beam.sdk.extensions.sbe.SbeSchema
- 
Creates a new  SbeSchema from the given intermediate representation. 
- fromIr(Ir) - Static method in class org.apache.beam.sdk.extensions.sbe.SerializableIr
- 
Creates a new instance from ir.
 
- fromJsonFile(File) - Method in class org.apache.beam.sdk.io.cdap.ConfigWrapper
- 
- fromJsonString(String) - Method in class org.apache.beam.sdk.io.cdap.ConfigWrapper
- 
- fromJsonString(String, Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
-  
- fromLong(Long, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.LocalTimestampMicrosConversion
-  
- fromLong(Long, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.LocalTimestampMillisConversion
-  
- fromLong(Long, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimeMicrosConversion
-  
- fromLong(Long, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimestampMicrosConversion
-  
- fromLong(Long, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimestampMillisConversion
-  
- fromLong(Long, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimeMicrosConversion
-  
- fromLong(Long, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimestampConversion
-  
- fromLong(Long, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimestampMicrosConversion
-  
- fromMap(Map<String, String>) - Static method in class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
- 
Returns a new configuration instance using provided flags. 
- fromModel(Message) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2Message
- 
From model  Message to hl7v2 message. 
- fromName(String) - Static method in enum org.apache.beam.io.debezium.Connectors
- 
Returns a connector class corresponding to the given connector name. 
- fromName(String) - Static method in class org.apache.beam.sdk.io.solace.data.Solace.Queue
-  
- fromName(String) - Static method in class org.apache.beam.sdk.io.solace.data.Solace.Topic
-  
- fromObject(StorageObject) - Static method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
- 
- fromOptions(DataflowPipelineOptions) - Static method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator
- 
Constructs a translator from the provided options. 
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.dataflow.DataflowRunner
- 
Construct a runner from the provided options. 
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.dataflow.TestDataflowRunner
- 
Constructs a runner from the provided options. 
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.dataflow.util.GcsStager
-  
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.direct.DirectRunner
- 
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.flink.FlinkRunner
- 
Construct a runner from the provided options. 
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.flink.TestFlinkRunner
-  
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.jet.JetRunner
-  
- fromOptions(PipelineOptions, Function<ClientConfig, JetInstance>) - Static method in class org.apache.beam.runners.jet.JetRunner
-  
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.portability.PortableRunner
- 
Constructs a runner from the provided options. 
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.portability.testing.TestPortableRunner
-  
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.portability.testing.TestUniversalRunner
- 
Constructs a runner from the provided options. 
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.prism.PrismRunner
- 
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.prism.TestPrismRunner
- 
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.spark.SparkRunner
- 
Creates and returns a new SparkRunner with specified options. 
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.spark.SparkRunnerDebugger
-  
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingRunner
- 
Creates and returns a new SparkStructuredStreamingRunner with specified options. 
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.spark.TestSparkRunner
-  
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.twister2.Twister2Runner
-  
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.twister2.Twister2TestRunner
-  
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.extensions.gcp.auth.GcpCredentialFactory
-  
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.extensions.gcp.auth.NoopCredentialFactory
-  
- fromOptions(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsFileSystemRegistrar
-  
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
-  
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
-  
- fromOptions(PipelineOptions) - Method in class org.apache.beam.sdk.io.aws2.s3.DefaultS3FileSystemSchemeRegistrar
-  
- fromOptions(PipelineOptions) - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemRegistrar
-  
- fromOptions(PipelineOptions) - Method in interface org.apache.beam.sdk.io.aws2.s3.S3FileSystemSchemeRegistrar
- 
- fromOptions(PipelineOptions) - Method in class org.apache.beam.sdk.io.azure.blobstore.AzureBlobStoreFileSystemRegistrar
-  
- fromOptions(PipelineOptions) - Method in class org.apache.beam.sdk.io.ClassLoaderFileSystem.ClassLoaderFileSystemRegistrar
-  
- fromOptions(PipelineOptions) - Method in interface org.apache.beam.sdk.io.FileSystemRegistrar
- 
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
- 
Creates an instance of this rule using provided options. 
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.io.googleads.GoogleAdsUserCredentialFactory
-  
- fromOptions(PipelineOptions) - Method in class org.apache.beam.sdk.io.hdfs.HadoopFileSystemRegistrar
-  
- fromOptions(PipelineOptions) - Method in class org.apache.beam.sdk.io.LocalFileSystemRegistrar
-  
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.PipelineRunner
- 
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.testing.CrashingRunner
-  
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.testing.TestPipeline
-  
- fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.transforms.resourcehints.ResourceHints
- 
- fromParams(String[]) - Static method in class org.apache.beam.runners.flink.FlinkJobServerDriver
-  
- fromParams(String[]) - Static method in class org.apache.beam.runners.spark.SparkJobServerDriver
-  
- fromParams(DefaultFilenamePolicy.Params) - Static method in class org.apache.beam.sdk.io.DefaultFilenamePolicy
- 
- fromPath(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription
- 
Creates a class representing a Pub/Sub subscription from the specified subscription path. 
- fromPath(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic
- 
Creates a class representing a Cloud Pub/Sub topic from the specified topic path. 
- fromPath(Path, boolean) - Static method in class org.apache.beam.sdk.io.LocalResources
-  
- fromProcessFunctionWithOutputType(ProcessFunction<InputT, OutputT>, TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.transforms.InferableFunction
-  
- fromProto(PubsubMessage) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessages
-  
- fromPTransform(RunnerApi.PTransform) - Static method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortRead
-  
- fromPTransform(RunnerApi.PTransform) - Static method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortWrite
-  
- fromQuery(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
- 
Reads results received after executing the given query. 
- fromQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
- 
- fromQuery(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- fromQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- fromQuery(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
- 
A query to be executed in Snowflake. 
- fromQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
-  
- fromRawEvents(Coder<T>, List<TestStream.Event<T>>) - Static method in class org.apache.beam.sdk.testing.TestStream
- 
For internal use only. 
- fromResourceName(String) - Static method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
- 
Creates a GcsPath from a OnePlatform resource name in string form. 
- fromRow(Row) - Static method in class org.apache.beam.sdk.values.Row
- 
Creates a row builder based on the specified row. 
- fromRowFunction(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.io.aws2.schemas.AwsSchemaProvider
-  
- fromRowFunction(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.annotations.DefaultSchema.DefaultSchemaProvider
- 
Given a type, returns a function that converts from a  Row object to that type. 
- fromRowFunction(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
- 
Deprecated. 
- fromRowFunction(TypeDescriptor<T>) - Method in interface org.apache.beam.sdk.schemas.SchemaProvider
- 
Given a type, returns a function that converts from a  Row object to that type. 
- fromRows(Class<OutputT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Convert
- 
- fromRows(TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Convert
- 
- fromS3Options(S3Options) - Static method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration
-  
- fromSerializableFunctionWithOutputType(SerializableFunction<InputT, OutputT>, TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.transforms.SimpleFunction
-  
- fromSnapshot(Snapshot) - Static method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo
-  
- fromSpec(Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
- 
Constructs a CloudObjectby copying the supplied serialized object spec, which must
 represent an SDK object serialized for transport via the Dataflow API.
 
- fromSpec(HCatalogIO.Read) - Static method in class org.apache.beam.sdk.io.hcatalog.HCatToRow
- 
- fromStandardParameters(ValueProvider<ResourceId>, String, String, boolean) - Static method in class org.apache.beam.sdk.io.DefaultFilenamePolicy
- 
- fromStaticMethods(Class<?>, Class<?>) - Static method in class org.apache.beam.sdk.coders.CoderProviders
- 
Creates a  CoderProvider from a class's  static <T> Coder<T> of(TypeDescriptor<T>,
 List<Coder<?>>) method. 
- fromString(String, boolean) - Static method in class org.apache.beam.sdk.io.LocalResources
-  
- fromString(ValueProvider<String>, boolean) - Static method in class org.apache.beam.sdk.io.LocalResources
-  
- fromSubscription(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
- 
Reads from the given subscription. 
- fromSubscription(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
- 
- fromSupplier(SerializableSupplier<Matcher<T>>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- fromTable(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
- 
A table name to be read in Snowflake. 
- fromTable(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
-  
- fromTableSchema(TableSchema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
- fromTableSchema(TableSchema, BigQueryUtils.SchemaConversionOptions) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
- fromTopic(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
- 
Creates and returns a transform for reading from a Cloud Pub/Sub topic. 
- fromTopic(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
- 
- fromUri(URI) - Static method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
- 
Creates a GcsPath from a URI. 
- fromUri(String) - Static method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
- 
Creates a GcsPath from a URI in string form. 
- FULL_RANGE - Static variable in class org.apache.beam.sdk.io.azure.cosmos.NormalizedRange
-  
- FullNameTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider
- 
Base class for table providers that look up table metadata using full table names, instead of
 querying it by parts of the name separately. 
- FullNameTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.FullNameTableProvider
-  
- fullOuterJoin(String, PCollection<KV<K, V1>>, PCollection<KV<K, V2>>, V1, V2) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
- 
Full Outer Join of two collections of KV elements. 
- fullOuterJoin(PCollection<KV<K, V1>>, PCollection<KV<K, V2>>, V1, V2) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
- 
Full Outer Join of two collections of KV elements. 
- fullOuterJoin(PCollection<RhsT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Join
- 
Perform a full outer join. 
- fullUpdate(String, String) - Static method in class org.apache.beam.sdk.io.mongodb.UpdateField
- 
Sets the limit of documents to find. 
- fullyExpand(Map<TupleTag<?>, PValue>) - Static method in class org.apache.beam.sdk.values.PValues
- 
- fun1(ScalaInterop.Fun1<T, V>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.utils.ScalaInterop
-  
- fun2(ScalaInterop.Fun2<T1, T2, V>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.utils.ScalaInterop
-  
- functionGroup - Variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ZetaSqlScalarFunctionImpl
- 
ZetaSQL function group identifier. 
- fuse(PipelineTranslator.UnresolvedTranslation<T, T2>) - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator.UnresolvedTranslation
-  
- Gauge - Interface in org.apache.beam.sdk.metrics
- 
A metric that reports the latest value out of reported values. 
- gauge(String, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
- 
Create a metric that can have its new value set, and is aggregated by taking the last reported
 value. 
- gauge(Class<?>, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
- 
Create a metric that can have its new value set, and is aggregated by taking the last reported
 value. 
- GaugeImpl - Class in org.apache.beam.runners.jet.metrics
- 
- GaugeImpl(MetricName) - Constructor for class org.apache.beam.runners.jet.metrics.GaugeImpl
-  
- GaugeResult - Class in org.apache.beam.sdk.metrics
- 
The result of a  Gauge metric. 
- GaugeResult() - Constructor for class org.apache.beam.sdk.metrics.GaugeResult
-  
- GaugeResult.EmptyGaugeResult - Class in org.apache.beam.sdk.metrics
- 
- GceMetadataUtil - Class in org.apache.beam.sdk.extensions.gcp.util
-  
- GceMetadataUtil() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.GceMetadataUtil
-  
- GcpCredentialFactory - Class in org.apache.beam.sdk.extensions.gcp.auth
- 
Construct an oauth credential to be used by the SDK and the SDK workers. 
- GcpIoPipelineOptionsRegistrar - Class in org.apache.beam.sdk.io.gcp.common
- 
A registrar containing the default GCP options. 
- GcpIoPipelineOptionsRegistrar() - Constructor for class org.apache.beam.sdk.io.gcp.common.GcpIoPipelineOptionsRegistrar
-  
- GcpOAuthScopesFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpOAuthScopesFactory
-  
- GcpOptions - Interface in org.apache.beam.sdk.extensions.gcp.options
- 
Options used to configure Google Cloud Platform specific options such as the project and
 credentials. 
- GcpOptions.DefaultProjectFactory - Class in org.apache.beam.sdk.extensions.gcp.options
- 
Attempts to infer the default project based upon the environment this application is executing
 within. 
- GcpOptions.EnableStreamingEngineFactory - Class in org.apache.beam.sdk.extensions.gcp.options
- 
EnableStreamingEngine defaults to false unless one of the two experiments is set. 
- GcpOptions.GcpOAuthScopesFactory - Class in org.apache.beam.sdk.extensions.gcp.options
- 
Returns the default set of OAuth scopes. 
- GcpOptions.GcpTempLocationFactory - Class in org.apache.beam.sdk.extensions.gcp.options
- 
- GcpOptions.GcpUserCredentialsFactory - Class in org.apache.beam.sdk.extensions.gcp.options
- 
Attempts to load the GCP credentials. 
- GcpPipelineOptionsRegistrar - Class in org.apache.beam.sdk.extensions.gcp.options
- 
A registrar containing the default GCP options. 
- GcpPipelineOptionsRegistrar() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcpPipelineOptionsRegistrar
-  
- GCPSecretSessionServiceFactory - Class in org.apache.beam.sdk.io.solace.broker
- 
This class implements a  SessionServiceFactory that retrieve the basic authentication
 credentials from a Google Cloud Secret Manager secret. 
- GCPSecretSessionServiceFactory() - Constructor for class org.apache.beam.sdk.io.solace.broker.GCPSecretSessionServiceFactory
-  
- GCPSecretSessionServiceFactory.Builder - Class in org.apache.beam.sdk.io.solace.broker
-  
- GcpTempLocationFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpTempLocationFactory
-  
- GcpUserCredentialsFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpUserCredentialsFactory
-  
- GCS_URI - Static variable in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
- 
Pattern that is used to parse a GCS URL. 
- GcsCountersOptions() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.GcsCountersOptions
-  
- GcsCreateOptions - Class in org.apache.beam.sdk.extensions.gcp.storage
- 
An abstract class that contains common configuration options for creating resources. 
- GcsCreateOptions() - Constructor for class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions
-  
- GcsCreateOptions.Builder - Class in org.apache.beam.sdk.extensions.gcp.storage
- 
- GcsFileSystemRegistrar - Class in org.apache.beam.sdk.extensions.gcp.storage
- 
AutoServiceregistrar for theGcsFileSystem.
 
- GcsFileSystemRegistrar() - Constructor for class org.apache.beam.sdk.extensions.gcp.storage.GcsFileSystemRegistrar
-  
- GcsOptions - Interface in org.apache.beam.sdk.extensions.gcp.options
- 
Options used to configure Google Cloud Storage. 
- GcsOptions.ExecutorServiceFactory - Class in org.apache.beam.sdk.extensions.gcp.options
- 
Returns the default ExecutorServiceto use within the Apache Beam SDK.
 
- GcsOptions.PathValidatorFactory - Class in org.apache.beam.sdk.extensions.gcp.options
- 
- GcsPath - Class in org.apache.beam.sdk.extensions.gcp.util.gcsfs
- 
Implements the Java NIO PathAPI for Google Cloud Storage paths.
 
- GcsPath(FileSystem, String, String) - Constructor for class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
- 
Constructs a GcsPath. 
- GcsPathValidator - Class in org.apache.beam.sdk.extensions.gcp.storage
- 
- GcsReadOptionsFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.GcsReadOptionsFactory
-  
- GcsResourceId - Class in org.apache.beam.sdk.extensions.gcp.storage
- 
ResourceId implementation for Google Cloud Storage. 
- GcsStager - Class in org.apache.beam.runners.dataflow.util
- 
Utility class for staging files to GCS. 
- gcsUploadBufferSizeBytes() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions
- 
The buffer size (in bytes) to use when uploading files to GCS. 
- GcsUtil - Class in org.apache.beam.sdk.extensions.gcp.util
- 
Provides operations on GCS. 
- GcsUtil.CreateOptions - Class in org.apache.beam.sdk.extensions.gcp.util
-  
- GcsUtil.CreateOptions.Builder - Class in org.apache.beam.sdk.extensions.gcp.util
-  
- GcsUtil.GcsCountersOptions - Class in org.apache.beam.sdk.extensions.gcp.util
-  
- GcsUtil.GcsReadOptionsFactory - Class in org.apache.beam.sdk.extensions.gcp.util
-  
- GcsUtil.GcsUtilFactory - Class in org.apache.beam.sdk.extensions.gcp.util
- 
- GcsUtil.StorageObjectOrIOException - Class in org.apache.beam.sdk.extensions.gcp.util
- 
- GcsUtilFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.GcsUtilFactory
-  
- generate(Schema) - Static method in class org.apache.beam.sdk.coders.RowCoderGenerator
-  
- generateInitialChangeStreamPartitions() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.ChangeStreamDao
- 
Returns the result from GenerateInitialChangeStreamPartitions API. 
- generateInitialPartitionsAction(ChangeStreamMetrics, ChangeStreamDao, Instant) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.ActionFactory
- 
- GenerateInitialPartitionsAction - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.action
- 
- GenerateInitialPartitionsAction(ChangeStreamMetrics, ChangeStreamDao, Instant) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.GenerateInitialPartitionsAction
-  
- generateRandom(String) - Static method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataTableNames
- 
Generates a unique name for the partition metadata table and its indexes. 
- generateRowKeyPrefix() - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.UniqueIdGenerator
- 
Return a random base64 encoded 8 byte string. 
- GenerateSequence - Class in org.apache.beam.sdk.io
- 
A  PTransform that produces longs starting from the given value, and either up to the
 given limit or until  Long.MAX_VALUE / until the given time elapses. 
- GenerateSequence() - Constructor for class org.apache.beam.sdk.io.GenerateSequence
-  
- GenerateSequence.External - Class in org.apache.beam.sdk.io
- 
Exposes GenerateSequence as an external transform for cross-language usage. 
- GenerateSequence.External.ExternalConfiguration - Class in org.apache.beam.sdk.io
- 
Parameters class to expose the transform to an external SDK. 
- GenerateSequenceConfiguration() - Constructor for class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration
-  
- GenerateSequenceSchemaTransformProvider - Class in org.apache.beam.sdk.providers
-  
- GenerateSequenceSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider
-  
- GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration - Class in org.apache.beam.sdk.providers
-  
- GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration.Builder - Class in org.apache.beam.sdk.providers
-  
- GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration.Rate - Class in org.apache.beam.sdk.providers
-  
- GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration.Rate.Builder - Class in org.apache.beam.sdk.providers
-  
- GenerateSequenceSchemaTransformProvider.GenerateSequenceSchemaTransform - Class in org.apache.beam.sdk.providers
-  
- GenerateSequenceTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.seqgen
- 
Sequence generator table provider. 
- GenerateSequenceTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.seqgen.GenerateSequenceTableProvider
-  
- generic(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns an  AvroCoder instance for the Avro schema. 
- generic() - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory
- 
- GenericDatumFactory() - Constructor for class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory.GenericDatumFactory
-  
- GenericDlq - Class in org.apache.beam.sdk.schemas.io
- 
Helper to generate a DLQ transform to write PCollection to an external system. 
- GenericDlqProvider - Interface in org.apache.beam.sdk.schemas.io
- 
A Provider for generic DLQ transforms that handle deserialization failures. 
- get(JobInfo) - Method in interface org.apache.beam.runners.fnexecution.control.ExecutableStageContext.Factory
- 
- get(JobInfo) - Method in class org.apache.beam.runners.fnexecution.control.ReferenceCountingExecutableStageContextFactory
-  
- get(K, W) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.BagUserStateHandler
- 
Returns an Iterableof values representing the bag user state for the given key and
 window.
 
- get(W) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.IterableSideInputHandler
- 
Returns an Iterableof values representing the side input for the given window.
 
- get(W) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.MultimapSideInputHandler
- 
Returns an Iterableof keys representing the side input for the given window.
 
- get(K, W) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.MultimapSideInputHandler
- 
Returns an Iterableof values representing the side input for the given key and
 window.
 
- get() - Method in class org.apache.beam.runners.portability.CloseableResource
- 
Gets the underlying resource. 
- get() - Static method in class org.apache.beam.runners.spark.io.EmptyCheckpointMark
-  
- get(PCollectionView<T>, BoundedWindow) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.CachedSideInputReader
-  
- get(BoundedWindow) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.SideInputValues.ByWindow
-  
- get(BoundedWindow) - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.SideInputValues
-  
- get(BoundedWindow) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.SideInputValues.Global
-  
- get(PCollectionView<T>, BoundedWindow) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.SparkSideInputReader
-  
- get(PCollectionView<T>, BoundedWindow) - Method in class org.apache.beam.runners.spark.util.CachedSideInputReader
-  
- get(Long) - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
- 
- get(PCollectionView<T>, BoundedWindow) - Method in class org.apache.beam.runners.spark.util.SparkSideInputReader
-  
- get() - Method in class org.apache.beam.runners.twister2.translators.functions.Twister2SinkFunction
-  
- get(PCollectionView<T>, BoundedWindow) - Method in class org.apache.beam.runners.twister2.utils.Twister2SideInputReader
-  
- get(int) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedList
-  
- get(Object) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedMap
-  
- get(int) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedRow
-  
- get() - Method in class org.apache.beam.sdk.io.aws2.schemas.AwsBuilderFactory
-  
- get() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.estimator.BytesThroughputEstimator
- 
Returns the estimated throughput bytes for this run. 
- get() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.estimator.BytesThroughputEstimator
- 
Returns the estimated throughput bytes for now. 
- get() - Method in interface org.apache.beam.sdk.io.gcp.spanner.changestreams.estimator.ThroughputEstimator
- 
Returns the estimated throughput for now. 
- get() - Method in class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
-  
- get() - Method in interface org.apache.beam.sdk.options.ValueProvider
- 
- get() - Method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
-  
- get() - Method in class org.apache.beam.sdk.options.ValueProvider.RuntimeValueProvider
-  
- get() - Method in class org.apache.beam.sdk.options.ValueProvider.StaticValueProvider
-  
- get(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.AutoValueSchema.AbstractGetterTypeSupplier
-  
- get(ObjectT) - Method in interface org.apache.beam.sdk.schemas.FieldValueGetter
-  
- get(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema.GetterTypeSupplier
-  
- get(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema.SetterTypeSupplier
-  
- get(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.JavaFieldSchema.JavaFieldTypeSupplier
-  
- get(Object) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
-  
- get(TypeDescriptor<?>) - Method in interface org.apache.beam.sdk.schemas.utils.FieldValueTypeSupplier
- 
Return all the FieldValueTypeInformations. 
- get(TypeDescriptor<?>, Schema) - Method in interface org.apache.beam.sdk.schemas.utils.FieldValueTypeSupplier
- 
Return all the FieldValueTypeInformations. 
- get(K) - Method in interface org.apache.beam.sdk.state.MapState
- 
A deferred lookup, using null values if the item is not found. 
- get(K) - Method in interface org.apache.beam.sdk.state.MultimapState
- 
A deferred lookup, returns an empty iterable if the item is not found. 
- get(String) - Method in interface org.apache.beam.sdk.state.TimerMap
-  
- get(TupleTag<V>) - Method in class org.apache.beam.sdk.transforms.CombineFns.CoCombineResult
- 
Returns the value represented by the given  TupleTag. 
- get(TupleTag<T>) - Method in interface org.apache.beam.sdk.transforms.DoFn.MultiOutputReceiver
- 
- get() - Method in interface org.apache.beam.sdk.transforms.Materializations.IterableView
- 
Returns an iterable for all values. 
- get() - Method in interface org.apache.beam.sdk.transforms.Materializations.MultimapView
- 
Returns an iterable of all keys. 
- get(K) - Method in interface org.apache.beam.sdk.transforms.Materializations.MultimapView
- 
Returns an iterable of all the values for the specified key. 
- get(int) - Method in class org.apache.beam.sdk.values.PCollectionList
- 
Returns the  PCollection at the given index (origin zero). 
- get(String) - Method in class org.apache.beam.sdk.values.PCollectionRowTuple
- 
- get(TupleTag<T>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
- 
- get(String) - Method in class org.apache.beam.sdk.values.PCollectionTuple
- 
- get() - Method in class org.apache.beam.sdk.values.PCollectionViews.ValueOrMetadata
-  
- get(int) - Method in class org.apache.beam.sdk.values.TupleTagList
- 
Returns the  TupleTag at the given index (origin zero). 
- getAcceptedIssuers() - Method in class org.apache.beam.sdk.io.splunk.CustomX509TrustManager
-  
- getAccessKey() - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- getAccountName() - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- getAccum() - Method in interface org.apache.beam.sdk.state.CombiningState
- 
Read the merged accumulator for this state cell. 
- getAccumTVariable() - Method in class org.apache.beam.sdk.extensions.sql.impl.LazyAggregateCombineFn
-  
- getAccumTVariable() - Method in class org.apache.beam.sdk.extensions.sql.TypedCombineFnDelegate
-  
- getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.runners.flink.FlinkStreamingAggregationsTranslators.ConcatenateAsIterable
-  
- getAccumulatorCoder(CoderRegistry, Coder<TimestampedValue<KV<EventKeyT, KV<Long, EventT>>>>) - Method in class org.apache.beam.sdk.extensions.ordered.combiner.DefaultSequenceCombiner
-  
- getAccumulatorCoder(CoderRegistry, Coder) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.CountMinSketchFn
-  
- getAccumulatorCoder(CoderRegistry, Coder) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.TDigestQuantilesFn
-  
- getAccumulatorCoder(CoderRegistry, Coder<InputT>) - Method in class org.apache.beam.sdk.extensions.sql.impl.LazyAggregateCombineFn
-  
- getAccumulatorCoder(CoderRegistry, Coder<Boolean>) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CountIf.CountIfFn
-  
- getAccumulatorCoder(CoderRegistry, Coder<Row>) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CovarianceFn
-  
- getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.VarianceFn
-  
- getAccumulatorCoder(CoderRegistry, Coder<InputT>) - Method in class org.apache.beam.sdk.extensions.sql.TypedCombineFnDelegate
-  
- getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
-  
- getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn
- 
Deprecated. 
- getAccumulatorCoder(CoderRegistry, Coder<Double>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
-  
- getAccumulatorCoder(CoderRegistry, Coder<V>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
-  
- getAccumulatorCoder(CoderRegistry, Coder<Integer>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
-  
- getAccumulatorCoder(CoderRegistry, Coder<Long>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
-  
- getAccumulatorCoder(CoderRegistry, Coder<InputT>) - Method in interface org.apache.beam.sdk.transforms.CombineFnBase.GlobalCombineFn
- 
Returns the Coderto use for accumulatorAccumTvalues, or null if it is not
 able to be inferred.
 
- getAccumulatorCoder(CoderRegistry, Coder<DataT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
-  
- getAccumulatorCoder(CoderRegistry, Coder<DataT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
-  
- getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
-  
- getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
-  
- getActiveWorkRefreshPeriodMillis() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getAdditionalInputs() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.ParDoSingle
-  
- getAdditionalInputs() - Method in class org.apache.beam.sdk.io.WriteFiles
-  
- getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
- 
- getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
- 
- getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
- 
- getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
- 
- getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.PTransform
- 
- getAdditionalOutputTags() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-  
- getAddresses() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getAlgorithm() - Method in enum org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType
- 
Returns the string representation of this type. 
- getAlgorithm() - Method in class org.apache.beam.sdk.io.aws2.s3.SSECustomerKey
-  
- getAll() - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO
- 
Retrieve all HL7v2 Messages from a PCollection of message IDs (such as from PubSub notification
 subscription). 
- getAll(TupleTag<V>) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
- 
Returns the values from the table represented by the given TupleTag<V>as anIterable<V>(which may be empty if there are no results).
 
- getAll(String) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
- 
- getAll() - Method in class org.apache.beam.sdk.values.PCollectionList
- 
- getAll() - Method in class org.apache.beam.sdk.values.PCollectionRowTuple
- 
- getAll() - Method in class org.apache.beam.sdk.values.PCollectionTuple
- 
- getAll() - Method in class org.apache.beam.sdk.values.TupleTagList
- 
- getAllFields() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
If true, all fields are being accessed. 
- getAllIds(String, String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- getAllJobs() - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
-  
- getAllMetadata() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
- getAllowDuplicates() - Method in class org.apache.beam.sdk.transforms.Redistribute.RedistributeArbitrarily
-  
- getAllowDuplicates() - Method in class org.apache.beam.sdk.transforms.Redistribute.RedistributeByKey
-  
- getAllowedLateness() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- getAllowedTimestampSkew() - Method in class org.apache.beam.sdk.transforms.DoFn
- 
- getAllowedTimestampSkew() - Method in class org.apache.beam.sdk.transforms.WithTimestamps
- 
- getAllowlist() - Method in class org.apache.beam.sdk.expansion.service.ExpansionServiceConfig
-  
- getAllowNonRestoredState() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getAllPartitionsCreatedAfter(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao
- 
- getAllRows(String, String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- getAllWorkerStatuses(long, TimeUnit) - Method in class org.apache.beam.runners.fnexecution.status.BeamWorkerStatusGrpcService
- 
Get all the statuses from all connected SDK harnesses within specified timeout. 
- getAlpha() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPFieldRef
-  
- getAlsoStartLoopbackWorker() - Method in interface org.apache.beam.sdk.expansion.service.ExpansionServiceOptions
-  
- getAnnotatedConstructor(Class<?>) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
-  
- getAnnotatedCreateMethod(Class<?>) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
-  
- getAnnotations() - Method in class org.apache.beam.sdk.transforms.PTransform
- 
Returns annotations map to provide additional hints to the runner. 
- getApiKey() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getApiPrefix() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
Generates the API endpoint prefix based on the set values. 
- getApiRootUrl() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
The root URL for the Dataflow API. 
- getApiServiceDescriptor() - Method in class org.apache.beam.sdk.fn.server.GrpcFnServer
- 
Get an  Endpoints.ApiServiceDescriptor describing the endpoint this  GrpcFnServer is bound
 to. 
- getAppend() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider.Configuration
-  
- getAppId() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getApplicationName() - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher.TableAndQuery
-  
- getAppliedFn(CoderRegistry, Coder<? extends KV<K, ? extends Iterable<InputT>>>, WindowingStrategy<?, ?>) - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
- 
- getApplyMethod(ScalarFn) - Static method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFnReflector
- 
- getAppName() - Method in interface org.apache.beam.sdk.options.ApplicationNameOptions
- 
Name of application, for display purposes. 
- getAppProfileId() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableConfig
- 
Returns the app profile being read from. 
- getApproximateArrivalTimestamp() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisRecord
-  
- getArgument() - Method in class org.apache.beam.sdk.schemas.logicaltypes.Date
-  
- getArgument() - Method in class org.apache.beam.sdk.schemas.logicaltypes.DateTime
-  
- getArgument() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
-  
- getArgument() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
-  
- getArgument() - Method in class org.apache.beam.sdk.schemas.logicaltypes.PassThroughLogicalType
-  
- getArgument() - Method in class org.apache.beam.sdk.schemas.logicaltypes.Time
-  
- getArgument() - Method in class org.apache.beam.sdk.schemas.logicaltypes.UuidLogicalType
-  
- getArgument() - Method in interface org.apache.beam.sdk.schemas.Schema.LogicalType
- 
An optional argument to configure the type. 
- getArguments() - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getArgumentType() - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.TZTimeOnly
-  
- getArgumentType() - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.TZTimestamp
-  
- getArgumentType() - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.UTCTimeOnly
-  
- getArgumentType() - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.UTCTimestamp
-  
- getArgumentType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.Date
-  
- getArgumentType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.DateTime
-  
- getArgumentType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
-  
- getArgumentType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.MicrosInstant
-  
- getArgumentType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
-  
- getArgumentType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.PassThroughLogicalType
-  
- getArgumentType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.PythonCallable
-  
- getArgumentType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.SchemaLogicalType
-  
- getArgumentType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.Time
-  
- getArgumentType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.UuidLogicalType
-  
- getArgumentType() - Method in interface org.apache.beam.sdk.schemas.Schema.LogicalType
- 
A schema type representing how to interpret the argument. 
- getArgumentTypes(Method) - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
Returns a list of argument types for the given method, which must be a part of the class. 
- getArray(String) - Method in class org.apache.beam.sdk.values.Row
- 
Get an array value by field name, IllegalStateExceptionis thrown if schema doesn't
 match.
 
- getArray(int) - Method in class org.apache.beam.sdk.values.Row
- 
Get an array value by field index, IllegalStateExceptionis thrown if schema doesn't
 match.
 
- getArtifact(ArtifactApi.GetArtifactRequest, StreamObserver<ArtifactApi.GetArtifactResponse>) - Method in class org.apache.beam.runners.fnexecution.artifact.ArtifactRetrievalService
-  
- getArtifact(RunnerApi.ArtifactInformation) - Static method in class org.apache.beam.runners.fnexecution.artifact.ArtifactRetrievalService
-  
- getArtifactPort() - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver.ServerConfiguration
-  
- getArtifactStagingPath() - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver.ServerConfiguration
-  
- getATN() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- getATN() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- getAttachedMode() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getAttachmentBytes() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record
- 
Gets the attachment data of the message as a byte array, if any. 
- getAttempted() - Method in class org.apache.beam.sdk.metrics.MetricResult
- 
Return the value of this metric across all attempts of executing all parts of the pipeline. 
- getAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
- 
Returns the given attribute value. 
- getAttributeId() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration
-  
- getAttributeId() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration
-  
- getAttributeMap() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
- 
Returns the full map of attributes. 
- getAttributeMap() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration
-  
- getAttributes() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration
-  
- getAttributes() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration
-  
- getAttributes() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration
-  
- getAttributes() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration
-  
- getAttributesMap() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration
-  
- getAttributesMap() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration
-  
- getAuthenticator() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getAuthenticator() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getAuthToken(String, String) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverConfiguration
- 
Certain embedded scenarios and so on actually allow for having no authentication at all. 
- getAutoCommit() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- getAutoOffsetResetConfig() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
-  
- getAutoscalingAlgorithm() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
- 
The autoscaling algorithm to use for the workerpool. 
- getAutoSharding() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- getAutosharding() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration
-  
- getAutoWatermarkInterval() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getAvroBytesToRowFunction(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
- getAvroFilterFormatFunction(boolean) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.PortableBigQueryDestinations
-  
- getAwsCredentialsProvider() - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
- 
- getAwsRegion() - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
- 
Region used to configure AWS service clients. 
- getAzureConnectionString() - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- getAzureCredentialsProvider() - Method in interface org.apache.beam.sdk.io.azure.options.AzureOptions
- 
The credential instance that should be used to authenticate against Azure services. 
- getBacking() - Method in class org.apache.beam.sdk.fn.data.WeightedList
-  
- getBacklogBytes(String) - Method in class org.apache.beam.sdk.io.solace.broker.BasicAuthSempClient
-  
- getBacklogBytes(String) - Method in class org.apache.beam.sdk.io.solace.broker.SempBasicAuthClientExecutor
-  
- getBacklogBytes(String) - Method in interface org.apache.beam.sdk.io.solace.broker.SempClient
- 
Retrieves the size of the backlog (in bytes) for the specified queue. 
- getBacklogCheckTime() - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicy.PartitionContext
- 
The time at which latest offset for the partition was fetched in order to calculate backlog. 
- getBadRecordErrorHandler() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getBadRecordErrorHandler() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- getBadRecordErrorHandler() - Method in class org.apache.beam.sdk.io.WriteFiles
-  
- getBadRecordRouter() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- getBadRecordRouter() - Method in class org.apache.beam.sdk.io.WriteFiles
-  
- getBagUserStateSpecs() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
- 
Get a mapping from PTransform id to user state input id to  bag user
 states that are used during execution. 
- getBaseAutoValueClass(TypeDescriptor<?>) - Static method in class org.apache.beam.sdk.schemas.utils.AutoValueUtils
-  
- getBaseName() - Method in class org.apache.beam.sdk.metrics.LabeledMetricNameUtils.ParsedMetricName
-  
- getBaseType() - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.TZTimeOnly
-  
- getBaseType() - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.TZTimestamp
-  
- getBaseType() - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.UTCTimeOnly
-  
- getBaseType() - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.UTCTimestamp
-  
- getBaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.Date
-  
- getBaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.DateTime
-  
- getBaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
-  
- getBaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.MicrosInstant
-  
- getBaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
-  
- getBaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.PassThroughLogicalType
-  
- getBaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.PythonCallable
-  
- getBaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.SchemaLogicalType
-  
- getBaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.Time
-  
- getBaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.UuidLogicalType
-  
- getBaseType() - Method in interface org.apache.beam.sdk.schemas.Schema.LogicalType
- 
- getBaseValue(String, Class<T>) - Method in class org.apache.beam.sdk.values.Row
- 
Returns the base type for this field. 
- getBaseValue(String) - Method in class org.apache.beam.sdk.values.Row
- 
Returns the base type for this field. 
- getBaseValue(int, Class<T>) - Method in class org.apache.beam.sdk.values.Row
- 
Returns the base type for this field. 
- getBaseValue(int) - Method in class org.apache.beam.sdk.values.Row
- 
Returns the base type for this field. 
- getBaseValues() - Method in class org.apache.beam.sdk.values.Row
- 
Return a list of data values. 
- getBatchClient() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerAccessor
-  
- getBatchCombinePerKeyOperator(FlinkStreamingTranslationContext, PCollection<KV<K, InputT>>, Map<Integer, PCollectionView<?>>, List<PCollectionView<?>>, Coder<WindowedValue<KV<K, AccumT>>>, CombineFnBase.GlobalCombineFn<InputT, AccumT, ?>, <any>, TypeInformation<WindowedValue<KV<K, OutputT>>>) - Static method in class org.apache.beam.runners.flink.FlinkStreamingAggregationsTranslators
-  
- getBatchDuration() - Method in class org.apache.beam.runners.spark.io.CreateStream
-  
- getBatches() - Method in class org.apache.beam.runners.spark.io.CreateStream
- 
Get the underlying queue representing the mock stream of micro-batches. 
- getBatching() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration
-  
- getBatchingParams() - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches
- 
Returns user supplied parameters for batching. 
- getBatchingParams() - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches.WithShardedKey
- 
Returns user supplied parameters for batching. 
- getBatchInitialCount() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
- 
The initial size of a batch; used in the absence of the QoS system having significant data to
 determine a better batch size. 
- getBatchIntervalMillis() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- getBatchMaxBytes() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
- 
The maximum number of bytes to include in a batch. 
- getBatchMaxCount() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
- 
The maximum number of writes to include in a batch. 
- getBatchService() - Method in interface org.apache.beam.sdk.io.snowflake.services.SnowflakeServices
-  
- getBatchService() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeServicesImpl
-  
- getBatchSize() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration
-  
- getBatchSize() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteConfiguration
-  
- getBatchSize() - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches.BatchingParams
-  
- getBatchSizeBytes() - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText
-  
- getBatchSizeBytes() - Method in class org.apache.beam.sdk.extensions.ml.DLPInspectText
-  
- getBatchSizeBytes() - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText
-  
- getBatchSizeBytes() - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches.BatchingParams
-  
- getBatchTargetLatency() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
- 
Target latency for batch requests. 
- getBeamRelInput(RelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils
-  
- getBeamSchemaFromProto(String, String) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoByteUtils
- 
Retrieves a Beam Schema from a Protocol Buffer message. 
- getBeamSchemaFromProtoSchema(String, String) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoByteUtils
- 
Parses the given Protocol Buffers schema string, retrieves the Descriptor for the specified
 message name, and constructs a Beam Schema from it. 
- getBeamSqlTable() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
-  
- getBeamSqlUdfs() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.UdfUdafProvider
- 
- getBearerToken() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getBigQueryEndpoint() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
- 
BQ endpoint to use. 
- getBigQueryLocation() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.TestBigQueryOptions
-  
- getBigQueryProject() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getBigtableChangeStreamInstanceId() - Method in interface org.apache.beam.sdk.io.gcp.bigtable.changestreams.BigtableChangeStreamTestOptions
-  
- getBigtableClientOverride() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableConfig
- 
Returns the Bigtable client override. 
- getBigtableOptions() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
- getBigtableOptions() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- getBlobServiceEndpoint() - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
- 
The Azure Blobstore service endpoint used by the Blob service client. 
- getBlobstoreClientFactoryClass() - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- getBody() - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsMessage
- 
Message body. 
- getBody() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getBoolean(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getBoolean(Map<String, Object>, String, Boolean) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getBoolean() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- getBoolean(String) - Method in class org.apache.beam.sdk.values.Row
- 
- getBoolean(int) - Method in class org.apache.beam.sdk.values.Row
- 
Get a Booleanvalue by field index,ClassCastExceptionis thrown if schema
 doesn't match.
 
- getBootstrapServers() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
-  
- getBootstrapServers() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
- 
Sets the bootstrap servers for the Kafka consumer. 
- getBootstrapServers() - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration
-  
- getBoundednessOfRelNode(RelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
- 
This method returns the Boundedness of a RelNode. 
- getBoundedTrie(MetricName) - Method in class org.apache.beam.runners.jet.metrics.JetMetricsContainer
-  
- getBoundedTrie(MetricName) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
- 
Return the  BoundedTrie that should be used for implementing the given  metricName in this container. 
- getBoundedTries() - Method in class org.apache.beam.sdk.metrics.MetricQueryResults
- 
Return the metric results for the bounded tries that matched the filter. 
- getBqStreamingApiLoggingFrequencySec() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getBranch() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getBroadcastSizeEstimate() - Method in class org.apache.beam.runners.spark.util.SideInputBroadcast
-  
- getBucket() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
- 
Returns the bucket name associated with this GCS path, or an empty string if this is a relative
 path component. 
- getBucket(GcsPath) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
- 
Get the  Bucket from Cloud Storage path or propagates an exception. 
- getBucketKeyEnabled() - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- getBucketKeyEnabled() - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration
- 
Whether to use an S3 Bucket Key for object encryption with server-side encryption using AWS KMS
 (SSE-KMS) or not. 
- getBufferSize() - Method in class org.apache.beam.sdk.fn.stream.BufferingStreamObserver
-  
- getBuilderCreator(Class<?>, Schema, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.AutoValueUtils
- 
Try to find an accessible builder class for creating an AutoValue class. 
- getBuiltinMethods() - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BeamBuiltinFunctionProvider
-  
- getBulkDirective() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Document
-  
- getBulkEndPoint() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getBulkIO() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
-  
- getBundle(OutputReceiverFactory, StateRequestHandler, BundleProgressHandler) - Method in interface org.apache.beam.runners.fnexecution.control.StageBundleFactory
- 
Get a new  bundle for processing the data in an executable stage. 
- getBundle(OutputReceiverFactory, StateRequestHandler, BundleProgressHandler, BundleFinalizationHandler) - Method in interface org.apache.beam.runners.fnexecution.control.StageBundleFactory
-  
- getBundle(OutputReceiverFactory, StateRequestHandler, BundleProgressHandler, BundleFinalizationHandler, BundleCheckpointHandler) - Method in interface org.apache.beam.runners.fnexecution.control.StageBundleFactory
-  
- getBundle(OutputReceiverFactory, TimerReceiverFactory, StateRequestHandler, BundleProgressHandler) - Method in interface org.apache.beam.runners.fnexecution.control.StageBundleFactory
- 
Get a new  bundle for processing the data in an executable stage. 
- getBundle(OutputReceiverFactory, TimerReceiverFactory, StateRequestHandler, BundleProgressHandler, BundleFinalizationHandler) - Method in interface org.apache.beam.runners.fnexecution.control.StageBundleFactory
-  
- getBundle(OutputReceiverFactory, TimerReceiverFactory, StateRequestHandler, BundleProgressHandler, BundleFinalizationHandler, BundleCheckpointHandler) - Method in interface org.apache.beam.runners.fnexecution.control.StageBundleFactory
-  
- getBundle() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirBundleParameter
- 
FHIR R4 bundle resource object as a string. 
- getBundleProcessorCacheTimeout() - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
-  
- getBundleSize() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- getByte() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- getByte(String) - Method in class org.apache.beam.sdk.values.Row
- 
Get a  Schema.TypeName.BYTE value by field name,  IllegalStateException is thrown if
 schema doesn't match. 
- getByte(int) - Method in class org.apache.beam.sdk.values.Row
- 
Get a  Schema.TypeName.BYTE value by field index,  ClassCastException is thrown if
 schema doesn't match. 
- getBytes(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getBytes(Map<String, Object>, String, byte[]) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getBytes() - Method in class org.apache.beam.sdk.io.range.ByteKey
- 
Returns a newly-allocated  byte[] representing this  ByteKey. 
- getBytes(String) - Method in class org.apache.beam.sdk.values.Row
- 
Get a  Schema.TypeName.BYTES value by field name,  IllegalStateException is thrown if
 schema doesn't match. 
- getBytes(int) - Method in class org.apache.beam.sdk.values.Row
- 
- getBytesPerOffset() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
- 
Returns approximately how many bytes of data correspond to a single offset in this source. 
- getCacheTokens() - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandler
- 
Retrieves a list of valid cache tokens. 
- getCallable() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaRowUdf.Configuration
-  
- getCaseEnumType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
- 
- getCaseSensitive() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getCaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType.Value
- 
Returns the enumeration that specified which OneOf field is set. 
- getCatalog() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- getCatalogConfig() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getCatalogName() - Method in class org.apache.beam.sdk.io.iceberg.IcebergCatalogConfig
-  
- getCatalogName() - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration
-  
- getCatalogName() - Method in class org.apache.beam.sdk.io.iceberg.SchemaTransformConfiguration
-  
- getCatalogProperties() - Method in class org.apache.beam.sdk.io.iceberg.IcebergCatalogConfig
-  
- getCatalogProperties() - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration
-  
- getCatalogProperties() - Method in class org.apache.beam.sdk.io.iceberg.SchemaTransformConfiguration
-  
- getCause() - Method in class org.apache.beam.runners.jet.FailedRunningPipelineResults
-  
- getCellsMutatedPerColumn(String, String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerSchema
- 
Return the total number of cells affected when the specified column is mutated. 
- getCellsMutatedPerRow(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerSchema
- 
Return the total number of cells affected with the given row is deleted. 
- getCEPFieldRefFromParKeys(ImmutableBitSet) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPUtils
- 
Transform the partition columns into serializable CEPFieldRef. 
- getCepKind() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPOperator
-  
- getCEPPatternFromPattern(Schema, RexNode, Map<String, RexNode>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPUtils
- 
Construct a list of CEPPatterns from aRexNode.
 
- getChangeSequenceNumber() - Method in class org.apache.beam.sdk.io.gcp.bigquery.RowMutationInformation
- 
The value supplied to the BigQuery _CHANGE_SEQUENCE_NUMBERpseudo-column.
 
- getChangeStreamContinuationTokens() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.NewPartition
-  
- getChangeStreamContinuationTokens() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.PartitionRecord
-  
- getChangeStreamDao() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.DaoFactory
-  
- getChangeStreamDao() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.DaoFactory
- 
Creates and returns a singleton DAO instance for querying a partition change stream. 
- getChangeStreamName() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.DaoFactory
-  
- getChangeStreamName() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration
-  
- getChangeStreamNamePrefix() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
- 
Return the prefix used to identify the rows belonging to this job. 
- getChangeStreamNamePrefix() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
-  
- getChannelFactory() - Method in class org.apache.beam.sdk.io.CompressedSource
-  
- getChannelNames() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- getChannelzShowOnlyWindmillServiceChannels() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getCharset() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.XmlConfiguration
-  
- getCheckpointDir() - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
-  
- getCheckpointDurationMillis() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- getCheckpointingInterval() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getCheckpointingInterval() - Method in interface org.apache.beam.sdk.io.kafka.KafkaIO.Read.FakeFlinkPipelineOptions
-  
- getCheckpointingMode() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getCheckpointMark() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-  
- getCheckpointMark() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.UnboundedReaderImpl
-  
- getCheckpointMark() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
- 
- getCheckpointMarkCoder() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
-  
- getCheckpointMarkCoder() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.UnboundedSourceImpl
-  
- getCheckpointMarkCoder() - Method in class org.apache.beam.sdk.io.solace.read.UnboundedSolaceSource
-  
- getCheckpointMarkCoder() - Method in class org.apache.beam.sdk.io.UnboundedSource
- 
Returns a  Coder for encoding and decoding the checkpoints for this source. 
- getCheckpointTimeoutMillis() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getCheckStopReadingFn() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getChildPartitions() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChildPartitionsRecord
- 
List of child partitions yielded within this record. 
- getChildRels(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
-  
- getClass(String, String) - Static method in class org.apache.beam.sdk.schemas.transforms.providers.StringCompiler
-  
- getClasses() - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
Returns a set of  TypeDescriptors, one for each superclass (including this class). 
- getClassName() - Method in class org.apache.beam.runners.dataflow.util.CloudObject
- 
Gets the name of the Java class that this CloudObject represents. 
- getClazz() - Method in class org.apache.beam.sdk.schemas.utils.ReflectUtils.ClassWithSchema
-  
- getClient(String) - Static method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
-  
- getClientBuilderFactory() - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
-  
- getClientFactory() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration
-  
- getClientInfo(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- getClientInfo() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- getClock() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration
-  
- getCloningBehavior() - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- getCloseStream() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.StreamProgress
-  
- getClosingBehavior() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- getClosure() - Method in class org.apache.beam.sdk.transforms.Contextful
- 
Returns the closure. 
- getClusterId() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getClustering() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- getClusterName() - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
-  
- getClusterType() - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- getCode() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.TypeCode
- 
Returns the type code of the column. 
- getCodec(DestinationT) - Method in class org.apache.beam.sdk.extensions.avro.io.DynamicAvroDestinations
- 
Return an AVRO codec for a given destination. 
- getCodeJarPathname() - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
-  
- getCoder() - Method in class org.apache.beam.runners.fnexecution.control.RemoteOutputReceiver
-  
- getCoder() - Method in class org.apache.beam.runners.fnexecution.data.RemoteInputDestination
-  
- getCoder(Class<T>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
- 
Returns the  Coder to use for values of the given class. 
- getCoder(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
- 
Returns the  Coder to use for values of the given type. 
- getCoder(TypeDescriptor<OutputT>, TypeDescriptor<InputT>, Coder<InputT>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
- 
- getCoder(Class<? extends T>, Class<T>, Map<Type, ? extends Coder<?>>, TypeVariable<?>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
- 
- getCoder() - Method in class org.apache.beam.sdk.coders.DelegateCoder
- 
Returns the coder used to encode/decode the intermediate values produced/consumed by the coding
 functions of this DelegateCoder.
 
- getCoder() - Method in class org.apache.beam.sdk.fn.data.DataEndpoint
-  
- getCoder() - Method in class org.apache.beam.sdk.fn.data.TimerEndpoint
-  
- getCoder(CoderRegistry) - Method in class org.apache.beam.sdk.io.kafka.ConfluentSchemaRegistryDeserializerProvider
-  
- getCoder(CoderRegistry) - Method in interface org.apache.beam.sdk.io.kafka.DeserializerProvider
-  
- getCoder() - Method in interface org.apache.beam.sdk.io.singlestore.SingleStoreIO.RowMapperWithCoder
-  
- getCoder(Pipeline) - Static method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord
-  
- getCoder() - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Record
- 
The coder for the record, or null if there is no coder. 
- getCoder() - Static method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
- 
- getCoder() - Method in class org.apache.beam.sdk.values.PCollection
- 
Returns the  Coder used by this  PCollection to encode and decode the values
 stored in it. 
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.AtomicCoder
- 
If this is a  Coder for a parameterized type, returns the list of  Coders being
 used for each of the parameters in the same order they appear within the parameterized type's
 type signature. 
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.Coder
- 
If this is a  Coder for a parameterized type, returns the list of  Coders being
 used for each of the parameters in the same order they appear within the parameterized type's
 type signature. 
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.CustomCoder
- 
If this is a  Coder for a parameterized type, returns the list of  Coders being
 used for each of the parameters in the same order they appear within the parameterized type's
 type signature. 
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.KvCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.MapCoder
- 
If this is a  Coder for a parameterized type, returns the list of  Coders being
 used for each of the parameters in the same order they appear within the parameterized type's
 type signature. 
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.NullableCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.ShardedKeyCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.SnappyCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.SortedMapCoder
- 
If this is a  Coder for a parameterized type, returns the list of  Coders being
 used for each of the parameters in the same order they appear within the parameterized type's
 type signature. 
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.TimestampPrefixingWindowCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.VarLongCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.coders.ZstdCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.io.hadoop.WritableCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.io.kafka.TopicPartitionCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.io.ReadableFileCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.testing.TestStream.TestStreamCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.values.FailsafeValueInSingleWindow.Coder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.values.PCollectionViews.ValueOrMetadataCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
-  
- getCoderArguments() - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
-  
- getCoderInternal() - Method in interface org.apache.beam.sdk.values.PCollectionView
- 
- getCoderInternal() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
-  
- getCoderProvider() - Static method in class org.apache.beam.sdk.coders.SerializableCoder
- 
- getCoderProvider() - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
- getCoderProvider() - Static method in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
- 
- getCoderProvider() - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
- 
- getCoderProvider() - Static method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteResultCoder
-  
- getCoderProvider() - Static method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.OffsetByteRangeCoder
-  
- getCoderProvider() - Static method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.SubscriptionPartitionCoder
-  
- getCoderProvider() - Static method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.UuidCoder
-  
- getCoderProvider() - Static method in class org.apache.beam.sdk.io.hadoop.WritableCoder
- 
- getCoderProvider() - Static method in class org.apache.beam.sdk.io.splunk.SplunkEventCoder
-  
- getCoderProviders() - Method in interface org.apache.beam.sdk.coders.CoderProviderRegistrar
- 
- getCoderProviders() - Method in class org.apache.beam.sdk.coders.DefaultCoder.DefaultCoderProviderRegistrar
-  
- getCoderProviders() - Method in class org.apache.beam.sdk.coders.SerializableCoder.SerializableCoderProviderRegistrar
-  
- getCoderProviders() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtobufCoderProviderRegistrar
-  
- getCoderProviders() - Method in class org.apache.beam.sdk.io.amqp.AmqpMessageCoderProviderRegistrar
-  
- getCoderProviders() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryCoderProviderRegistrar
-  
- getCoderProviders() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubCoderProviderRegistrar
-  
- getCoderProviders() - Method in class org.apache.beam.sdk.io.hadoop.WritableCoder.WritableCoderProviderRegistrar
-  
- getCoderProviders() - Method in class org.apache.beam.sdk.io.hbase.HBaseCoderProviderRegistrar
-  
- getCoderRegistry() - Method in class org.apache.beam.sdk.Pipeline
- 
- getCoderTranslators() - Method in class org.apache.beam.sdk.extensions.avro.AvroGenericCoderRegistrar
-  
- getCoderURNs() - Method in class org.apache.beam.sdk.extensions.avro.AvroGenericCoderRegistrar
-  
- getCoGbkResultSchema() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
- 
- getCohorts() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.CalcRelSplitter
- 
Returns a list of sets of expressions that should be on the same level. 
- getCollations() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
-  
- getCollection() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple.TaggedKeyedPCollection
- 
Returns the underlying PCollection of this TaggedKeyedPCollection. 
- getCollectionElementType() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
-  
- getColumnDelimiter() - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText
-  
- getColumnDelimiter() - Method in class org.apache.beam.sdk.extensions.ml.DLPInspectText
-  
- getColumnDelimiter() - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText
-  
- getColumns() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- getColumns() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration
-  
- getColumns(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerSchema
-  
- getColumns() - Method in class org.apache.beam.sdk.io.snowflake.data.SnowflakeTableSchema
-  
- getCombineFn() - Method in class org.apache.beam.sdk.extensions.sql.impl.UdafImpl
-  
- getCombineFn() - Method in class org.apache.beam.sdk.transforms.Combine.GloballyAsSingletonView
-  
- getComment() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table
-  
- getCommitDeadline() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getCommitRetrySettings() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getCommitted() - Method in class org.apache.beam.sdk.metrics.MetricResult
- 
Return the value of this metric across all successfully completed parts of the pipeline. 
- getCommittedOrNull() - Method in class org.apache.beam.sdk.metrics.MetricResult
- 
Return the value of this metric across all attempts of executing all parts of the pipeline. 
- getCommitTimestamp() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao.TransactionResult
- 
Returns the commit timestamp of the read / write transaction. 
- getCommitTimestamp() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
The timestamp at which the modifications within were committed in Cloud Spanner. 
- getComponents() - Method in class org.apache.beam.sdk.coders.AtomicCoder
- 
Returns the list of  Coders that are components of this  Coder. 
- getComponents() - Method in class org.apache.beam.sdk.coders.StructuredCoder
- 
Returns the list of  Coders that are components of this  Coder. 
- getComponents(AvroGenericCoder) - Method in class org.apache.beam.sdk.extensions.avro.AvroGenericCoderTranslator
-  
- getComponents() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
-  
- getComponents() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Path
- 
Hierarchy list of component paths making up the full path, starting with the top-level child
 component path. 
- getComponents() - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
-  
- getComponents() - Method in class org.apache.beam.sdk.values.FailsafeValueInSingleWindow.Coder
-  
- getComponents() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
-  
- getComponents() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
-  
- getComponentType() - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
Returns the component type if this type is an array type, otherwise returns null.
 
- getCompression() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
- 
- getCompression() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration
- 
- getCompression() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
-  
- getCompressionCodecName() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.ParquetConfiguration
-  
- getComputeNumShards() - Method in class org.apache.beam.sdk.io.WriteFiles
-  
- getConfig() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceProviderFromDataSourceConfiguration
-  
- getConfigProperties() - Method in class org.apache.beam.sdk.io.iceberg.IcebergCatalogConfig
-  
- getConfigProperties() - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration
-  
- getConfigProperties() - Method in class org.apache.beam.sdk.io.iceberg.SchemaTransformConfiguration
-  
- getConfiguration() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
-  
- getConfiguration() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
-  
- getConfiguration() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
-  
- getConfiguration() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
-  
- getConfiguration() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.WriteRowMutations
-  
- getConfigurationMap() - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Configuration Map Getter. 
- getConfigurationRow() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransform
-  
- getConfigurationRow() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteSchemaTransformProvider.BigQueryWriteSchemaTransform
-  
- getConfiguredLoggerFromOptions(SdkHarnessOptions) - Static method in interface org.apache.beam.sdk.options.SdkHarnessOptions
- 
Configure log manager's default log level and log level overrides from the sdk harness options,
 and return the list of configured loggers. 
- getConfluentSchemaRegistrySubject() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
-  
- getConfluentSchemaRegistryUrl() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
-  
- getConnection(InfluxDbIO.DataSourceConfiguration, boolean) - Static method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO
-  
- getConnectionInitSql() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- getConnectionInitSql() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration
-  
- getConnectionProperties() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- getConnectionProperties() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration
-  
- getConnector() - Method in enum org.apache.beam.io.debezium.Connectors
- 
Class connector to debezium. 
- getConnectStringPrefix() - Method in class org.apache.beam.sdk.extensions.sql.impl.JdbcDriver
-  
- getConnectTimeout() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getConstructorCreator(TypeDescriptor<?>, Schema, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.AutoValueUtils
- 
Try to find an accessible constructor for creating an AutoValue class. 
- getConstructorCreator(TypeDescriptor<?>, Constructor, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
-  
- getConstructorCreator(TypeDescriptor<T>, Constructor<T>, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
-  
- getConsumerConfig() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getConsumerConfigUpdates() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
-  
- getConsumerFactoryFn() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getConsumerFactoryFn() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- getConsumerPollingTimeout() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getContainerImageBaseRepository() - Method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
- 
Provides the version/tag for constructing the container image path. 
- getContent() - Method in class org.apache.beam.sdk.io.tika.ParseResult
- 
Returns the extracted text. 
- getContentEncoding() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getContentType() - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.CreateOptions
- 
The content type for the created file, eg "text/plain". 
- getContentType() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getContext() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
-  
- getContext() - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Gets the context of a plugin. 
- getContiguousSequenceRangeReevaluationFrequency() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler.OrderedProcessingGlobalSequenceHandler
- 
How frequently the combiner should reevaluate the maximum range? This parameter only affects
 the behaviour of streaming pipelines. 
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterAll
-  
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterEach
-  
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterFirst
-  
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
-  
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
-  
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterSynchronizedProcessingTime
-  
- getContinuationTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-  
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-  
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
-  
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.DefaultTrigger
-  
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.Never.NeverTrigger
-  
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
-  
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
-  
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.ReshuffleTrigger
-  
- getContinuationTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
- 
Return a trigger to use after a  GroupByKey to preserve the intention of this trigger. 
- getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
- 
- getContinuationTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.Trigger.OnceTrigger
-  
- getConversionOptions(ObjectNode) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryTableProvider
-  
- getConvertedSchemaInformation(Schema, TypeDescriptor<T>) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroSchemaInformationProvider
-  
- getConvertedSchemaInformation(Schema, TypeDescriptor<T>, SchemaRegistry) - Static method in class org.apache.beam.sdk.schemas.utils.ConvertHelpers
- 
Get the coder used for converting from an inputSchema to a given type. 
- getConvertedSchemaInformation(Schema, TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.utils.RowSchemaInformationProvider
-  
- getConvertedSchemaInformation(Schema, TypeDescriptor<T>) - Method in interface org.apache.beam.sdk.schemas.utils.SchemaInformationProvider
-  
- getConvertedType() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.DateConversion
-  
- getConvertedType() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.LocalTimestampMicrosConversion
-  
- getConvertedType() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.LocalTimestampMillisConversion
-  
- getConvertedType() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimeMicrosConversion
-  
- getConvertedType() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimeMillisConversion
-  
- getConvertedType() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimestampMicrosConversion
-  
- getConvertedType() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimestampMillisConversion
-  
- getConvertedType() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.DateConversion
-  
- getConvertedType() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimeConversion
-  
- getConvertedType() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimeMicrosConversion
-  
- getConvertedType() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimestampConversion
-  
- getConvertedType() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimestampMicrosConversion
-  
- getConvertPrimitive(Schema.FieldType, TypeDescriptor<?>, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.ConvertHelpers
- 
Returns a function to convert a Row into a primitive type. 
- getCorrelationId() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getCosmosClientBuilder() - Method in interface org.apache.beam.sdk.io.azure.cosmos.CosmosOptions
-  
- getCosmosKey() - Method in interface org.apache.beam.sdk.io.azure.cosmos.CosmosOptions
- 
The Azure Cosmos key used to perform authentication for accessing resource. 
- getCosmosServiceEndpoint() - Method in interface org.apache.beam.sdk.io.azure.cosmos.CosmosOptions
- 
The Azure Cosmos service endpoint used by the Cosmos client. 
- getCount() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSortRel
-  
- getCount() - Method in class org.apache.beam.sdk.metrics.DistributionResult
-  
- getCount() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
-  
- getCountBackoffs() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
Count invocations of BackOff.nextBackOffMillis().
 
- getCountCacheReadFailures() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
Count  Cache read failures. 
- getCountCacheReadNonNulls() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
Count associated non-null values resulting from  Cache reads. 
- getCountCacheReadNulls() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
Count associated null values resulting from  Cache reads. 
- getCountCacheReadRequests() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
Count number of attempts to read from the  Cache. 
- getCountCacheWriteFailures() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
Count  Cache write failures. 
- getCountCacheWriteRequests() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
Count number of attempts to write to the  Cache. 
- getCountCacheWriteSuccesses() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
Count  Cache write successes. 
- getCountCalls() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
- getCountEndPoint() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getCounter(MetricName) - Method in class org.apache.beam.runners.jet.metrics.JetMetricsContainer
-  
- getCounter(MetricName) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
- 
Return the  Counter that should be used for implementing the given  metricName in
 this container. 
- getCounters(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
-  
- getCounters(MetricFilter) - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.WithMetricsSupport
-  
- getCounters() - Method in class org.apache.beam.sdk.metrics.MetricQueryResults
- 
Return the metric results for the counters that matched the filter. 
- getCountFailures() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
Count failures resulting from  Call's successful  Caller invocation. 
- getCountRequests() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
Count incoming request elements processed by  Call's  DoFn. 
- getCountResponses() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
Count outgoing responses resulting from  Call's successful  Caller invocation. 
- getCountryOfResidence() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
-  
- getCountSetup() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
- getCountShouldBackoff() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
- getCountSleeps() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
Count invocations of Sleeper.sleep(long).
 
- getCountTeardown() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
- getCpu() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
-  
- getCpuRate() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
-  
- getCreatedAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata
- 
The time at which this partition was first detected and created in the metadata table. 
- getCreatedAtIndexName() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataTableNames
-  
- getCreateDisposition() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- getCreateDisposition() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.WriteBuilder.Configuration
-  
- getCreateDisposition() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceConfig
-  
- getCreateFromSnapshot() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
- 
If set, the snapshot from which the job should be created. 
- getCreateTime() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2Message
- 
Gets create time. 
- getCreator(TypeDescriptor<T>, Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Get an object creator for an AVRO-generated SpecificRecord. 
- getCredential() - Method in interface org.apache.beam.sdk.extensions.gcp.auth.CredentialFactory
-  
- getCredential() - Method in class org.apache.beam.sdk.extensions.gcp.auth.GcpCredentialFactory
- 
Returns a default GCP Credentialsor null when it fails.
 
- getCredential() - Method in class org.apache.beam.sdk.extensions.gcp.auth.NoopCredentialFactory
-  
- getCredential() - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsUserCredentialFactory
- 
- getCredentialFactoryClass() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
- 
The class of the credential factory that should be created and used to create credentials. 
- getCredentials() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getCrossProduct() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaExplodeTransformProvider.Configuration
-  
- getCsvConfiguration() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration
-  
- getCsvFormat() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTableProvider.CsvToRow
-  
- getCsvRecord() - Method in class org.apache.beam.sdk.io.csv.CsvIOParseError
- 
The CSV record associated with the caught Exception.
 
- getCurrent() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-  
- getCurrent() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
-  
- getCurrent() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
- 
Gets the current record from the delegate reader. 
- getCurrent() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.UnboundedReaderImpl
-  
- getCurrent() - Method in class org.apache.beam.sdk.io.Source.Reader
- 
- getCurrentBlock() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource.AvroReader
-  
- getCurrentBlock() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
- 
- getCurrentBlockOffset() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource.AvroReader
-  
- getCurrentBlockOffset() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
- 
Returns the largest offset such that starting to read from that offset includes the current
 block. 
- getCurrentBlockSize() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource.AvroReader
-  
- getCurrentBlockSize() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
- 
Returns the size of the current block in bytes as it is represented in the underlying file,
 if possible. 
- getCurrentBundle() - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- getCurrentBundleTimestamp() - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- getCurrentContainer() - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
- 
- getCurrentDirectory() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
-  
- getCurrentDirectory() - Method in class org.apache.beam.sdk.io.ClassLoaderFileSystem.ClassLoaderResourceId
-  
- getCurrentDirectory() - Method in interface org.apache.beam.sdk.io.fs.ResourceId
- 
Returns the ResourceIdthat represents the current directory of thisResourceId.
 
- getCurrentOffset() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
-  
- getCurrentOffset() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-  
- getCurrentOffset() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
- 
- getCurrentParent() - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
- 
Gets the parent composite transform to the current transform, if one exists. 
- getCurrentRateLimit() - Method in class org.apache.beam.sdk.io.sparkreceiver.WrappedSupervisor
-  
- getCurrentRecord() - Method in class org.apache.beam.sdk.io.BlockBasedSource.Block
- 
Returns the current record. 
- getCurrentRecordId() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
- 
Returns a unique identifier for the current record. 
- getCurrentRecordOffset() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
-  
- getCurrentRelativeTime() - Method in interface org.apache.beam.sdk.state.Timer
- 
- getCurrentRowAsStruct() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.ChangeStreamResultSet
- 
Returns the record at the current pointer as a Struct.
 
- getCurrentSchemaPlus() - Method in class org.apache.beam.sdk.extensions.sql.impl.JdbcConnection
- 
Calcite-created SchemaPluswrapper for the current schema.
 
- getCurrentSource() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-  
- getCurrentSource() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource.AvroReader
-  
- getCurrentSource() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
- 
Returns a Sourcedescribing the same input that thisReadercurrently reads
 (including items already read).
 
- getCurrentSource() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
-  
- getCurrentSource() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.UnboundedReaderImpl
-  
- getCurrentSource() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-  
- getCurrentSource() - Method in class org.apache.beam.sdk.io.Source.Reader
- 
Returns a Sourcedescribing the same input that thisReadercurrently reads
 (including items already read).
 
- getCurrentSource() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
- 
- getCurrentTimestamp() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-  
- getCurrentTimestamp() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
- 
By default, returns the minimum possible timestamp. 
- getCurrentTimestamp() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-  
- getCurrentTimestamp() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.UnboundedReaderImpl
-  
- getCurrentTimestamp() - Method in class org.apache.beam.sdk.io.Source.Reader
- 
Returns the timestamp associated with the current data item. 
- getCurrentToken() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.StreamProgress
-  
- getCurrentTransform() - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
-  
- getCurrentTransform() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- getCurrentTransform() - Method in class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- getCursor() - Method in class org.apache.beam.sdk.io.redis.RedisCursor
-  
- getCustomBeamRequirement() - Method in interface org.apache.beam.sdk.extensions.python.PythonExternalTransformOptions
-  
- getCustomerId() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Order
-  
- getCustomerProvidedKey() - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- getCustomError(HttpRequestWrapper, HttpResponseWrapper) - Method in class org.apache.beam.sdk.extensions.gcp.util.CustomHttpErrors
-  
- getCustomError() - Method in class org.apache.beam.sdk.extensions.gcp.util.CustomHttpErrors.MatcherAndError
-  
- getDanglingDataSets() - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.BatchTranslationContext
-  
- getData() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisRecord
-  
- getData() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2Message
- 
Gets data. 
- getData(Row) - Method in interface org.apache.beam.sdk.io.iceberg.DynamicDestinations
-  
- getDataAsBytes() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisRecord
-  
- getDatabase() - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration
-  
- getDatabase() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getDatabase() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceConfig
- 
Getting a Snowflake database. 
- getDatabase() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getDatabase() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getDatabaseAdminClient() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerAccessor
-  
- getDatabaseClient() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerAccessor
-  
- getDatabaseId() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
-  
- getDatabaseId() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
-  
- getDatabaseId() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-  
- getDatabaseId() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
-  
- getDatabaseId() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration
-  
- getDatabaseId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getDatabaseId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration
-  
- getDatabaseId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration
-  
- getDatabaseRole() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getDataBoostEnabled() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getDataCatalogEndpoint() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogPipelineOptions
- 
DataCatalog endpoint. 
- getDataCatalogSegments() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.SubscriptionPath
-  
- getDataCatalogSegments() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.TopicPath
- 
Returns the data catalog segments. 
- getDataClient() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.BigtableChangeStreamAccessor
-  
- getDataCoder() - Method in class org.apache.beam.runners.dataflow.DataflowRunner.StreamingPCollectionViewWriterFn
-  
- getDataflowClient() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
An instance of the Dataflow client. 
- getDataflowEndpoint() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
Dataflow endpoint to use. 
- getDataflowEndpoint() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
- 
Dataflow endpoint to use. 
- getDataflowJobFile() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
The path to write the translated Dataflow job specification out to at job submission time. 
- getDataflowKmsKey() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
- 
GCP  Cloud KMS key for Dataflow pipelines and
 buckets created by GcpTempLocationFactory. 
- getDataflowOptions() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-  
- getDataflowRunnerInfo() - Static method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
- 
- getDataflowServiceOptions() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
- 
Service options are set by the user and configure the service. 
- getDataflowWorkerJar() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- getDataResource() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HealthcareIOError
-  
- getDataSchema() - Method in interface org.apache.beam.sdk.io.iceberg.DynamicDestinations
-  
- getDataset(PCollection<T>) - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator.TranslationState
-  
- getDataset(PCollection<T>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- getDataset(String) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getDataset(String, String) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getDataset(String, Map<String, String>) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getDataset(String, String, Map<String, String>) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getDataset(String, String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
- 
Gets the specified  Dataset resource by dataset ID. 
- getDataset(String, String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.DatasetServiceImpl
- 
Gets the specified  Dataset resource by dataset ID. 
- getDataset(String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- getDataSetOrThrow(String) - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.BatchTranslationContext
-  
- getDatasetService(BigQueryOptions) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices
- 
- getDatasetService(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl
-  
- getDatasetService(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
-  
- getDataSource() - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.DataSourceConfiguration
-  
- getDataSource() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getDataSourceConfiguration() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadConfiguration
-  
- getDataSourceConfiguration() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteConfiguration
-  
- getDataSourceConfiguration() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getDataSourceProviderFn() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceConfig
- 
Getting a DataSource provider function for connection credentials. 
- getDataStreamOrThrow(String) - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext
-  
- getDataType() - Method in class org.apache.beam.sdk.io.snowflake.data.SnowflakeColumn
-  
- getDateTime() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- getDateTime(String) - Method in class org.apache.beam.sdk.values.Row
- 
- getDateTime(int) - Method in class org.apache.beam.sdk.values.Row
- 
- getDatumFactory() - Method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns the datum factory used for encoding/decoding. 
- getDatumWriterFactory(DestinationT) - Method in class org.apache.beam.sdk.extensions.avro.io.DynamicAvroDestinations
- 
- getDayOfMonth() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-  
- getDayOfMonth() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-  
- getDbSize() - Method in class org.apache.beam.sdk.io.redis.RedisCursor
-  
- getDebeziumConnectionProperties() - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration
-  
- getDecimal() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- getDecimal(String) - Method in class org.apache.beam.sdk.values.Row
- 
- getDecimal(int) - Method in class org.apache.beam.sdk.values.Row
- 
Get a BigDecimalvalue by field index,ClassCastExceptionis thrown if schema
 doesn't match.
 
- getDef() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner.NonCumulativeCostImpl
-  
- getDef() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.RelMdNodeStats
-  
- getDefault() - Static method in class org.apache.beam.runners.fnexecution.logging.Slf4jLogWriter
-  
- getDefaultCoder(TypeDescriptor<?>, CoderRegistry) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
- 
Returns the default coder for a given type descriptor. 
- getDefaultDestination() - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
- 
Returns the default destination. 
- getDefaultEnvironmentConfig() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- getDefaultEnvironmentType() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- getDefaultHeaders() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getDefaultJobName() - Static method in class org.apache.beam.runners.jobsubmission.PortablePipelineJarUtils
-  
- getDefaultOutputCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.runners.flink.FlinkStreamingAggregationsTranslators.ConcatenateAsIterable
-  
- getDefaultOutputCoder(CoderRegistry, Coder) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.TDigestQuantilesFn
-  
- getDefaultOutputCoder(CoderRegistry, Coder<InputT>) - Method in class org.apache.beam.sdk.extensions.sql.TypedCombineFnDelegate
-  
- getDefaultOutputCoder() - Method in class org.apache.beam.sdk.io.Source
- 
- getDefaultOutputCoder(CoderRegistry, Coder<Double>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
-  
- getDefaultOutputCoder(CoderRegistry, Coder<V>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
-  
- getDefaultOutputCoder(CoderRegistry, Coder<Integer>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
-  
- getDefaultOutputCoder(CoderRegistry, Coder<Long>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
-  
- getDefaultOutputCoder(CoderRegistry, Coder<InputT>) - Method in interface org.apache.beam.sdk.transforms.CombineFnBase.GlobalCombineFn
- 
Returns the Coderto use by default for outputOutputTvalues, or null if it
 is not able to be inferred.
 
- getDefaultOutputCoder() - Method in class org.apache.beam.sdk.transforms.PTransform
- 
- getDefaultOutputCoder(InputT) - Method in class org.apache.beam.sdk.transforms.PTransform
- 
- getDefaultOutputCoder(InputT, PCollection<T>) - Method in class org.apache.beam.sdk.transforms.PTransform
- 
- getDefaultOutputCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
-  
- getDefaultOverrides(boolean) - Static method in class org.apache.beam.runners.spark.SparkTransformOverrides
-  
- getDefaultPrecision(SqlTypeName) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelDataTypeSystem
-  
- getDefaultSdkHarnessLogLevel() - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
- 
This option controls the default log level of all loggers without a log level override. 
- getDefaultTimezone() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
-  
- getDefaultValue() - Method in interface org.apache.beam.sdk.values.PCollectionViews.HasDefaultValue
-  
- getDefaultValue() - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn
- 
Deprecated. Returns the default value that was specified. 
- getDefaultValue() - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn2
- 
Returns the default value that was specified. 
- getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.extensions.sql.impl.TVFSlidingWindowFn
-  
- getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-  
- getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
-  
- getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-  
- getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
- 
Return a  WindowMappingFn that returns the earliest window that contains the end of the
 main-input window. 
- getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
- 
Returns the default  WindowMappingFn to use to map main input windows to side input
 windows. 
- getDefaultWorkerLogLevel() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
- 
Deprecated. This option controls the default log level of all loggers without a log level override. 
- getDeidentifyConfig() - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText
-  
- getDeidentifyTemplateName() - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText
-  
- getDelay() - Method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform.Delay
-  
- getDelimiter() - Method in class org.apache.beam.sdk.io.csv.providers.CsvWriteTransformProvider.CsvWriteConfiguration
-  
- getDelimiters() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
-  
- getDeliveryMode() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getDeliveryMode() - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- getDependencies() - Method in class org.apache.beam.sdk.expansion.service.ExpansionServiceConfig
-  
- getDependencies(RunnerApi.FunctionSpec, PipelineOptions) - Method in interface org.apache.beam.sdk.expansion.service.TransformProvider
-  
- getDependencies(ConfigT, PipelineOptions) - Method in interface org.apache.beam.sdk.transforms.ExternalTransformBuilder
- 
List the dependencies needed for this transform. 
- getDescription() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
- 
If the field has a description, returns the description for the field. 
- getDescription() - Method in class org.apache.beam.sdk.schemas.Schema.Field
- 
Returns the field's description. 
- getDescription() - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Failure
- 
The description of what was being attempted when the failure occurred. 
- getDescriptor(String) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
-  
- getDescriptorFromTableSchema(TableSchema, boolean, boolean) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProto
-  
- getDescriptorFromTableSchema(TableSchema, boolean, boolean) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProto
- 
Given a BigQuery TableSchema, returns a protocol-buffer Descriptor that can be used to write
 data using the BigQuery Storage API. 
- getDeserializer(Map<String, ?>, boolean) - Method in class org.apache.beam.sdk.io.kafka.ConfluentSchemaRegistryDeserializerProvider
-  
- getDeserializer(Map<String, ?>, boolean) - Method in interface org.apache.beam.sdk.io.kafka.DeserializerProvider
-  
- getDesiredNumUnboundedSourceSplits() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
The desired number of initial splits for UnboundedSources. 
- getDestination() - Method in class org.apache.beam.runners.dataflow.util.PackageUtil.StagedFile
- 
Staged target for this file. 
- getDestination(String, String) - Method in interface org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService.ArtifactDestinationProvider
-  
- getDestination(UserT) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
- 
Returns an object that represents at a high level the destination being written to. 
- getDestination() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-  
- getDestination() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
- 
Return the user destination object for this writer. 
- getDestination(ValueInSingleWindow<T>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
- 
Returns an object that represents at a high level which table is being written to. 
- getDestination(ValueInSingleWindow<Row>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.PortableBigQueryDestinations
-  
- getDestination() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record
- 
Gets the destination (topic or queue) to which the message was sent. 
- getDestinationCoder() - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
- 
Returns the coder for DestinationT.
 
- getDestinationCoder() - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
- 
Returns the coder for DestinationT.
 
- getDestinationFile(boolean, FileBasedSink.DynamicDestinations<?, DestinationT, ?>, int, FileBasedSink.OutputFileHints) - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-  
- getDestinationFn() - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- getDiagnostics() - Method in exception org.apache.beam.sdk.schemas.transforms.providers.StringCompiler.CompileException
-  
- getDictionary(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getDictionary(Map<String, Object>, String, Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getDir() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.OrderKey
-  
- getDirectoryTreatment() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
-  
- getDisableAutoCommit() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- getDisableMetrics() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getDiskSizeGb() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
- 
Remote worker disk size, in gigabytes, or 0 to use the default size. 
- getDistribution(MetricName) - Method in class org.apache.beam.runners.jet.metrics.JetMetricsContainer
-  
- getDistribution() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
-  
- getDistribution(MetricName) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
- 
Return the  Distribution that should be used for implementing the given  metricName in this container. 
- getDistributions() - Method in class org.apache.beam.sdk.metrics.MetricQueryResults
- 
Return the metric results for the distributions that matched the filter. 
- getDlqTransform(String) - Static method in class org.apache.beam.sdk.schemas.io.GenericDlq
-  
- getDocToBulk() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
-  
- getDocumentCount() - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
-  
- getDoFnRunner(PipelineOptions, DoFn<InputT, OutputT>, SideInputReader, AbstractParDoP.JetOutputManager, TupleTag<OutputT>, List<TupleTag<?>>, Coder<InputT>, Map<TupleTag<?>, Coder<?>>, WindowingStrategy<?, ?>, DoFnSchemaInformation, Map<String, PCollectionView<?>>) - Method in class org.apache.beam.runners.jet.processors.ParDoP
-  
- getDoFnRunner(PipelineOptions, DoFn<KV<?, ?>, OutputT>, SideInputReader, AbstractParDoP.JetOutputManager, TupleTag<OutputT>, List<TupleTag<?>>, Coder<KV<?, ?>>, Map<TupleTag<?>, Coder<?>>, WindowingStrategy<?, ?>, DoFnSchemaInformation, Map<String, PCollectionView<?>>) - Method in class org.apache.beam.runners.jet.processors.StatefulParDoP
-  
- getDoFnSchemaInformation(DoFn<?, ?>, PCollection<?>) - Static method in class org.apache.beam.sdk.transforms.ParDo
- 
Extract information on how the DoFn uses schemas. 
- getDouble() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- getDouble(String) - Method in class org.apache.beam.sdk.values.Row
- 
- getDouble(int) - Method in class org.apache.beam.sdk.values.Row
- 
- getDriverClassName() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- getDriverClassName() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration
-  
- getDriverJars() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- getDriverJars() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration
-  
- getDrop() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- getDrop() - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration
-  
- getDrop() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider.Configuration
-  
- getDumpHeapOnOOM() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
If true, save a heap dump before killing a thread or process which is GC thrashing
 or out of memory. 
- getDuplicateCount() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus
-  
- getDynamicDestinations() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSink
-  
- getDynamicDestinations() - Method in class org.apache.beam.sdk.io.FileBasedSink
- 
- getEarliestBufferedSequence() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus
-  
- getEarliestHL7v2SendTime(String, String) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Gets the earliest HL7v2 send time. 
- getEarliestHL7v2SendTime(String, String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- getEarlyTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-  
- getElasticsearchHttpPort() - Method in interface org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOITCommon.ElasticsearchPipelineOptions
-  
- getElasticsearchServer() - Method in interface org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOITCommon.ElasticsearchPipelineOptions
-  
- getElemCoder() - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
-  
- getElement() - Method in class org.apache.beam.sdk.io.gcp.bigquery.AvroWriteRequest
-  
- getElementByteSize() - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches.BatchingParams
-  
- getElementCoders() - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
-  
- getElementConverters() - Method in class org.apache.beam.sdk.transforms.DoFnSchemaInformation
- 
The schema of the @Element parameter. 
- getElementCount() - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
- 
The number of elements after which this trigger may fire. 
- getElements() - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration.Rate
-  
- getElements() - Method in class org.apache.beam.sdk.testing.TestStream.ElementEvent
-  
- getElements() - Method in class org.apache.beam.sdk.transforms.Create.Values
-  
- getElementType() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
- 
If the field is a container type, returns the element type. 
- getEmptyMatchTreatment() - Method in class org.apache.beam.sdk.io.FileBasedSource
-  
- getEmptyMatchTreatment() - Method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
-  
- getEmptyMatchTreatment() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
-  
- getEmulatorHost() - Method in interface org.apache.beam.sdk.io.gcp.firestore.FirestoreOptions
- 
A host port pair to allow connecting to a Cloud Firestore emulator instead of the live service. 
- getEmulatorHost() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getEnableBucketReadMetricCounter() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
- 
If true, reports number of bytes read from each gcs bucket. 
- getEnableBucketWriteMetricCounter() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
- 
If true, reports number of bytes written to each gcs bucket. 
- getEnableHeapDumps() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
- 
If true and PipelineOption tempLocation is set, save a heap dump before shutting
 down the JVM due to GC thrashing or out of memory. 
- getEnableLogViaFnApi() - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
- 
This option controls whether logging will be redirected through the FnApi. 
- getEnableSparkMetricSinks() - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
-  
- getEnableStableInputDrain() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getEnableStorageReadApiV2() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getEnableWebUI() - Method in interface org.apache.beam.runners.prism.PrismPipelineOptions
-  
- getEncodedElementByteSize(RandomAccessData) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
-  
- getEncodedElementByteSize(ByteString) - Method in class org.apache.beam.runners.fnexecution.wire.ByteStringCoder
-  
- getEncodedElementByteSize(BigDecimal) - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
- 
Returns the size in bytes of the encoded value using this coder. 
- getEncodedElementByteSize(Integer) - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
- 
Returns the size in bytes of the encoded value using this coder. 
- getEncodedElementByteSize(Long) - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
- 
Returns the size in bytes of the encoded value using this coder. 
- getEncodedElementByteSize(Short) - Method in class org.apache.beam.sdk.coders.BigEndianShortCoder
- 
Returns the size in bytes of the encoded value using this coder. 
- getEncodedElementByteSize(BigInteger) - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
- 
Returns the size in bytes of the encoded value using this coder. 
- getEncodedElementByteSize(Boolean) - Method in class org.apache.beam.sdk.coders.BooleanCoder
-  
- getEncodedElementByteSize(byte[]) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
-  
- getEncodedElementByteSize(Byte) - Method in class org.apache.beam.sdk.coders.ByteCoder
- 
Returns the size in bytes of the encoded value using this coder. 
- getEncodedElementByteSize(T) - Method in class org.apache.beam.sdk.coders.Coder
- 
Returns the size in bytes of the encoded value using this coder. 
- getEncodedElementByteSize(Double) - Method in class org.apache.beam.sdk.coders.DoubleCoder
- 
Returns the size in bytes of the encoded value using this coder. 
- getEncodedElementByteSize(Float) - Method in class org.apache.beam.sdk.coders.FloatCoder
- 
Returns the size in bytes of the encoded value using this coder. 
- getEncodedElementByteSize(Instant) - Method in class org.apache.beam.sdk.coders.InstantCoder
-  
- getEncodedElementByteSize(T) - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
- 
Overridden to short-circuit the default StructuredCoderbehavior of encoding and
 counting the bytes.
 
- getEncodedElementByteSize(T) - Method in class org.apache.beam.sdk.coders.NullableCoder
- 
Overridden to short-circuit the default StructuredCoderbehavior of encoding and
 counting the bytes.
 
- getEncodedElementByteSize(String) - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
- 
Returns the size in bytes of the encoded value using this coder. 
- getEncodedElementByteSize(Integer) - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
-  
- getEncodedElementByteSize(Integer) - Method in class org.apache.beam.sdk.coders.VarIntCoder
-  
- getEncodedElementByteSize(Long) - Method in class org.apache.beam.sdk.coders.VarLongCoder
-  
- getEncodedElementByteSize(Void) - Method in class org.apache.beam.sdk.coders.VoidCoder
-  
- getEncodedElementByteSize(ByteString) - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
-  
- getEncodedElementByteSize(HyperLogLogPlus) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.HyperLogLogPlusCoder
-  
- getEncodedElementByteSize(EncodedBoundedWindow) - Method in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow.Coder
-  
- getEncodedElementByteSize(BigQueryInsertError) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertErrorCoder
-  
- getEncodedElementByteSize(TableRow) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
-  
- getEncodedElementByteSize(OffsetRange) - Method in class org.apache.beam.sdk.io.range.OffsetRange.Coder
-  
- getEncodedElementByteSizeUsingCoder(Coder<T>, T) - Static method in class org.apache.beam.sdk.coders.Coder
-  
- getEncodedRecord() - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Record
- 
Nullable to account for failing to encode, or if there is no coder for the record at the time
 of failure. 
- getEncodedTypeDescriptor() - Method in class org.apache.beam.runners.fnexecution.wire.ByteStringCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.BigEndianShortCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.ByteCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.Coder
- 
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.CollectionCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.DelegateCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.DequeCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.DoubleCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.DurationCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.FloatCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.InstantCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.IterableCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.KvCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.ListCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.MapCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.NullableCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.SerializableCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.SetCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.SortedMapCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.VarIntCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.VarLongCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.VoidCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertErrorCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.io.range.OffsetRange.Coder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.io.splunk.SplunkEventCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.io.xml.JAXBCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
-  
- getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
-  
- getEncodedWindow() - Method in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow
-  
- getEncodingPositions() - Method in class org.apache.beam.sdk.schemas.Schema
- 
Gets the encoding positions for this schema. 
- getEnd() - Method in class org.apache.beam.sdk.extensions.ordered.ContiguousSequenceRange
-  
- getEnd() - Method in class org.apache.beam.sdk.io.cassandra.RingRange
-  
- getEnd() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient.HL7v2MessagePages
-  
- getEnd() - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration
-  
- getEndAtTimestamp() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration
-  
- getEndKey() - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
- 
- getEndOffset() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
- 
Returns the specified ending offset of the source. 
- getEndOffset() - Method in interface org.apache.beam.sdk.io.sparkreceiver.HasOffset
-  
- getEndpoint() - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
- 
Endpoint used to configure AWS service clients. 
- getEndTime() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.PartitionRecord
-  
- getEndTimestamp() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata
- 
The end time for querying this given partition. 
- getEnvironment() - Method in class org.apache.beam.runners.fnexecution.environment.ProcessEnvironment
-  
- getEnvironment() - Method in interface org.apache.beam.runners.fnexecution.environment.RemoteEnvironment
- 
Return the environment that the remote handles. 
- getEnvironment() - Method in class org.apache.beam.runners.fnexecution.environment.RemoteEnvironment.SimpleRemoteEnvironment
-  
- getEnvironment() - Method in class org.apache.beam.runners.fnexecution.environment.StaticRemoteEnvironment
-  
- getEnvironment() - Method in class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- getEnvironmentCacheMillis() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- getEnvironmentExpirationMillis() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- getEnvironmentId() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- getEnvironmentOption(PortablePipelineOptions, String) - Static method in interface org.apache.beam.sdk.options.PortablePipelineOptions
- 
Return the value for the specified environment option or empty string if not present. 
- getEnvironmentOptions() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- getEquivalentFieldType(TableSchema.ColumnType) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema
- 
Returns Beam equivalent of ClickHouse column type. 
- getEquivalentSchema(TableSchema) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema
- 
Returns Beam equivalent of ClickHouse schema. 
- getError() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertError
-  
- getError() - Method in class org.apache.beam.sdk.io.solace.data.Solace.PublishResult
- 
The error details if the message could not be published. 
- getError() - Method in class org.apache.beam.sdk.io.tika.ParseResult
- 
Returns the parse error, if the file was parsed unsuccessfully. 
- getError() - Method in class org.apache.beam.sdk.schemas.io.Failure
- 
Information about the cause of the failure. 
- getErrorAsString() - Method in class org.apache.beam.sdk.io.tika.ParseResult
- 
- getErrorHandling() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- getErrorHandling() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration
-  
- getErrorHandling() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration
-  
- getErrorHandling() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration
-  
- getErrorHandling() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration
-  
- getErrorHandling() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration
-  
- getErrorHandling() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
-  
- getErrorHandling() - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration
-  
- getErrorHandling() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaFilterTransformProvider.Configuration
-  
- getErrorHandling() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider.Configuration
-  
- getErrorInfo(IOException) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.DatasetServiceImpl
-  
- getErrorMessage() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageApiInsertError
-  
- getErrorMessage() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HealthcareIOError
-  
- getErrorRowSchema(Schema) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils
-  
- getErrors() - Method in class org.apache.beam.sdk.io.csv.CsvIOParseResult
- 
- getEstimatedLowWatermark() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.StreamProgress
-  
- getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.azure.cosmos.CosmosIO.BoundedCosmosBDSource
-  
- getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.BoundedSource
- 
An estimate of the total size (in bytes) of the data that would be read from this source. 
- getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
-  
- getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
-  
- getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageTableSource
-  
- getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
-  
- getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
-  
- getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-  
- getEvent() - Method in class org.apache.beam.sdk.extensions.ordered.UnprocessedEvent
-  
- getEventCoder(Pipeline, Coder<KV<KeyT, KV<Long, EventT>>>) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler
- 
Provide the event coder. 
- getEventExaminer() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler
-  
- getEvents() - Method in class org.apache.beam.sdk.testing.TestStream
- 
- getEx() - Method in class org.apache.beam.runners.jet.processors.FlattenP.Supplier
-  
- getException() - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Failure
- 
The exception itself, e.g. 
- getExceptionStacktrace() - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Failure
- 
The full stacktrace. 
- getExecutables() - Method in interface org.apache.beam.runners.direct.ExecutableGraph
-  
- getExecutableStageIntermediateId(PipelineNode.PTransformNode) - Static method in class org.apache.beam.runners.fnexecution.translation.PipelineTranslatorUtils
-  
- getExecuteStreamingSqlRetrySettings() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getExecutionEnvironment() - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.BatchTranslationContext
-  
- getExecutionEnvironment() - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext
-  
- getExecutionModeForBatch() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getExecutionRetryDelay() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getExecutorService() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
- 
- getExpansionPort() - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver.ServerConfiguration
-  
- getExpansionServiceConfig() - Method in interface org.apache.beam.sdk.expansion.service.ExpansionServiceOptions
-  
- getExpansionServiceConfigFile() - Method in interface org.apache.beam.sdk.expansion.service.ExpansionServiceOptions
-  
- getExpectedAssertions() - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
-  
- getExpectFileToNotExist() - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.CreateOptions
- 
If true, the created file is expected to not exist. 
- getExperiments() - Method in interface org.apache.beam.sdk.options.ExperimentalOptions
-  
- getExperimentValue(PipelineOptions, String) - Static method in interface org.apache.beam.sdk.options.ExperimentalOptions
- 
Return the value for the specified experiment or null if not present. 
- getExpiration() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getExpiration() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record
- 
Gets the message expiration time in milliseconds since the Unix epoch. 
- getExplanation() - Method in class org.apache.beam.sdk.extensions.ordered.UnprocessedEvent
-  
- getExplicitHashKey(T) - Method in interface org.apache.beam.sdk.io.aws2.kinesis.KinesisPartitioner.ExplicitPartitioner
- 
Required hash value (128-bit integer) to determine explicitly the shard a record is assigned
 to based on the hash key range of each shard. 
- getExplicitHashKey(T) - Method in interface org.apache.beam.sdk.io.aws2.kinesis.KinesisPartitioner
- 
Optional hash value (128-bit integer) to determine explicitly the shard a record is assigned to
 based on the hash key range of each shard. 
- getExpression(SchemaPlus, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- getExpression() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaRowUdf.Configuration
-  
- getExtendedSequenceNumber() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisRecord
-  
- getExtensionHosts() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
-  
- getExtensionRegistry() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
- 
Returns the  ExtensionRegistry listing all known Protocol Buffers extension messages to
  T registered with this  ProtoCoder. 
- getExternalSorterType() - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
- 
Returns the external sorter type. 
- getExtraInteger() - Method in class org.apache.beam.sdk.managed.testing.TestSchemaTransformProvider.Config
-  
- getExtraString() - Method in class org.apache.beam.sdk.managed.testing.TestSchemaTransformProvider.Config
-  
- getFactory() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.AvroConvertValueForGetter
-  
- getFactory() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.AvroConvertValueForSetter
-  
- getFactory(AwsOptions) - Static method in interface org.apache.beam.sdk.io.aws2.common.ClientBuilderFactory
- 
- getFactory() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
-  
- getFactory() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
-  
- getFailedBodies() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.ExecuteBundlesResult
-  
- getFailedBodies() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write.AbstractResult
-  
- getFailedBodies() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write.Result
- 
Gets failed bodies with err. 
- getFailedBundles() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.ExecuteBundlesResult
- 
Gets failed FhirBundleResponse wrapped inside HealthcareIOError. 
- getFailedFiles() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.ExecuteBundlesResult
-  
- getFailedFiles() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write.AbstractResult
-  
- getFailedFiles() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write.Result
- 
Gets failed file imports with err. 
- getFailedInserts() - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
- 
- getFailedInsertsWithErr() - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
- 
- getFailedInsertsWithErr() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.Write.Result
-  
- getFailedLatencyMetric() - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- getFailedMessages() - Method in class org.apache.beam.sdk.io.jms.WriteJmsResult
-  
- getFailedMutations() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteResult
-  
- getFailedReads() - Method in class org.apache.beam.sdk.io.gcp.healthcare.DicomIO.ReadStudyMetadata.Result
- 
Gets failed reads. 
- getFailedReads() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Read.Result
- 
Gets failed reads. 
- getFailedReads() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIOPatientEverything.Result
- 
Gets failed reads. 
- getFailedRowsTag() - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiLoads
-  
- getFailedRowsTupleTag() - Method in class org.apache.beam.sdk.io.gcp.bigquery.StreamingWriteTables
-  
- getFailedSearches() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Search.Result
- 
Gets failed searches. 
- getFailedStorageApiInserts() - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
- 
Return any rows that persistently fail to insert when using a storage-api method. 
- getFailedToParseLines() - Method in class org.apache.beam.sdk.transforms.JsonToRow.ParseResult
- 
- getFailOnCheckpointingErrors() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getFailsafeTableRow() - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWritePayload
-  
- getFailsafeTableRowPayload() - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWritePayload
-  
- getFailsafeValue() - Method in class org.apache.beam.sdk.values.FailsafeValueInSingleWindow
- 
Returns the failsafe value of this FailsafeValueInSingleWindow.
 
- getFailure() - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord
- 
Information about why the record failed. 
- getFailureCollector() - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getFailures() - Method in class org.apache.beam.io.requestresponse.Result
-  
- getFanout() - Method in class org.apache.beam.sdk.transforms.Combine.GloballyAsSingletonView
-  
- getFasterCopy() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getFetchSize() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- getFhirBundleParameter() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirBundleResponse
- 
FhirBundleParameter represents a FHIR bundle in JSON format to be executed on a FHIR store. 
- getFhirStore() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.ExecuteBundles
-  
- getField() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPMeasure
-  
- getField() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerSchema.KeyPart
-  
- getField() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- getField(int) - Method in class org.apache.beam.sdk.schemas.Schema
- 
Return a field by index. 
- getField(String) - Method in class org.apache.beam.sdk.schemas.Schema
-  
- getFieldAccessDescriptor() - Method in class org.apache.beam.sdk.transforms.DoFnSchemaInformation
- 
Effective FieldAccessDescriptor applied by DoFn. 
- getFieldCount() - Method in class org.apache.beam.sdk.schemas.Schema
- 
Return the count of fields. 
- getFieldCount() - Method in class org.apache.beam.sdk.values.Row
- 
Return the size of data fields. 
- getFieldCount() - Method in class org.apache.beam.sdk.values.RowWithGetters
-  
- getFieldCount() - Method in class org.apache.beam.sdk.values.RowWithStorage
-  
- getFieldDescription(T) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- getFieldId() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor
-  
- getFieldName() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor
-  
- getFieldNames() - Method in class org.apache.beam.sdk.schemas.Schema
- 
Return the list of all field names. 
- getFieldOptionById(int) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
-  
- getFieldRef(CEPOperation) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPUtils
- 
a function that finds a pattern reference recursively. 
- getFieldRename() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor
-  
- getFields() - Method in class org.apache.beam.sdk.schemas.Schema
-  
- getFields() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaExplodeTransformProvider.Configuration
-  
- getFields() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider.Configuration
-  
- getFields(Class<?>) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
-  
- getFieldsAccessed() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
-  
- getFieldType(Schema, CEPOperation) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPUtils
-  
- getFieldType(OneOfType.Value) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
-  
- getFieldTypes(TypeDescriptor<T>, Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Get field types for an AVRO-generated SpecificRecord or a POJO. 
- getFieldTypes(TypeDescriptor<?>, Schema, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
-  
- getFieldTypes(TypeDescriptor<?>, Schema, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
-  
- getFileDescriptor(String) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
-  
- getFileDescriptorPath() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration
-  
- getFileDescriptorPath() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration
-  
- getFileDescriptorPath() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
-  
- getFileDescriptorPath() - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration
-  
- getFileFormat() - Method in class org.apache.beam.sdk.io.iceberg.IcebergDestination
- 
File format for created files. 
- getFileInputSplitMaxSizeMB() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getFileLocation() - Method in class org.apache.beam.sdk.io.tika.ParseResult
- 
Returns the absolute path to the input file. 
- getFilename() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
-  
- getFileName() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- getFilename() - Method in class org.apache.beam.sdk.io.ClassLoaderFileSystem.ClassLoaderResourceId
-  
- getFilename() - Method in class org.apache.beam.sdk.io.csv.CsvIOParseError
- 
The filename associated with the caught Exception.
 
- getFilename(BoundedWindow, PaneInfo, int, int, Compression) - Method in interface org.apache.beam.sdk.io.FileIO.Write.FileNaming
- 
Generates the filename. 
- getFilename() - Method in interface org.apache.beam.sdk.io.fs.ResourceId
- 
Returns the name of the file or directory denoted by this ResourceId.
 
- getFilenamePolicy(DestinationT) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
- 
- getFilenamePrefix() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration
-  
- getFilenameSuffix() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration
-  
- getFileOrPatternSpec() - Method in class org.apache.beam.sdk.io.FileBasedSource
-  
- getFileOrPatternSpecProvider() - Method in class org.apache.beam.sdk.io.FileBasedSource
-  
- getFilePattern() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTable
-  
- getFilepattern() - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformConfiguration
- 
The filepattern used to match and read files. 
- getFilePattern() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
-  
- getFilesList() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceConfig
- 
Getting list of names of staged files. 
- getFilesList() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeStreamingServiceConfig
- 
Getter for a list of staged files which are will be loaded to Snowflake. 
- getFilesToStage() - Method in interface org.apache.beam.sdk.options.FileStagingOptions
- 
List of local files to make available to workers. 
- getFileSystem() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- getFilter() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getFilterFormatFunction(boolean) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.PortableBigQueryDestinations
-  
- getFinishBundleBeforeCheckpointing() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getFinishedAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata
- 
The time at which the connector finished processing this partition. 
- getFirestoreDb() - Method in interface org.apache.beam.sdk.io.gcp.firestore.FirestoreOptions
- 
The Firestore database ID to connect to. 
- getFirestoreHost() - Method in interface org.apache.beam.sdk.io.gcp.firestore.FirestoreOptions
- 
A host port pair to allow connecting to a Cloud Firestore instead of the default live service. 
- getFirestoreProject() - Method in interface org.apache.beam.sdk.io.gcp.firestore.FirestoreOptions
- 
The Firestore project ID to connect to. 
- getFlatJsonRows(Schema) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery
- 
Loads rows from BigQuery into  Rows with given  Schema. 
- getFlexRSGoal() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
- 
This option controls Flexible Resource Scheduling mode. 
- getFlinkConfDir() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getFlinkMaster() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
- 
The url of the Flink JobManager on which to execute pipelines. 
- getFloat() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- getFloat(String) - Method in class org.apache.beam.sdk.values.Row
- 
Get a  Schema.TypeName.FLOAT value by field name,  IllegalStateException is thrown if
 schema doesn't match. 
- getFloat(int) - Method in class org.apache.beam.sdk.values.Row
- 
- getFn() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.ParDoSingle
-  
- getFn() - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
-  
- getFn() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
- 
- getFn() - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
- 
- getFn() - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
- 
- getFn() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-  
- getFn() - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
-  
- getFnApiDevContainerVersion() - Method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
- 
Provides the version/tag for dev SDK FnAPI container image. 
- getFnApiEnvironmentMajorVersion() - Method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
- 
Provides the FnAPI environment's major version number. 
- getForceSlotSharingGroup() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getForceUnalignedCheckpointEnabled() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getFormat() - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformConfiguration
- 
The format of the file(s) to read. 
- getFormat() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration
-  
- getFormat() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration
-  
- getFormat() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration
-  
- getFormat() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration
-  
- getFormat() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration
-  
- getFormat() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
-  
- getFormat() - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration
-  
- getFormatClass() - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Gets InputFormat or OutputFormat class for a plugin. 
- getFormatClass() - Method in enum org.apache.beam.sdk.io.cdap.PluginConstants.Format
-  
- getFormatClass() - Method in enum org.apache.beam.sdk.io.cdap.PluginConstants.Hadoop
-  
- getFormatName() - Method in enum org.apache.beam.sdk.io.cdap.PluginConstants.Format
-  
- getFormatProviderClass() - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Gets InputFormatProvider or OutputFormatProvider class for a plugin. 
- getFormatProviderClass() - Method in enum org.apache.beam.sdk.io.cdap.PluginConstants.FormatProvider
-  
- getFormatProviderName() - Method in enum org.apache.beam.sdk.io.cdap.PluginConstants.FormatProvider
-  
- getFqName(String, String, Iterable<String>) - Static method in class org.apache.beam.sdk.metrics.Lineage
- 
Assemble fully qualified name ( FQN). 
- getFqName(String, Iterable<String>) - Static method in class org.apache.beam.sdk.metrics.Lineage
- 
Assemble the FQN of given system, and segments. 
- getFractionConsumed() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
-  
- getFractionConsumed() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
- 
Returns a value in [0, 1] representing approximately what fraction of the  current source this reader has read so far, or  null if such an
 estimate is not available. 
- getFractionConsumed() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-  
- getFractionConsumed() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-  
- getFractionConsumed() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-  
- getFractionConsumed() - Method in interface org.apache.beam.sdk.io.range.RangeTracker
- 
- getFractionOfBlockConsumed() - Method in class org.apache.beam.sdk.io.BlockBasedSource.Block
- 
Returns the fraction of the block already consumed, if possible, as a value in [0,
 1].
 
- getFrom(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.estimator.BytesThroughputEstimator
- 
Returns the estimated throughput bytes for a specified time. 
- getFrom(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.estimator.NullThroughputEstimator
- 
Always returns 0. 
- getFrom(Timestamp) - Method in interface org.apache.beam.sdk.io.gcp.spanner.changestreams.estimator.ThroughputEstimator
- 
Returns the estimated throughput for a specified time. 
- getFrom() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRange
- 
Returns the range start timestamp (inclusive). 
- getFrom() - Method in class org.apache.beam.sdk.io.range.OffsetRange
-  
- getFromRowFunction(Class<T>) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
-  
- getFromRowFunction() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDynamicMessageSchema
-  
- getFromRowFunction() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
- 
Returns the toRow conversion function. 
- getFromRowFunction(Class<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
Retrieve the function that converts a  Row object to the specified type. 
- getFromRowFunction(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
Retrieve the function that converts a  Row object to the specified type. 
- getFromRowFunction() - Method in class org.apache.beam.sdk.values.PCollection
- 
Returns the attached schema's fromRowFunction. 
- getFromSnapshotExclusive() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getFromSnapshotInclusive() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getFromSnapshotRefExclusive() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getFromSnapshotRefInclusive() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getFullName(PTransform<?, ?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
- 
Returns the full name of the currently being translated transform. 
- getFullPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.SubscriptionPath
-  
- getFullPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.TopicPath
-  
- getFunction() - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields.Fanout
-  
- getFunction() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaRowUdf
-  
- getFunctionNames() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- getFunctions(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- getGapDuration() - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-  
- getGauge(MetricName) - Method in class org.apache.beam.runners.jet.metrics.JetMetricsContainer
-  
- getGauge(MetricName) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
- 
Return the  Gauge that should be used for implementing the given  metricName in
 this container. 
- getGauges(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
-  
- getGauges(MetricFilter) - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.WithMetricsSupport
-  
- getGauges() - Method in class org.apache.beam.sdk.metrics.MetricQueryResults
- 
Return the metric results for the gauges that matched the filter. 
- getGcloudCancelCommand(DataflowPipelineOptions, String) - Static method in class org.apache.beam.runners.dataflow.util.MonitoringUtil
-  
- getGcpCredential() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
- 
The credential instance that should be used to authenticate against GCP services. 
- getGcpOauthScopes() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
- 
- getGcpTempLocation() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
- 
A GCS path for storing temporary files in GCP. 
- getGcsEndpoint() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
- 
GCS endpoint to use. 
- getGcsHttpRequestReadTimeout() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- getGcsHttpRequestWriteTimeout() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- getGcsPerformanceMetrics() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
- 
If true, reports metrics of certain operations, such as batch copies. 
- getGcsReadCounterPrefix() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- getGcsRewriteDataOpBatchLimit() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- getGcsUploadBufferSizeBytes() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
- 
The buffer size (in bytes) to use when uploading files to GCS. 
- getGcsUtil() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
- 
The GcsUtil instance that should be used to communicate with Google Cloud Storage. 
- getGcsWriteCounterPrefix() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- getGCThrashingPercentagePerPeriod() - Method in interface org.apache.beam.sdk.options.MemoryMonitorOptions
- 
The GC thrashing threshold percentage. 
- getGenericRecordToRowFunction(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
- getGetOffsetFn() - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
- getGetReceiverArgsFromConfigFn() - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
- getGetters(TypeDescriptor<T>, Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Get generated getters for an AVRO-generated SpecificRecord or a POJO. 
- getGetters(TypeDescriptor<T>, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
- 
- getGetters(TypeDescriptor<T>, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
-  
- getGetters() - Method in class org.apache.beam.sdk.values.RowWithGetters
-  
- getGetterTarget() - Method in class org.apache.beam.sdk.values.RowWithGetters
-  
- getGlobalConfigRefreshPeriod() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getGlobalSequenceCombiner() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler.OrderedProcessingGlobalSequenceHandler
- 
Provide the global sequence combiner. 
- getGoogleAdsClientId() - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsOptions
- 
OAuth 2.0 Client ID identifying the application. 
- getGoogleAdsClientSecret() - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsOptions
- 
OAuth 2.0 Client Secret for the specified Client ID. 
- getGoogleAdsCredential() - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsOptions
- 
The credential instance that should be used to authenticate against the Google Ads API. 
- getGoogleAdsCredentialFactoryClass() - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsOptions
- 
The class of the credential factory to create credentials if none have been explicitly set. 
- getGoogleAdsDeveloperToken() - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsOptions
- 
Google Ads developer token for the user connecting to the Google Ads API. 
- getGoogleAdsEndpoint() - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsOptions
- 
Host endpoint to use for connections to the Google Ads API. 
- getGoogleAdsRefreshToken() - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsOptions
- 
OAuth 2.0 Refresh Token for the user connecting to the Google Ads API. 
- getGoogleApiTrace() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions
- 
This option enables tracing of API calls to Google services used within the Apache Beam SDK. 
- getGoogleCloudStorageReadOptions() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- getGrammarFileName() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- getGrammarFileName() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- getGroupFilesFileLoad() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
- 
Choose to use a GBK when gathering a list of files in batch FILE_LOAD. 
- getGroupingTableMaxSizeMb() - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
- 
Size (in MB) of each grouping table used to pre-combine elements. 
- getGson() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.DataChangeRecordToRow
-  
- getHadoopConfiguration() - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Gets a plugin Hadoop configuration. 
- getHasError() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Document
-  
- getHashCode() - Method in class org.apache.beam.io.debezium.KafkaSourceConsumerFn
-  
- getHdfsConfiguration() - Method in interface org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptions
-  
- getHeaderAccessor() - Static method in class org.apache.beam.sdk.fn.server.GrpcContextHeaderAccessorProvider
-  
- getHeaderColumns() - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText
-  
- getHeaderColumns() - Method in class org.apache.beam.sdk.extensions.ml.DLPInspectText
-  
- getHeaderColumns() - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText
-  
- getHeaders() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
-  
- getHeaders() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getHeartbeatMillis() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata
- 
The number of milliseconds after the stream is idle, which a heartbeat record will be emitted
 in the change stream query. 
- getHighWatermark() - Method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.SparkWatermarks
-  
- getHintMaxNumWorkers() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
- 
A hint to the QoS system for the intended max number of workers for a pipeline. 
- getHistogram(MetricName, HistogramData.BucketType) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
- 
Return the  Histogram that should be used for implementing the given  metricName
 in this container. 
- getHistograms(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
-  
- getHistograms(MetricFilter) - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.WithMetricsSupport
-  
- getHL7v2Message(String) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Gets a Hl7v2 message by its name from a Hl7v2 store. 
- getHL7v2Message() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2ReadResponse
- 
Gets hl7v2Message. 
- getHL7v2Message(String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
- 
Gets HL7v2 message. 
- getHl7v2MessageId() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2ReadParameter
- 
HL7v2MessageId string. 
- getHL7v2Store(String) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Gets an HL7v2 store. 
- getHL7v2Store(String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
- 
Gets HL7v2 store. 
- getHoldability() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- getHost() - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration
-  
- getHost() - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver.ServerConfiguration
-  
- getHost() - Method in class org.apache.beam.sdk.expansion.service.ExpansionServer
- 
- getHost() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getHttpClient() - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- getHttpClientConfiguration() - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
- 
- getHttpPipeline() - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- getHTTPReadTimeout() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getHTTPWriteTimeout() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getHumanReadableJsonRecord() - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Record
- 
The failing record, encoded as JSON. 
- getIcebergCatalog() - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration
-  
- getIcebergCatalog() - Method in class org.apache.beam.sdk.io.iceberg.SchemaTransformConfiguration
-  
- getId() - Method in interface org.apache.beam.runners.fnexecution.control.RemoteBundle
- 
Get an id used to represent this bundle. 
- getId() - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor.ActiveBundle
- 
Returns an id used to represent this bundle. 
- getId() - Method in class org.apache.beam.runners.jobsubmission.JobInvocation
-  
- getId() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
-  
- getId() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
-  
- getId() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Order
-  
- getId() - Method in interface org.apache.beam.sdk.fn.IdGenerator
-  
- getId() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.ProjectPath
-  
- getId() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.SchemaPath
-  
- getId() - Method in class org.apache.beam.sdk.values.TupleTag
- 
Returns the id of this TupleTag.
 
- getId() - Method in class org.apache.beam.sdk.values.ValueWithRecordId
-  
- getIdAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration
-  
- getIdAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
- 
Get the id attribute. 
- getIdAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
- 
Get the id attribute. 
- getIdAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration
-  
- getIdentifier() - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.TZTimeOnly
-  
- getIdentifier() - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.TZTimestamp
-  
- getIdentifier() - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.UTCTimeOnly
-  
- getIdentifier() - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.UTCTimestamp
-  
- getIdentifier() - Method in class org.apache.beam.sdk.schemas.logicaltypes.Date
-  
- getIdentifier() - Method in class org.apache.beam.sdk.schemas.logicaltypes.DateTime
-  
- getIdentifier() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
-  
- getIdentifier() - Method in class org.apache.beam.sdk.schemas.logicaltypes.MicrosInstant
-  
- getIdentifier() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
-  
- getIdentifier() - Method in class org.apache.beam.sdk.schemas.logicaltypes.PassThroughLogicalType
-  
- getIdentifier() - Method in class org.apache.beam.sdk.schemas.logicaltypes.PythonCallable
-  
- getIdentifier() - Method in class org.apache.beam.sdk.schemas.logicaltypes.SchemaLogicalType
-  
- getIdentifier() - Method in class org.apache.beam.sdk.schemas.logicaltypes.Time
-  
- getIdentifier() - Method in class org.apache.beam.sdk.schemas.logicaltypes.UuidLogicalType
-  
- getIdentifier() - Method in interface org.apache.beam.sdk.schemas.Schema.LogicalType
- 
The unique identifier for this type. 
- getIdleShutdownTimeout() - Method in interface org.apache.beam.runners.prism.PrismPipelineOptions
-  
- getImpersonateServiceAccount() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
- 
All API requests will be made as the given service account or target service account in an
 impersonation delegation chain instead of the currently selected account. 
- getImplementor() - Method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
-  
- getImplementor(boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.UdafImpl
-  
- getImplementor() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.CastFunctionImpl
-  
- getInboundObserver() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer
-  
- getIncompatibleGlobalWindowErrorMessage() - Method in class org.apache.beam.sdk.extensions.sql.TypedCombineFnDelegate
-  
- getIncompatibleGlobalWindowErrorMessage() - Method in interface org.apache.beam.sdk.transforms.CombineFnBase.GlobalCombineFn
- 
Returns the error message for not supported default values in Combine.globally(). 
- getIncompatibleGlobalWindowErrorMessage() - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
-  
- getIndex() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPFieldRef
-  
- getIndex() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.OrderKey
-  
- getIndex() - Method in class org.apache.beam.sdk.extensions.sql.impl.utils.SerializableRexInputRef
-  
- getIndex() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getIndex() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- getIndex() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration
-  
- getIndex(TupleTag<?>) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
- 
Returns the index for the given tuple tag, if the tag is present in this schema, -1 if it
 isn't. 
- getIndex() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
- 
The zero-based index of this trigger firing that produced this pane. 
- getIndexes() - Method in class org.apache.beam.sdk.extensions.sql.impl.utils.SerializableRexFieldAccess
-  
- getInferMaps() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.SchemaConversionOptions
- 
/** Controls whether to use the map or row FieldType for a TableSchema field that appears to
 represent a map (it is an array of structs containing only keyandvaluefields).
 
- getInflightWaitSeconds() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StreamAppendClient
- 
If the previous call to appendRows blocked due to flow control, returns how long the call
 blocked for. 
- getIngestManager() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeStreamingServiceConfig
- 
Getter for ingest manager which serves API to load data in streaming mode and retrieve a report
 about loaded data. 
- getInitialBackoff() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
- 
The initial backoff duration to be used before retrying a request for the first time. 
- getInitializedProducer(SolaceIO.SubmissionMode) - Method in class org.apache.beam.sdk.io.solace.broker.JcsmpSessionService
-  
- getInitializedProducer(SolaceIO.SubmissionMode) - Method in class org.apache.beam.sdk.io.solace.broker.SessionService
- 
Returns a MessageProducer object for publishing messages to Solace. 
- getInitialRestriction(Map<String, String>) - Method in class org.apache.beam.io.debezium.KafkaSourceConsumerFn
-  
- getInitialRestriction(PulsarSourceDescriptor) - Method in class org.apache.beam.sdk.io.pulsar.ReadFromPulsarDoFn
-  
- getInitialRestriction(InputT) - Method in class org.apache.beam.sdk.transforms.Watch.WatchGrowthFn
-  
- getInitialWatermarkEstimatorState(Instant) - Method in class org.apache.beam.io.debezium.KafkaSourceConsumerFn
-  
- getInitialWatermarkEstimatorState(InitialPipelineState) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.DetectNewPartitionsDoFn
-  
- getInitialWatermarkEstimatorState(PartitionRecord) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.ReadChangeStreamPartitionDoFn
-  
- getInitialWatermarkEstimatorState(PartitionMetadata) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.DetectNewPartitionsDoFn
-  
- getInitialWatermarkEstimatorState(PartitionMetadata) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.ReadChangeStreamPartitionDoFn
-  
- getInitialWatermarkEstimatorState(Instant) - Method in class org.apache.beam.sdk.io.pulsar.ReadFromPulsarDoFn
-  
- getInitialWatermarkEstimatorState(Instant) - Method in class org.apache.beam.sdk.transforms.Watch.WatchGrowthFn
-  
- getInput(PTransform<InputT, ?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
-  
- getInput() - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator.UnresolvedTranslation
-  
- getInput() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- getInput(PTransform<T, ?>) - Method in class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- getInput(RelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils
-  
- getInputDataSet(PValue) - Method in class org.apache.beam.runners.twister2.Twister2BatchTranslationContext
-  
- getInputDataSet(PValue) - Method in class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- getInputDoc() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Document
-  
- getInputFile() - Method in interface org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.WordCountOptions
-  
- getinputFormatClass() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
-  
- getinputFormatKeyClass() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
-  
- getInputFormatProvider() - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getinputFormatValueClass() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
-  
- getInputId(PipelineNode.PTransformNode) - Static method in class org.apache.beam.runners.fnexecution.translation.PipelineTranslatorUtils
-  
- getInputReceivers() - Method in interface org.apache.beam.runners.fnexecution.control.RemoteBundle
- 
Get a map of PCollection ids to  receivers which consume input elements,
 forwarding them to the remote environment. 
- getInputReceivers() - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor.ActiveBundle
- 
Get a map of PCollection ids to  receivers which consume input
 elements, forwarding them to the remote environment. 
- getInputs(PTransform<InputT, ?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
- 
Returns the input of the currently being translated transform. 
- getInputs() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- getInputs() - Method in class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- getInputSchema() - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getInputSchemas() - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getInputTVariable() - Method in class org.apache.beam.sdk.extensions.sql.TypedCombineFnDelegate
-  
- getInputType() - Method in class org.apache.beam.sdk.extensions.sql.impl.UdafImpl
-  
- getInputType() - Method in class org.apache.beam.sdk.extensions.sql.TypedCombineFnDelegate
-  
- getInputType() - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
- 
Returns a  TypeDescriptor capturing what is known statically about the input type of
 this  CombineFn instance's most-derived class. 
- getInputTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.DoFn
- 
Returns a  TypeDescriptor capturing what is known statically about the input type of
 this  DoFn instance's most-derived class. 
- getInputTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.InferableFunction
- 
- getInputValueCoder(Coder<KV<K, V>>) - Static method in class org.apache.beam.sdk.transforms.GroupByKey
- 
Returns the Coderof the values of the input to this transform.
 
- getInsertBundleParallelism() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getInsertCount() - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- getInsertDefault() - Method in class org.apache.beam.sdk.transforms.Combine.GloballyAsSingletonView
-  
- getInsertErrors() - Method in class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy.Context
-  
- getInspectConfig() - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText
-  
- getInspectConfig() - Method in class org.apache.beam.sdk.extensions.ml.DLPInspectText
-  
- getInspectConfig() - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText
-  
- getInspectTemplateName() - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText
-  
- getInspectTemplateName() - Method in class org.apache.beam.sdk.extensions.ml.DLPInspectText
-  
- getInspectTemplateName() - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText
-  
- getInstance() - Static method in class org.apache.beam.runners.spark.metrics.MetricsAccumulator
-  
- getInstance(SparkSession) - Static method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsAccumulator
- 
- getInstance() - Static method in class org.apache.beam.sdk.io.googleads.DefaultGoogleAdsClientFactory
-  
- getInstance() - Static method in class org.apache.beam.sdk.metrics.NoOpCounter
-  
- getInstance() - Static method in class org.apache.beam.sdk.metrics.NoOpGauge
-  
- getInstance() - Static method in class org.apache.beam.sdk.metrics.NoOpHistogram
-  
- getInstance(String, String) - Static method in class org.apache.beam.sdk.schemas.transforms.providers.StringCompiler
-  
- getInstanceAdminClient() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.BigtableChangeStreamAccessor
-  
- getInstanceConfigId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerAccessor
-  
- getInstanceId() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableConfig
- 
Returns the instance id being written to. 
- getInstanceId() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProvider.BigtableReadSchemaTransformConfiguration
-  
- getInstanceId() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProvider.BigtableWriteSchemaTransformConfiguration
-  
- getInstanceId() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration
-  
- getInstanceId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getInstanceId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration
-  
- getInstanceId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration
-  
- getInstructionId() - Method in class org.apache.beam.sdk.fn.data.LogicalEndpoint
-  
- getInstructionRequestHandler() - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient
-  
- getInstructionRequestHandler() - Method in interface org.apache.beam.runners.fnexecution.control.StageBundleFactory
-  
- getInstructionRequestHandler() - Method in class org.apache.beam.runners.fnexecution.environment.ProcessEnvironment
-  
- getInstructionRequestHandler() - Method in interface org.apache.beam.runners.fnexecution.environment.RemoteEnvironment
- 
- getInstructionRequestHandler() - Method in class org.apache.beam.runners.fnexecution.environment.RemoteEnvironment.SimpleRemoteEnvironment
-  
- getInstructionRequestHandler() - Method in class org.apache.beam.runners.fnexecution.environment.StaticRemoteEnvironment
-  
- getInt(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getInt(Map<String, Object>, String, Integer) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getInt16() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- getInt16(String) - Method in class org.apache.beam.sdk.values.Row
- 
Get a  Schema.TypeName.INT16 value by field name,  IllegalStateException is thrown if
 schema doesn't match. 
- getInt16(int) - Method in class org.apache.beam.sdk.values.Row
- 
- getInt32() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- getInt32(String) - Method in class org.apache.beam.sdk.values.Row
- 
Get a  Schema.TypeName.INT32 value by field name,  IllegalStateException is thrown if
 schema doesn't match. 
- getInt32(int) - Method in class org.apache.beam.sdk.values.Row
- 
- getInt64() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- getInt64(String) - Method in class org.apache.beam.sdk.values.Row
- 
Get a  Schema.TypeName.INT64 value by field name,  IllegalStateException is thrown if
 schema doesn't match. 
- getInt64(int) - Method in class org.apache.beam.sdk.values.Row
- 
- getInterface() - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-  
- getInterfaces() - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
Returns a set of  TypeDescriptors, one for each interface implemented by this class. 
- getIntersectingPartition(Range.ByteStringRange, Range.ByteStringRange) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ByteStringRangeHelper
- 
Return the overlapping parts of 2 partitions. 
- getIo() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
-  
- getIr() - Method in class org.apache.beam.sdk.extensions.sbe.SbeSchema
-  
- getIrOptions() - Method in class org.apache.beam.sdk.extensions.sbe.SbeSchema
-  
- getIsLocalChannelProvider() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getIsWindmillServiceDirectPathEnabled() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getIterable(String) - Method in class org.apache.beam.sdk.values.Row
- 
Get an iterable value by field name, IllegalStateExceptionis thrown if schema doesn't
 match.
 
- getIterable(int) - Method in class org.apache.beam.sdk.values.Row
- 
Get an iterable value by field index, IllegalStateExceptionis thrown if schema doesn't
 match.
 
- getIterableComponentType(TypeDescriptor) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
- 
For an array T[] or a subclass of Iterable, return a TypeDescriptor describing T. 
- getJarPath() - Method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
- 
Optional Beam filesystem path to the jar containing the bytecode for this function. 
- getJavaClass(RelDataType) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamJavaTypeFactory
-  
- getJavaClassLookupAllowlist() - Method in interface org.apache.beam.sdk.expansion.service.ExpansionServiceOptions
-  
- getJavaClassLookupAllowlistFile() - Method in interface org.apache.beam.sdk.expansion.service.ExpansionServiceOptions
-  
- getJAXBClass() - Method in class org.apache.beam.sdk.io.xml.JAXBCoder
-  
- getJdbcType() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- getJdbcType() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration
-  
- getJdbcUrl() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- getJdbcUrl() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration
-  
- getJdkAddOpenModules() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
- 
Open modules needed for reflection that access JDK internals with Java 9+ 
- getJdkAddOpenModules() - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
- 
Open modules needed for reflection that access JDK internals with Java 9+ 
- getJetDefaultParallelism() - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
-  
- getJetLocalMode() - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
-  
- getJetProcessorsCooperative() - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
-  
- getJetServers() - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
-  
- getJfrRecordingDurationSec() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- getJmsCorrelationID() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-  
- getJmsDeliveryMode() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-  
- getJmsDestination() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-  
- getJmsExpiration() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-  
- getJmsMessageID() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-  
- getJmsPriority() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-  
- getJmsRedelivered() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-  
- getJmsReplyTo() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-  
- getJmsTimestamp() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-  
- getJmsType() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-  
- getJob(String) - Method in class org.apache.beam.runners.dataflow.DataflowClient
- 
Gets the Dataflow  Job with the given  jobId. 
- getJob() - Method in exception org.apache.beam.runners.dataflow.DataflowJobException
- 
Returns the failed job. 
- getJob() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator.JobSpecification
-  
- getJob(JobReference) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.JobService
- 
- getJob(JobReference) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
-  
- getJobCheckIntervalInSecs() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getJobEndpoint() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- getJobFileZip() - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- getJobId() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
- 
Get the id of this job. 
- getJobId() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
- 
The identity of the Dataflow job. 
- getJobId() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
-  
- getJobInfo() - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.BatchTranslationContext
-  
- getJobInfo() - Method in interface org.apache.beam.runners.flink.FlinkPortablePipelineTranslator.TranslationContext
-  
- getJobInfo() - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext
-  
- getJobLabelsMap() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getJobMessages(String, long) - Method in class org.apache.beam.runners.dataflow.util.MonitoringUtil
- 
Return job messages sorted in ascending order by timestamp. 
- getJobMetrics(String) - Method in class org.apache.beam.runners.dataflow.DataflowClient
- 
- getJobMetrics(JobApi.GetJobMetricsRequest, StreamObserver<JobApi.GetJobMetricsResponse>) - Method in class org.apache.beam.runners.jobsubmission.InMemoryJobService
-  
- getJobMetrics(JobApi.GetJobMetricsRequest, StreamObserver<JobApi.GetJobMetricsResponse>) - Method in class org.apache.beam.runners.portability.testing.TestJobService
-  
- getJobMonitoringPageURL(String, String) - Static method in class org.apache.beam.runners.dataflow.util.MonitoringUtil
- 
- getJobMonitoringPageURL(String, String, String) - Static method in class org.apache.beam.runners.dataflow.util.MonitoringUtil
-  
- getJobName() - Method in interface org.apache.beam.sdk.options.PipelineOptions
-  
- getJobs(JobApi.GetJobsRequest, StreamObserver<JobApi.GetJobsResponse>) - Method in class org.apache.beam.runners.jobsubmission.InMemoryJobService
-  
- getJobServerConfig() - Method in interface org.apache.beam.runners.portability.testing.TestPortablePipelineOptions
-  
- getJobServerDriver() - Method in interface org.apache.beam.runners.portability.testing.TestPortablePipelineOptions
-  
- getJobServerTimeout() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- getJobServerUrl() - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver
-  
- getJobService(BigQueryOptions) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices
- 
- getJobService(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl
-  
- getJobService(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
-  
- getJobType() - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- getJoinColumns(boolean, List<Pair<RexNode, RexNode>>, int, Schema) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamJoinTransforms
-  
- getJsonBytesToRowFunction(Schema) - Static method in class org.apache.beam.sdk.schemas.utils.JsonUtils
- 
- getJsonClustering() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- getJsonFactory() - Static method in class org.apache.beam.sdk.extensions.gcp.util.Transport
-  
- getJsonFactory() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- getJsonStringToRowFunction(Schema) - Static method in class org.apache.beam.sdk.schemas.utils.JsonUtils
- 
- getJsonTimePartitioning() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- getJsonToRowWithErrFn() - Method in class org.apache.beam.sdk.transforms.JsonToRow.JsonToRowWithErrFn.ParseWithError
-  
- getKeep() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- getKeep() - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration
-  
- getKeep() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaFilterTransformProvider.Configuration
-  
- getKey(Coder<K>) - Method in class org.apache.beam.runners.flink.adapter.FlinkKey
-  
- getKey() - Method in interface org.apache.beam.runners.local.Bundle
- 
Returns the key that was output in the most recent GroupByKeyin the execution of this
 bundle.
 
- getKey() - Method in class org.apache.beam.runners.local.StructuralKey
- 
- getKey() - Method in class org.apache.beam.runners.spark.stateful.SparkStateInternals
-  
- getKey() - Method in class org.apache.beam.sdk.io.aws2.s3.SSECustomerKey
-  
- getKey() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirSearchParameter
-  
- getKey() - Method in class org.apache.beam.sdk.metrics.MetricResult
-  
- getKey() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
-  
- getKey() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
- 
The key for the display item. 
- getKey() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
- 
The key for the display item. 
- getKey() - Method in class org.apache.beam.sdk.values.KV
- 
Returns the key of this  KV. 
- getKey() - Method in class org.apache.beam.sdk.values.ShardedKey
-  
- getKeyClass() - Method in enum org.apache.beam.sdk.io.cdap.PluginConstants.Hadoop
-  
- getKeyCoder() - Method in class org.apache.beam.sdk.coders.KvCoder
-  
- getKeyCoder() - Method in class org.apache.beam.sdk.coders.MapCoder
-  
- getKeyCoder() - Method in class org.apache.beam.sdk.coders.SortedMapCoder
-  
- getKeyCoder(Pipeline, Coder<KV<KeyT, KV<Long, EventT>>>) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler
- 
Provide the key coder. 
- getKeyCoder() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
-  
- getKeyCoder() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getKeyCoder(Coder<KV<K, V>>) - Static method in class org.apache.beam.sdk.transforms.GroupByKey
- 
Returns the Coderof the keys of the input to this transform, which is also used as theCoderof the keys of the output of this transform.
 
- getKeyCoder() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
- 
- getKeyDeserializerProvider() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getKeyedCollections() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
- 
- getKeyedResources() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Search.Result
- 
Gets resources with input SearchParameter key. 
- getKeyParts(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerSchema
-  
- getKeyRange() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
- 
Returns the range of keys that will be read from the table. 
- getKeys() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
-  
- getKeySerializer() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- getKeySet() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- getKeysJson() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.Mod
- 
The primary keys of this specific modification. 
- getKeystorePassword() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getKeystorePath() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getKeyTranslationFunction() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
-  
- getKeyTypeDescriptor() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
-  
- getKind() - Method in class org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters
-  
- getKind() - Method in class org.apache.beam.sdk.io.gcp.datastore.DataStoreV1SchemaIOProvider.DataStoreV1SchemaIO
-  
- getKind() - Method in interface org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.PubsubClientFactory
- 
Return the display name for this factory. 
- getKind() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier
-  
- getKind() - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields.Fanout
-  
- getKindString() - Method in class org.apache.beam.sdk.io.BoundedReadFromUnboundedSource
-  
- getKindString() - Method in class org.apache.beam.sdk.io.Read.Bounded
-  
- getKindString() - Method in class org.apache.beam.sdk.io.Read.Unbounded
-  
- getKindString() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
-  
- getKindString() - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
-  
- getKindString() - Method in class org.apache.beam.sdk.transforms.Combine.PerKeyWithHotKeyFanout
-  
- getKindString() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-  
- getKindString() - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
-  
- getKindString() - Method in class org.apache.beam.sdk.transforms.PTransform
- 
Returns the name to use by default for this PTransform(not including the names of any
 enclosingPTransforms).
 
- getKindString() - Method in class org.apache.beam.sdk.transforms.Tee
-  
- getKindString() - Method in class org.apache.beam.sdk.transforms.windowing.Window
-  
- getKindString() - Method in class org.apache.beam.sdk.values.PValueBase
- 
Returns a  String capturing the kind of this  PValueBase. 
- getKinesisIOConsumerArns() - Method in interface org.apache.beam.sdk.io.aws2.kinesis.KinesisIOOptions
- 
Used to enable / disable EFO. 
- getKmsKey() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransformConfiguration
-  
- getKmsKey() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- getKV() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
-  
- getLabel() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
- 
Retrieve the optional label for an item. 
- getLabel() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
- 
The optional label for an item. 
- getLabels() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
- 
Labels that will be applied to the billing records for this job. 
- getLabels() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2Message
- 
Gets labels. 
- getLanguage() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaFilterTransformProvider.Configuration
-  
- getLanguage() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider.Configuration
-  
- getLanguageOptions() - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
-  
- getLastContiguousSequenceRange() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus
-  
- getLastEmitted() - Method in class org.apache.beam.sdk.io.CountingSource.CounterMark
- 
Returns the last value emitted by the reader. 
- getLastFieldId() - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- getLastProcessedSequence() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus
-  
- getLastRunTimestamp() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.StreamProgress
-  
- getLastUpdated() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.NewPartition
-  
- getLastWatermarkedBatchTime() - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
-  
- getLatencyNanos() - Method in class org.apache.beam.sdk.io.solace.data.Solace.PublishResult
- 
The publishing latency in nanoseconds. 
- getLatencyTrackingInterval() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getLatestBufferedSequence() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus
-  
- getLatestHL7v2SendTime(String, String) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Gets the latest HL7v2 send time. 
- getLatestHL7v2SendTime(String, String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- getLateTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-  
- getLeaves() - Method in class org.apache.beam.runners.twister2.Twister2PipelineExecutionEnvironment
-  
- getLeaves() - Method in class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- getLegacyDevContainerVersion() - Method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
- 
Provides the version/tag for legacy SDK FnAPI container image. 
- getLegacyEnvironmentMajorVersion() - Method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
- 
Provides the legacy environment's major version number. 
- getLength() - Method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeVarchar
-  
- getLength() - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
-  
- getLength() - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedString
-  
- getLevel() - Method in class org.apache.beam.sdk.schemas.transforms.providers.LoggingTransformProvider.Configuration
-  
- getLimitCountOfSortRel() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.AbstractBeamCalcRel
-  
- getLinkUrl() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
- 
Retrieve the optional link URL for an item. 
- getLinkUrl() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
- 
The optional link URL for an item. 
- getList() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier
-  
- getListeners() - Method in interface org.apache.beam.runners.spark.SparkContextOptions
-  
- getListOfMaps(Map<String, Object>, String, List<Map<String, Object>>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getLiteralGqlQuery() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-  
- getLoadBalanceBundles() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- getLocalhost() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-  
- getLocalJobServicePortFile() - Method in interface org.apache.beam.runners.portability.testing.TestUniversalRunner.Options
- 
A file containing the job service port, since Gradle needs to know this filename statically
 to provide it in Beam testing options. 
- getLocalValue() - Method in class org.apache.beam.runners.flink.metrics.MetricsAccumulator
-  
- getLocalWindmillHostport() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getLocation() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table
-  
- getLocation() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration
-  
- getLocation() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration
-  
- getLocation() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- getLocation() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration
-  
- getLogicalStartTime() - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getLogicalType() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
-  
- getLogicalType(Class<LogicalTypeT>) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
Helper function for retrieving the concrete logical type subclass. 
- getLogicalTypeName() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.DateConversion
-  
- getLogicalTypeName() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.LocalTimestampMicrosConversion
-  
- getLogicalTypeName() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.LocalTimestampMillisConversion
-  
- getLogicalTypeName() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimeMicrosConversion
-  
- getLogicalTypeName() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimeMillisConversion
-  
- getLogicalTypeName() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimestampMicrosConversion
-  
- getLogicalTypeName() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimestampMillisConversion
-  
- getLogicalTypeName() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.DateConversion
-  
- getLogicalTypeName() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimeConversion
-  
- getLogicalTypeName() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimeMicrosConversion
-  
- getLogicalTypeName() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimestampConversion
-  
- getLogicalTypeName() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimestampMicrosConversion
-  
- getLogicalTypeValue(String, Class<T>) - Method in class org.apache.beam.sdk.values.Row
- 
Returns the Logical Type input type for this field. 
- getLogicalTypeValue(int, Class<T>) - Method in class org.apache.beam.sdk.values.Row
- 
Returns the Logical Type input type for this field. 
- getLoginTimeout() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getLoginTimeout() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getLogLevel() - Method in class org.apache.beam.sdk.schemas.transforms.providers.LoggingTransformProvider.Configuration
-  
- getLogMdc() - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
- 
Whether to include SLF4J MDC in log entries. 
- getLong(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getLong(Map<String, Object>, String, Long) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getLowWatermark() - Method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.SparkWatermarks
-  
- getLowWatermark() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.NewPartition
-  
- getMainOutputTag() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.ParDoSingle
-  
- getMainOutputTag() - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- getMainOutputTag() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-  
- getMainTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
- 
The main trigger, which will continue firing until the "until" trigger fires. 
- getManifestListLocation() - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo
-  
- getMap() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier
-  
- getMap(String) - Method in class org.apache.beam.sdk.values.Row
- 
Get a MAP value by field name, IllegalStateExceptionis thrown if schema doesn't match.
 
- getMap(int) - Method in class org.apache.beam.sdk.values.Row
- 
Get a MAP value by field index, IllegalStateExceptionis thrown if schema doesn't
 match.
 
- getMapKeyType() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
- 
If the field is a map type, returns the key type. 
- getMapKeyType() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
-  
- getMapping() - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.Globally
-  
- getMapping() - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.PerKey
-  
- getMapType(TypeDescriptor, int) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
-  
- getMapValueType() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
- 
If the field is a map type, returns the key type. 
- getMapValueType() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
-  
- getMatcher() - Method in class org.apache.beam.sdk.extensions.gcp.util.CustomHttpErrors.MatcherAndError
-  
- getMatchUpdatedFiles() - Method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
-  
- getMaterialization() - Method in class org.apache.beam.sdk.transforms.ViewFn
- 
Gets the materialization of this  ViewFn. 
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryListFromMultimapViewFn
-  
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryListViewFn
-  
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryMapFromVoidKeyViewFn
-  
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryMapViewFn
-  
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryMultimapFromVoidKeyViewFn
-  
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryMultimapViewFn
-  
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.IterableBackedListViewFn
-  
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn
- 
Deprecated. 
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn2
-  
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.ListViewFn
- 
Deprecated. 
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.ListViewFn2
-  
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.MapViewFn
- 
Deprecated. 
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.MapViewFn2
-  
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.MultimapViewFn
- 
Deprecated. 
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.MultimapViewFn2
-  
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn
- 
Deprecated. 
- getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn2
-  
- getMax() - Method in class org.apache.beam.sdk.metrics.DistributionResult
-  
- getMaxAttempts() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
- 
The maximum number of times a request will be attempted for a complete successful result. 
- getMaxBufferingDuration() - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches.BatchingParams
-  
- getMaxBufferingDurationMilliSec() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getMaxBundlesFromWindmillOutstanding() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
- 
Maximum number of bundles outstanding from windmill before the worker stops requesting. 
- getMaxBundleSize() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getMaxBundleTimeMills() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getMaxBytesFromWindmillOutstanding() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
- 
Maximum number of bytes outstanding from windmill before the worker stops requesting. 
- getMaxCacheMemoryUsage(PipelineOptions) - Method in class org.apache.beam.sdk.options.SdkHarnessOptions.DefaultMaxCacheMemoryUsageMb
-  
- getMaxCacheMemoryUsage(PipelineOptions) - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions.MaxCacheMemoryUsageMb
-  
- getMaxCacheMemoryUsageMb() - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
- 
Size (in MB) for the process wide cache within the SDK harness. 
- getMaxCacheMemoryUsageMbClass() - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
- 
An instance of this class will be used to specify the maximum amount of memory to allocate to a
 cache within an SDK harness instance. 
- getMaxCacheMemoryUsagePercent() - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
- 
Size (in % [0 - 100]) for the process wide cache within the SDK harness. 
- getMaxCommitDelay() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getMaxConnectionPoolConnections() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getMaxCumulativeBackoff() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getMaxElementCountToTriggerContinuousSequenceRangeReevaluation() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler.OrderedProcessingGlobalSequenceHandler
- 
Number of new elements to trigger the re-evaluation. 
- getMaxEndOffset(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
-  
- getMaxEndOffset(PipelineOptions) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
- 
Returns the actual ending offset of the current source. 
- getMaxInvocationHistory() - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver.ServerConfiguration
-  
- getMaxLength() - Method in class org.apache.beam.sdk.schemas.logicaltypes.VariableBytes
-  
- getMaxLength() - Method in class org.apache.beam.sdk.schemas.logicaltypes.VariableString
-  
- getMaxNumericPrecision() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelDataTypeSystem
-  
- getMaxNumericScale() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelDataTypeSystem
-  
- getMaxNumRecords() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getMaxNumWorkers() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
- 
The maximum number of workers to use for the workerpool. 
- getMaxOutputElementsPerBundle() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler
- 
Returns the maximum number of elements which will be output per each bundle. 
- getMaxParallelism() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getMaxPrecision(SqlTypeName) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelDataTypeSystem
-  
- getMaxPreviewRecords() - Method in class org.apache.beam.sdk.io.cdap.context.BatchSourceContextImpl
-  
- getMaxReadTime() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getMaxReadTimeSeconds() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
-  
- getMaxRecordsPerBatch() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- getMaxStackTraceDepthToReport() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getMaxStreamingBatchSize() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getMaxStreamingRowsToBatch() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getMD5() - Method in class org.apache.beam.sdk.io.aws2.s3.SSECustomerKey
-  
- getMean() - Method in class org.apache.beam.sdk.metrics.DistributionResult
-  
- getMean() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
-  
- getMemoryMB() - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
- 
Returns the configured size of the memory buffer. 
- getMemoryMB() - Method in class org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options
- 
Returns the configured size of the memory buffer. 
- getMessage() - Method in class org.apache.beam.io.requestresponse.ApiIOError
- 
The Exceptionmessage.
 
- getMessage() - Method in exception org.apache.beam.sdk.coders.Coder.NonDeterministicException
-  
- getMessage() - Method in class org.apache.beam.sdk.io.csv.CsvIOParseError
- 
The caught Throwable.getMessage().
 
- getMessage() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.OutgoingMessage
- 
Underlying Message. 
- getMessage() - Method in exception org.apache.beam.sdk.transforms.windowing.IncompatibleWindowException
-  
- getMessageBacklog() - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicy.PartitionContext
- 
Current backlog in messages (latest offset of the partition - last processed record offset). 
- getMessageConverter(DestinationT, BigQueryServices.DatasetService) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiDynamicDestinationsTableRow
-  
- getMessageId() - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsMessage
- 
SQS message id. 
- getMessageId() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
- 
Returns the messageId of the message populated by Cloud Pub/Sub. 
- getMessageId() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getMessageId() - Method in class org.apache.beam.sdk.io.solace.data.Solace.CorrelationKey
-  
- getMessageId() - Method in class org.apache.beam.sdk.io.solace.data.Solace.PublishResult
- 
The message id of the message that was published. 
- getMessageId() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record
- 
Gets the unique identifier of the message, a string for an application-specific message
 identifier. 
- getMessageName() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration
-  
- getMessageName() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration
-  
- getMessageName() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
-  
- getMessageName() - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration
-  
- getMessageRecord() - Method in class org.apache.beam.sdk.io.pulsar.PulsarMessage
-  
- getMessageStream(JobApi.JobMessagesRequest, StreamObserver<JobApi.JobMessagesResponse>) - Method in class org.apache.beam.runners.jobsubmission.InMemoryJobService
-  
- getMessageType() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
- 
Returns the Protocol Buffers  Message type this  ProtoCoder supports. 
- getMessageType() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2Message
- 
Gets message type. 
- getMetadata(DestinationT) - Method in class org.apache.beam.sdk.extensions.avro.io.DynamicAvroDestinations
- 
Return AVRO file metadata for a given destination. 
- getMetaData() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- getMetadata(MetadataEntity) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getMetadata(MetadataScope, MetadataEntity) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getMetadata() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
- 
Returns the MatchResult.Metadataof the file.
 
- getMetadata() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirBundleParameter
- 
String representing the metadata of the Bundle to be written. 
- getMetadata() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2ReadParameter
- 
String representing the metadata of the messageId to be read. 
- getMetadata() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2ReadResponse
- 
Gets metadata. 
- getMetadata() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.ChangeStreamResultSet
- 
Returns the gathered metadata for the change stream query so far. 
- getMetadata() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
The connector execution metadata for this record. 
- getMetadata() - Method in class org.apache.beam.sdk.io.tika.ParseResult
- 
Returns the extracted metadata. 
- getMetadata(String) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
- getMetadata() - Method in class org.apache.beam.sdk.values.PCollectionViews.ValueOrMetadata
-  
- getMetadataCoder() - Method in class org.apache.beam.sdk.io.ReadableFileCoder
-  
- getMetadataQuery() - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
-  
- getMetadataString(String) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
- getMetadataTable() - Method in interface org.apache.beam.sdk.io.gcp.spanner.SpannerIO.SpannerChangeStreamOptions
- 
Returns the name of the metadata table. 
- getMetadataTableAdminDao() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.DaoFactory
-  
- getMetadataTableDao() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.DaoFactory
-  
- getMetadataTableDebugString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.DaoFactory
-  
- getMetaStore() - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlCli
-  
- getMeters(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
-  
- getMeters(MetricFilter) - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.WithMetricsSupport
-  
- getMethod() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-  
- getMethod() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-  
- getMethod() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- getMethods(Class<?>) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
- 
Returns the list of non private/protected, non-static methods in the class, caching the
 results. 
- getMethodsMap(Class<?>) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
-  
- getMetricGaugeName(String, int) - Static method in class org.apache.beam.sdk.io.kafka.KafkaSinkMetrics
- 
Creates an MetricName based on topic name and partition id. 
- getMetricLabels() - Method in class org.apache.beam.sdk.metrics.LabeledMetricNameUtils.ParsedMetricName
-  
- getMetrics() - Method in class org.apache.beam.runners.jobsubmission.JobInvocation
-  
- getMetrics() - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getMetricsEnvironmentStateForCurrentThread() - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
- 
Returns the container holder for the current thread. 
- getMetricsGraphiteHost() - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
-  
- getMetricsGraphitePort() - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
-  
- getMetricsHttpSinkUrl() - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
-  
- getMetricsMapName(long) - Static method in class org.apache.beam.runners.jet.metrics.JetMetricsContainer
-  
- getMetricsPushPeriod() - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
-  
- getMetricsSink() - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
-  
- getMimeType() - Method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
- 
Deprecated. 
- getMimeType() - Method in interface org.apache.beam.sdk.io.FileBasedSink.OutputFileHints
- 
Returns the MIME type that should be used for the files that will hold the output data. 
- getMin() - Method in class org.apache.beam.sdk.metrics.DistributionResult
-  
- getMinBundleSize() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
- 
Returns the minimum bundle size that should be used when splitting the source into sub-sources. 
- getMinConnectionPoolConnections() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getMinCpuPlatform() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
- 
Specifies a Minimum CPU platform for VM instances. 
- getMinimumTimestamp() - Method in interface org.apache.beam.runners.local.Bundle
- 
Return the minimum timestamp among elements in this bundle. 
- getMinPauseBetweenCheckpoints() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getMinReadTimeMillis() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- getMissingPartitionsFrom(List<Range.ByteStringRange>, ByteString, ByteString) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ByteStringRangeHelper
- 
Return missing partitions within partitions that are within start and end. 
- getMissingPartitionsFromEntireKeySpace(List<Range.ByteStringRange>) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ByteStringRangeHelper
- 
Return missing partitions from the entire keyspace. 
- getMode() - Method in class org.apache.beam.sdk.io.FileBasedSource
-  
- getMode() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- getModeNames() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- getModifiableCollection() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteTable
-  
- getMods() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
The modifications within this record. 
- getModType() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
The type of operation that caused the modifications within this record. 
- getMonitoringInfos() - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
- 
Return the cumulative values for any metrics in this container as MonitoringInfos. 
- getMonthOfYear() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-  
- getMutableOutput(TupleTag<T>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- getMutationInformation() - Method in class org.apache.beam.sdk.io.gcp.bigquery.RowMutation
-  
- GetMutationsFromBeamRow() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProvider.GetMutationsFromBeamRow
-  
- getMutationType() - Method in class org.apache.beam.sdk.io.gcp.bigquery.RowMutationInformation
-  
- getName() - Method in enum org.apache.beam.io.debezium.Connectors
- 
The name of this connector class. 
- getName(int) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- getName() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
-  
- getName() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPMeasure
-  
- getName() - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-  
- getName() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table
-  
- getName() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2Message
- 
Gets name. 
- getName() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.SubscriptionPath
-  
- getName() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.TopicPath
-  
- getName() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ColumnType
- 
The name of the column. 
- getName() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerSchema.Column
-  
- getName() - Method in class org.apache.beam.sdk.io.snowflake.data.SnowflakeColumn
-  
- getName() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Destination
- 
Gets the name of the destination. 
- getName() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Queue
-  
- getName() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Topic
-  
- getName() - Method in class org.apache.beam.sdk.metrics.DelegatingCounter
-  
- getName() - Method in class org.apache.beam.sdk.metrics.DelegatingDistribution
-  
- getName() - Method in class org.apache.beam.sdk.metrics.DelegatingGauge
-  
- getName() - Method in class org.apache.beam.sdk.metrics.DelegatingHistogram
-  
- getName() - Method in interface org.apache.beam.sdk.metrics.Metric
- 
- getName() - Method in class org.apache.beam.sdk.metrics.MetricName
- 
The name of this metric. 
- getName() - Method in class org.apache.beam.sdk.metrics.MetricNameFilter
- 
- getName() - Method in class org.apache.beam.sdk.metrics.MetricResult
- 
Return the name of the metric. 
- getName() - Method in class org.apache.beam.sdk.metrics.NoOpCounter
-  
- getName() - Method in class org.apache.beam.sdk.metrics.NoOpGauge
-  
- getName() - Method in class org.apache.beam.sdk.metrics.NoOpHistogram
-  
- getName() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
- 
Returns the field name. 
- getName() - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
-  
- getName() - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedString
-  
- getName() - Method in class org.apache.beam.sdk.schemas.logicaltypes.VariableBytes
-  
- getName() - Method in class org.apache.beam.sdk.schemas.logicaltypes.VariableString
-  
- getName() - Method in class org.apache.beam.sdk.schemas.Schema.Field
- 
Returns the field name. 
- getName() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaRowUdf.Configuration
-  
- getName() - Method in class org.apache.beam.sdk.transforms.PTransform
- 
Returns the transform name. 
- getName() - Method in class org.apache.beam.sdk.values.PCollection
- 
- getName() - Method in interface org.apache.beam.sdk.values.PValue
- 
Returns the name of this  PValue. 
- getName() - Method in class org.apache.beam.sdk.values.PValueBase
- 
- getNameCount() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- getNameOverride(String, T) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- getNameOverride() - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
-  
- getNameOverride() - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
-  
- getNamespace() - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getNamespace() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-  
- getNamespace() - Method in class org.apache.beam.sdk.metrics.MetricName
- 
The namespace associated with this metric. 
- getNamespace() - Method in class org.apache.beam.sdk.metrics.MetricNameFilter
- 
- getNamespace() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
-  
- getNamespace() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
- 
The namespace for the display item. 
- getNamespace() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
- 
The namespace for the display item. 
- getNeedsAttributes() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
-  
- getNeedsMessageId() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
-  
- getNeedsOrderingKey() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
-  
- getNestedFieldsAccessed() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
-  
- getNetwork() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
- 
- getNetworkTimeout() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- getNewBigqueryClient(String) - Static method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
-  
- getNewValuesJson() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.Mod
- 
The new column values after the modification was applied. 
- getNextId() - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.UniqueIdGenerator
- 
Return a random base64 encoded 8 byte string. 
- getNextOffset() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
-  
- getNextWindow() - Method in class org.apache.beam.sdk.extensions.timeseries.FillGaps.InterpolateData
-  
- getNodeStats(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelMetadataQuery
-  
- getNodeStats() - Method in interface org.apache.beam.sdk.extensions.sql.impl.planner.NodeStatsMetadata
-  
- getNodeStats(RelNode, RelMetadataQuery) - Method in interface org.apache.beam.sdk.extensions.sql.impl.planner.NodeStatsMetadata.Handler
-  
- getNodeStats(RelNode, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.RelMdNodeStats
-  
- getNodeStats(RelNode, BeamRelMetadataQuery) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils
-  
- getNonCumulativeCost(RelNode, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner.NonCumulativeCostImpl
-  
- getNonNullPrefix() - Method in class org.apache.beam.sdk.schemas.transforms.providers.LoggingTransformProvider.Configuration
-  
- getNonSpeculativeIndex() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
- 
The zero-based index of this trigger firing among non-speculative panes. 
- getNonWildcardPrefix(String) - Static method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
- 
Returns the prefix portion of the glob that doesn't contain wildcards. 
- getNotSupported() - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTableFilter
- 
Identify parts of a predicate that are not supported by the IO push-down capabilities to be
 preserved in a CalcfollowingBeamIOSourceRel.
 
- getNotSupported() - Method in class org.apache.beam.sdk.extensions.sql.meta.DefaultTableFilter
- 
Since predicate push-down is assumed not to be supported by default - return an unchanged list
 of filters to be preserved. 
- getNotSupported() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryFilter
-  
- getNotSupported() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableFilter
-  
- getNullable() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
-  
- getNullableValue() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow
- 
Workaround for autovalue code generation, which does not allow type variables to be
 instantiated with nullable actual parameters. 
- getNullFirst() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.OrderKey
-  
- getNullParams() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamSqlUnparseContext
-  
- getNum() - Method in class org.apache.beam.runners.spark.io.ConsoleIO.Write.Unbound
-  
- getNumber() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
- 
Optionally returns the field index. 
- getNumber() - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields.Fanout
-  
- getNumber() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-  
- getNumber() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-  
- getNumber() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-  
- getNumberOfBufferedEvents() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus
-  
- getNumberOfExecutionRetries() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getNumberOfPartitionsInTransaction() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
The total number of partitions for the given transaction. 
- getNumberOfReceivedEvents() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus
-  
- getNumberOfRecordsInTransaction() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
The total number of data change records for the given transaction. 
- getNumberOfRecordsRead() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.ChangeStreamResultSetMetadata
- 
Returns the total number of records read from the change stream so far. 
- getNumberOfRecordsRead() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata
- 
The number of records read in the partition change stream query before reading this record. 
- getNumberOfWorkerHarnessThreads() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
Number of threads to use on the Dataflow worker harness. 
- getNumberOverride(int, T) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- getNumBytes() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.WriteSuccessSummary
-  
- getNumBytes() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.WriteSuccessSummary
-  
- getNumConcurrentCheckpoints() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getNumEntities(PipelineOptions, String, String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
- 
Returns Number of entities available for reading. 
- getNumExtractJobCalls() - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
-  
- getNumQuerySplits() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-  
- getNumRows(BigQueryOptions, TableReference) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
- 
It returns the number of rows for a given table. 
- getNumSampledBytesPerFile() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
-  
- getNumShards() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration
-  
- getNumShards() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- getNumShardsProvider() - Method in class org.apache.beam.sdk.io.WriteFiles
-  
- getNumStorageWriteApiStreamAppendClients() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getNumStorageWriteApiStreams() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getNumStreamingKeys() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getNumStreams() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- getNumWorkers() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
- 
Number of workers to use when executing the Dataflow job. 
- getNumWrites() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.WriteSuccessSummary
-  
- getNumWrites() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.WriteSuccessSummary
-  
- getOAuthToken() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getOauthToken() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getOauthToken() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getObject(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getObject(Map<String, Object>, String, Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getObject() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
- 
Returns the object name associated with this GCS path, or an empty string if no object is
 specified. 
- getObject(GcsPath) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
- 
- getObjectMapper() - Static method in class org.apache.beam.sdk.extensions.sql.TableUtils
-  
- getObjectReuse() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getObjects(List<GcsPath>) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
- 
- getObservedTime() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HealthcareIOError
-  
- getObservedTimestamp() - Method in class org.apache.beam.io.requestresponse.ApiIOError
- 
The observed timestamp of the error. 
- getObservedTimestamp() - Method in class org.apache.beam.sdk.io.csv.CsvIOParseError
- 
The date and time when the Exceptionoccurred.
 
- getOffset() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
-  
- getOffset() - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-  
- getOffset() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-  
- getOffset() - Method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform.AlignTo
-  
- getOffsetConsumerConfig() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getOffsetDeduplication() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getOffsetLimit() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark
-  
- getOffsetLimit() - Method in interface org.apache.beam.sdk.io.UnboundedSource.CheckpointMark
-  
- getOldValuesJson() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.Mod
- 
The old column values before the modification was applied. 
- getOnCreateMatcher() - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-  
- getOneOfSchema() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
- 
Returns the schema of the underlying  Row that is used to represent the union. 
- getOneOfTypes() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- getOneRecord(Map<String, String>) - Method in class org.apache.beam.io.debezium.KafkaSourceConsumerFn
-  
- getOnly() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- getOnly() - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration
-  
- getOnly(TupleTag<V>) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
- 
If there is a singleton value for the given tag, returns it. 
- getOnly(String) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
- 
- getOnly(TupleTag<V>, V) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
- 
If there is a singleton value for the given tag, returns it. 
- getOnly(String, V) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
- 
- getOnSuccessMatcher() - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-  
- getOnTimeBehavior() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- getOperand0() - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
-  
- getOperandList() - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlCheckConstraint
-  
- getOperandList() - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlColumnDeclaration
-  
- getOperandList() - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlCreateExternalTable
-  
- getOperandList() - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlCreateFunction
-  
- getOperands() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPCall
-  
- getOperation() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPMeasure
-  
- getOperation() - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo
-  
- getOperationMode() - Method in class org.apache.beam.runners.twister2.BeamBatchTSetEnvironment
-  
- getOperator() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPCall
-  
- getOperator() - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlCheckConstraint
-  
- getOperator() - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlColumnDeclaration
-  
- getOperator() - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlCreateFunction
-  
- getOperatorChaining() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getOptionNames() - Method in class org.apache.beam.sdk.schemas.Schema.Options
-  
- getOptions() - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator.TranslationState
-  
- getOptions() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- getOptions() - Method in class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- getOptions() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getOptions() - Method in class org.apache.beam.sdk.Pipeline
-  
- getOptions() - Method in class org.apache.beam.sdk.schemas.Schema.Field
- 
- getOptions() - Method in class org.apache.beam.sdk.schemas.Schema
-  
- getOptions() - Method in class org.apache.beam.sdk.testing.TestPipeline
-  
- getOptionsId() - Method in interface org.apache.beam.sdk.options.PipelineOptions
- 
Provides a process wide unique ID for this  PipelineOptions object, assigned at graph
 construction time. 
- getOptionsSupplier() - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator.TranslationState
-  
- getOptionsSupplier() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- getOrCreate(BigtableConfig) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.BigtableChangeStreamAccessor
- 
Create a BigtableAccess if it doesn't exist and store it in the cache for faster access. 
- getOrCreate(SpannerConfig) - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerAccessor
-  
- getOrCreateReader(PipelineOptions, CheckpointMarkT) - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
-  
- getOrCreateSession(SparkStructuredStreamingPipelineOptions) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.SparkSessionFactory
- 
- getOrDefault(K, V) - Method in interface org.apache.beam.sdk.state.MapState
- 
A deferred lookup. 
- getOrderingKey() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
- 
Returns the ordering key of the message. 
- getOrdinalPosition() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ColumnType
- 
The position of the column in the table. 
- getOrphanedNewPartitions() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.reconciler.OrphanedMetadataCleaner
- 
Returns a list of NewPartition that have been around for a while and do not overlap with any
 missing partition. 
- getOrThrowException() - Method in class org.apache.beam.sdk.io.cdap.context.FailureCollectorWrapper
-  
- getOutboundObserver() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer
-  
- getOutName(int) - Method in class org.apache.beam.sdk.values.TupleTag
- 
If this TupleTagis tagging outputoutputIndexof aPTransform, returns
 the name that should be used by default for the output.
 
- getOutput(PTransform<?, OutputT>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
-  
- getOutput() - Method in interface org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.WordCountOptions
-  
- getOutput() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- getOutput(TupleTag<T>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- getOutput(PTransform<?, T>) - Method in class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- getOutput() - Method in class org.apache.beam.sdk.io.csv.CsvIOParseResult
- 
The  TPCollection as a result of successfully parsing CSV records. 
- getOutput() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.ErrorHandling
-  
- getOutput() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration.ErrorHandling
-  
- getOutput() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration.ErrorHandling
-  
- getOutput() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteResult
-  
- getOutput() - Method in class org.apache.beam.sdk.schemas.transforms.providers.ErrorHandling
-  
- getOutput() - Method in class org.apache.beam.sdk.transforms.errorhandling.ErrorHandler.DefaultErrorHandler
-  
- getOutput() - Method in interface org.apache.beam.sdk.transforms.errorhandling.ErrorHandler
-  
- getOutput() - Method in class org.apache.beam.sdk.transforms.errorhandling.ErrorHandler.PTransformErrorHandler
-  
- getOutputCoder() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
-  
- getOutputCoder(SerializableFunction<InputT, OutputT>, Coder<InputT>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
- 
- getOutputCoder() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
-  
- getOutputCoder() - Method in class org.apache.beam.sdk.io.azure.cosmos.CosmosIO.BoundedCosmosBDSource
-  
- getOutputCoder() - Method in class org.apache.beam.sdk.io.CompressedSource
- 
Returns the delegate source's output coder. 
- getOutputCoder() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
-  
- getOutputCoder() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.UnboundedSourceImpl
-  
- getOutputCoder() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
-  
- getOutputCoder() - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
-  
- getOutputCoder() - Method in class org.apache.beam.sdk.io.solace.read.UnboundedSolaceSource
-  
- getOutputCoder() - Method in class org.apache.beam.sdk.io.Source
- 
Returns the Coderto use for the data read from this source.
 
- getOutputCoder() - Method in class org.apache.beam.sdk.io.TextSource
-  
- getOutputCoder() - Method in class org.apache.beam.sdk.io.xml.XmlSource
-  
- getOutputCoders() - Method in class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- getOutputExecutablePath() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- getOutputFile() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
-  
- getOutputFilePrefix() - Method in interface org.apache.beam.sdk.extensions.sql.example.BeamSqlDataCatalogExample.DCExamplePipelineOptions
- 
Output file prefix. 
- getOutputFormatProvider() - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getOutputId(PipelineNode.PTransformNode) - Static method in class org.apache.beam.runners.fnexecution.translation.PipelineTranslatorUtils
-  
- getOutputKvCoder(Coder<KV<K, V>>) - Static method in class org.apache.beam.sdk.transforms.GroupByKey
- 
Returns the Coderof the output of this transform.
 
- getOutputOrNull(ErrorHandling) - Static method in class org.apache.beam.sdk.schemas.transforms.providers.ErrorHandling
-  
- getOutputParallelization() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- getOutputParallelization() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadConfiguration
-  
- getOutputPortSchemas() - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getOutputs(PTransform<?, OutputT>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
- 
Returns the output of the currently being translated transform. 
- getOutputs() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- getOutputs() - Method in class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- getOutputSchema() - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getOutputSchema(Schema, FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.utils.SelectHelpers
- 
- getOutputStrategyInternal(WindowingStrategy<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.Window
- 
- getOutputStream() - Method in class org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService.ArtifactDestination
-  
- getOutputTVariable() - Method in class org.apache.beam.sdk.extensions.sql.TypedCombineFnDelegate
-  
- getOutputType() - Method in class org.apache.beam.sdk.extensions.sql.impl.UdafImpl
-  
- getOutputType() - Method in class org.apache.beam.sdk.extensions.sql.TypedCombineFnDelegate
-  
- getOutputType() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaRowUdf
-  
- getOutputType() - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
- 
Returns a  TypeDescriptor capturing what is known statically about the output type of
 this  CombineFn instance's most-derived class. 
- getOutputTypeDescriptor() - Method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides.BatchStatefulDoFn
-  
- getOutputTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.DoFn
- 
Returns a  TypeDescriptor capturing what is known statically about the output type of
 this  DoFn instance's most-derived class. 
- getOutputTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.InferableFunction
- 
- getOverlappingPartitions(List<Range.ByteStringRange>) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ByteStringRangeHelper
- 
Return a list of overlapping partitions. 
- getOverloadRatio() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
- 
The target ratio between requests sent and successful requests. 
- getOverrideWindmillBinary() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
- 
Custom windmill_main binary to use with the streaming runner. 
- getPane() - Method in class org.apache.beam.sdk.values.FailsafeValueInSingleWindow
- 
Returns the pane of this FailsafeValueInSingleWindowin its window.
 
- getPane() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow
- 
Returns the pane of this ValueInSingleWindowin its window.
 
- getPaneInfo() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-  
- getParallelism() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getParallelism() - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- getParameters() - Method in class org.apache.beam.sdk.extensions.sql.impl.UdafImpl
-  
- getParameters() - Method in class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase
- 
Returns the parameters of this function. 
- getParameters() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.CastFunctionImpl
-  
- getParent() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
- 
Returns the parent path, or nullif this path does not have a parent.
 
- getParentId() - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo
-  
- getParentLowWatermark() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.PartitionRecord
-  
- getParentPartitions() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.NewPartition
-  
- getParentPartitions() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.PartitionRecord
-  
- getParents() - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
-  
- getParentTokens() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChildPartition
- 
The unique partition identifiers of the parent partitions where this child partition originated
 from. 
- getParentTokens() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata
- 
The unique partition identifiers of the parent partitions where this child partition originated
 from. 
- getParquetConfiguration() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration
-  
- getParseFn() - Method in class org.apache.beam.sdk.io.solace.read.UnboundedSolaceSource
-  
- getParser() - Method in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
- 
Get the memoized Parser, possibly initializing it lazily.
 
- getParser() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
- 
Get the memoized Parser, possibly initializing it lazily.
 
- getPartition() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.NewPartition
-  
- getPartition() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.PartitionRecord
-  
- getPartition() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.StreamPartitionWithWatermark
-  
- getPartition(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao
- 
Fetches the partition metadata row data for the given partition token. 
- getPartition(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao.InTransactionContext
- 
Fetches the partition metadata row data for the given partition token. 
- getPartition() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
-  
- getPartition() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
-  
- getPartitionCreatedAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata
- 
The time at which this partition was first detected and created in the metadata table. 
- getPartitionEndTimestamp() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata
- 
The end time for the partition change stream query, which produced this record. 
- getPartitionKey(T) - Method in interface org.apache.beam.sdk.io.aws2.kinesis.KinesisPartitioner.ExplicitPartitioner
-  
- getPartitionKey(T) - Method in interface org.apache.beam.sdk.io.aws2.kinesis.KinesisPartitioner
- 
Determines which shard in the stream the record is assigned to. 
- getPartitionKey() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisRecord
-  
- getPartitionMetadataAdminDao() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.DaoFactory
- 
Creates and returns a singleton DAO instance for admin operations over the partition metadata
 table. 
- getPartitionMetadataDao() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.DaoFactory
- 
Creates and returns a singleton DAO instance for accessing the partition metadata table. 
- getPartitionQueryTimeout() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getPartitionReadTimeout() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getPartitionRunningAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata
- 
The time at which the connector started processing this partition. 
- getPartitions() - Method in class org.apache.beam.runners.spark.io.SourceRDD.Bounded
-  
- getPartitions() - Method in class org.apache.beam.runners.spark.io.SourceRDD.Unbounded
-  
- getPartitions() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark
-  
- getPartitionScheduledAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata
- 
The time at which this partition was scheduled to be queried. 
- getPartitionSpec() - Method in class org.apache.beam.sdk.io.iceberg.IcebergTableCreateConfig
- 
Partition spec destination, in the event that it must be dynamically created. 
- getPartitionStartTimestamp() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata
- 
The start time for the partition change stream query, which produced this record. 
- getPartitionsToReconcile(Instant, Instant) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.reconciler.PartitionReconciler
- 
For missing partitions, try to organize the mismatched parent tokens in a way to fill the
 missing partitions. 
- getPartitionToken() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata
- 
The partition token that produced this change stream record. 
- getPartitionToken() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
The unique identifier of the partition that generated this record. 
- getPartitionToken() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata
- 
Unique partition identifier, which can be used to perform a change stream query. 
- getPassword() - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration
-  
- getPassword() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getPassword() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- getPassword() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration
-  
- getPassword() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getPassword() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getPassword() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getPath() - Method in class org.apache.beam.sdk.extensions.sql.impl.TableName
- 
Table path up to the leaf table name. 
- getPath() - Method in class org.apache.beam.sdk.io.csv.providers.CsvWriteTransformProvider.CsvWriteConfiguration
-  
- getPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.ProjectPath
-  
- getPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.SchemaPath
-  
- getPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.SubscriptionPath
-  
- getPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.TopicPath
-  
- getPath() - Method in class org.apache.beam.sdk.io.json.providers.JsonWriteTransformProvider.JsonWriteConfiguration
-  
- getPath() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaRowUdf.Configuration
-  
- getPath() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
-  
- getPath() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
- 
The path for the display item within a component hierarchy. 
- getPathPrefix() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getPathValidator() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
- 
The path validator instance that should be used to validate paths. 
- getPathValidatorClass() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
- 
The class of the validator that should be created and used to validate paths. 
- getPatientCompartments() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIOPatientEverything.Result
- 
Gets the patient compartment responses for GetPatientEverything requests. 
- getPatientEverything() - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO
- 
Get the patient compartment for a FHIR Patient using the GetPatientEverything/$everything API. 
- getPatientEverything(String, Map<String, Object>, String) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Fhir get patient everything http body. 
- getPatientEverything(String, Map<String, Object>, String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- getPatternCondition() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPPattern
-  
- getPatternVar() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPPattern
-  
- getPayload(AvroGenericCoder) - Method in class org.apache.beam.sdk.extensions.avro.AvroGenericCoderTranslator
-  
- getPayload() - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWritePayload
-  
- getPayload() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
- 
Returns the main PubSub message. 
- getPayload() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-  
- getPayload() - Method in class org.apache.beam.sdk.io.mqtt.MqttRecord
-  
- getPayload() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record
- 
Gets the payload of the message as a byte array. 
- getPayload() - Method in class org.apache.beam.sdk.schemas.io.Failure
- 
Bytes containing the payload which has failed. 
- getPayload() - Method in class org.apache.beam.sdk.schemas.logicaltypes.UnknownLogicalType
-  
- getPCollection() - Method in interface org.apache.beam.runners.local.Bundle
- 
Returns the PCollection that the elements of this bundle belong to. 
- getPCollection() - Method in interface org.apache.beam.sdk.values.PCollectionView
- 
For internal use only. 
- getPCollection() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
-  
- getPCollectionInputs() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
-  
- getPCollectionInputs() - Method in interface org.apache.beam.sdk.extensions.sql.impl.rel.BeamRelNode
-  
- getPerDestinationOutputFilenames() - Method in class org.apache.beam.sdk.io.WriteFilesResult
- 
Returns a  PCollection of all output filenames generated by this  WriteFiles
 organized by user destination type. 
- getPerElementConsumers(CollectionT) - Method in interface org.apache.beam.runners.direct.ExecutableGraph
-  
- getPerElementInputs(ExecutableT) - Method in interface org.apache.beam.runners.direct.ExecutableGraph
-  
- getPeriod() - Method in class org.apache.beam.sdk.extensions.sql.impl.TVFSlidingWindowFn
- 
Amount of time between generated windows. 
- getPeriod() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-  
- getPeriod() - Method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform.AlignTo
-  
- getPeriodicStatusPageOutputDirectory() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getPerWorkerCounter(MetricName) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
- 
Return the  Counter that should be used for implementing the given per-worker {@code metricName)
 in this container. 
- getPerWorkerGauge(MetricName) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
- 
Return the  Gauge that should be used for implementing the given per-worker  metricName in this container. 
- getPerWorkerHistogram(MetricName, HistogramData.BucketType) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
- 
Return the  Histogram that should be used for implementing the given per-worker  metricName in this container. 
- getPerWorkerMetricsUpdateReportingPeriodMillis() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getPgJsonb(int) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.ChangeStreamResultSet
- 
Returns the record at the current pointer as JsonB.
 
- getPipeline() - Method in class org.apache.beam.io.requestresponse.Result
-  
- getPipeline(JobApi.GetJobPipelineRequest, StreamObserver<JobApi.GetJobPipelineResponse>) - Method in class org.apache.beam.runners.jobsubmission.InMemoryJobService
-  
- getPipeline() - Method in class org.apache.beam.runners.jobsubmission.JobInvocation
- 
Retrieve the job's pipeline. 
- getPipeline() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedEventProcessorResult
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Write.Result
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.WriteBatches.Result
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.csv.CsvIOParseResult
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.gcp.healthcare.DicomIO.ReadStudyMetadata.Result
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.ExecuteBundlesResult
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Read.Result
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Search.Result
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write.AbstractResult
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write.Result
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIOPatientEverything.Result
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.Write.Result
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteResult
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteResult
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.jms.WriteJmsResult
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.solace.write.SolaceOutput
-  
- getPipeline() - Method in class org.apache.beam.sdk.io.WriteFilesResult
-  
- getPipeline() - Method in class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
-  
- getPipeline() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
-  
- getPipeline() - Method in class org.apache.beam.sdk.transforms.JsonToRow.ParseResult
-  
- getPipeline() - Method in class org.apache.beam.sdk.transforms.WithFailures.Result
-  
- getPipeline() - Method in class org.apache.beam.sdk.values.PBegin
-  
- getPipeline() - Method in class org.apache.beam.sdk.values.PCollectionList
-  
- getPipeline() - Method in class org.apache.beam.sdk.values.PCollectionRowTuple
-  
- getPipeline() - Method in class org.apache.beam.sdk.values.PCollectionTuple
-  
- getPipeline() - Method in class org.apache.beam.sdk.values.PDone
-  
- getPipeline() - Method in interface org.apache.beam.sdk.values.PInput
- 
- getPipeline() - Method in interface org.apache.beam.sdk.values.POutput
- 
- getPipeline() - Method in class org.apache.beam.sdk.values.PValueBase
-  
- getPipelineFromClasspath(String) - Static method in class org.apache.beam.runners.jobsubmission.PortablePipelineJarUtils
-  
- getPipelineName() - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getPipelineOptions() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineRegistrar.Options
-  
- getPipelineOptions() - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
- 
Returns the configured pipeline options. 
- getPipelineOptions() - Method in class org.apache.beam.runners.direct.DirectRegistrar.Options
-  
- getPipelineOptions() - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.BatchTranslationContext
-  
- getPipelineOptions() - Method in interface org.apache.beam.runners.flink.FlinkPortablePipelineTranslator.TranslationContext
-  
- getPipelineOptions() - Method in class org.apache.beam.runners.flink.FlinkRunner
- 
For testing. 
- getPipelineOptions() - Method in class org.apache.beam.runners.flink.FlinkRunnerRegistrar.Options
-  
- getPipelineOptions() - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext
-  
- getPipelineOptions() - Method in class org.apache.beam.runners.flink.TestFlinkRunner
-  
- getPipelineOptions() - Method in class org.apache.beam.runners.jet.JetRunnerRegistrar.Options
-  
- getPipelineOptions() - Method in class org.apache.beam.runners.portability.testing.TestPortablePipelineOptions.TestPortablePipelineOptionsRegistrar
-  
- getPipelineOptions() - Method in class org.apache.beam.runners.portability.testing.TestUniversalRunner.OptionsRegistrar
-  
- getPipelineOptions() - Method in class org.apache.beam.runners.prism.PrismRegistrar.Options
-  
- getPipelineOptions() - Method in class org.apache.beam.runners.spark.SparkRunnerRegistrar.Options
-  
- getPipelineOptions() - Method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingRunnerRegistrar.Options
-  
- getPipelineOptions() - Method in class org.apache.beam.runners.twister2.Twister2RunnerRegistrar.Options
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpPipelineOptionsRegistrar
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.python.PythonExternalTransformOptionsRegistrar
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlPipelineOptionsRegistrar
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.JdbcConnection
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
-  
- getPipelineOptions() - Method in interface org.apache.beam.sdk.extensions.sql.impl.rel.BeamRelNode
- 
Perform a DFS(Depth-First-Search) to find the PipelineOptions config. 
- getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamValuesRel
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogPipelineOptionsRegistrar
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIOOptions.KinesisIOOptionsRegistrar
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.io.aws2.options.AwsPipelineOptionsRegistrar
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.io.azure.options.AzurePipelineOptionsRegistrar
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.io.gcp.common.GcpIoPipelineOptionsRegistrar
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptionsRegistrar
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.options.DefaultPipelineOptionsRegistrar
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.options.ManualDockerEnvironmentOptions.Options
-  
- getPipelineOptions() - Method in interface org.apache.beam.sdk.options.PipelineOptionsRegistrar
-  
- getPipelineOptions() - Method in class org.apache.beam.sdk.options.RemoteEnvironmentOptions.Options
-  
- getPipelineOptions() - Method in interface org.apache.beam.sdk.state.StateContext
- 
- getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.CombineWithContext.Context
- 
Returns the  PipelineOptions specified with the  PipelineRunner invoking this
  KeyedCombineFn. 
- getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.DoFn.FinishBundleContext
- 
Returns the  PipelineOptions specified with the  PipelineRunner invoking this  DoFn. 
- getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.DoFn.StartBundleContext
- 
Returns the  PipelineOptions specified with the  PipelineRunner invoking this  DoFn. 
- getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
- 
Returns the  PipelineOptions specified with the  PipelineRunner invoking this  DoFn. 
- getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.resourcehints.ResourceHintsOptions.Options
-  
- getPipelineOptionsFromClasspath(String) - Static method in class org.apache.beam.runners.jobsubmission.PortablePipelineJarUtils
-  
- getPipelinePolicy() - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- getPipelineProto() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
- 
Get the Runner API pipeline proto if available. 
- getPipelineProto() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator.JobSpecification
-  
- getPipelineRunners() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineRegistrar.Runner
-  
- getPipelineRunners() - Method in class org.apache.beam.runners.direct.DirectRegistrar.Runner
-  
- getPipelineRunners() - Method in class org.apache.beam.runners.flink.FlinkRunnerRegistrar.Runner
-  
- getPipelineRunners() - Method in class org.apache.beam.runners.jet.JetRunnerRegistrar.Runner
-  
- getPipelineRunners() - Method in class org.apache.beam.runners.portability.PortableRunnerRegistrar
-  
- getPipelineRunners() - Method in class org.apache.beam.runners.portability.testing.TestUniversalRunner.RunnerRegistrar
-  
- getPipelineRunners() - Method in class org.apache.beam.runners.prism.PrismRegistrar.Runner
-  
- getPipelineRunners() - Method in class org.apache.beam.runners.spark.SparkRunnerRegistrar.Runner
-  
- getPipelineRunners() - Method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingRunnerRegistrar.Runner
-  
- getPipelineRunners() - Method in class org.apache.beam.runners.twister2.Twister2RunnerRegistrar.Runner
-  
- getPipelineUrl() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
- 
The URL of the staged portable pipeline. 
- getPlanner() - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
-  
- getPlannerName() - Method in interface org.apache.beam.sdk.extensions.sql.impl.BeamSqlPipelineOptions
-  
- getPluginClass() - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Gets the main class of a plugin. 
- getPluginConfig() - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Gets a plugin config. 
- getPluginProperties() - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getPluginProperties(String) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getPluginType() - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Gets a plugin type. 
- getPollIntervalMillis() - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformConfiguration
- 
The time, in milliseconds, to wait before polling for new files. 
- getPort() - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration
-  
- getPort() - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver.ServerConfiguration
-  
- getPort() - Method in class org.apache.beam.sdk.expansion.service.ExpansionServer
- 
- getPort() - Method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortRead
-  
- getPort() - Method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortWrite
-  
- getPortNumber() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getPortNumber() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getPositionForFractionConsumed(double) - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
- 
Returns a position Psuch that the range[start, P)represents approximately
 the given fraction of the range[start, end).
 
- getPrecision() - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.Globally
-  
- getPrecision() - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.PerKey
-  
- getPrecision() - Method in class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeNumber
-  
- getPredefinedCsvFormat() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.CsvConfiguration
- 
See CSVFormat.Predefined#values()for a list of allowed values.
 
- getPreferGroupByKeyToHandleHugeValues() - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
-  
- getPrefix() - Method in class org.apache.beam.sdk.extensions.sql.impl.TableName
- 
First element in the path. 
- getPrefix() - Method in class org.apache.beam.sdk.schemas.transforms.providers.LoggingTransformProvider.Configuration
-  
- getPrefixedEndpoint(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getPreviousWindow() - Method in class org.apache.beam.sdk.extensions.timeseries.FillGaps.InterpolateData
-  
- getPrimary() - Method in class org.apache.beam.sdk.transforms.splittabledofn.SplitResult
- 
Returns the primary restriction. 
- getPrimaryKey() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- getPriority() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getPriority() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record
- 
Gets the priority level of the message (0-255, higher is more important). 
- getPrismLocation() - Method in interface org.apache.beam.runners.prism.PrismPipelineOptions
-  
- getPrismLogLevel() - Method in interface org.apache.beam.runners.prism.PrismPipelineOptions
-  
- getPrismVersionOverride() - Method in interface org.apache.beam.runners.prism.PrismPipelineOptions
-  
- getPrivateKey() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getPrivateKeyPassphrase() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getPrivateKeyPassphrase() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getPrivateKeyPassphrase() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getPrivateKeyPath() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getPrivateKeyPath() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getProcessBundleDescriptor(String) - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClient
-  
- getProcessBundleDescriptor(BeamFnApi.GetProcessBundleDescriptorRequest, StreamObserver<BeamFnApi.ProcessBundleDescriptor>) - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService
-  
- getProcessBundleDescriptor() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
-  
- getProcessBundleDescriptor() - Method in interface org.apache.beam.runners.fnexecution.control.StageBundleFactory
-  
- getProcessingTimeAdvance() - Method in class org.apache.beam.sdk.testing.TestStream.ProcessingTimeEvent
-  
- getProcessor(BeamFnApi.ProcessBundleDescriptor, List<RemoteInputDestination>) - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient
- 
Provides  SdkHarnessClient.BundleProcessor that is capable of processing bundles not containing timers or
 state accesses such as:
 
   Side inputs
   User state
   Remote references
  
- getProcessor(BeamFnApi.ProcessBundleDescriptor, List<RemoteInputDestination>, StateDelegator) - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient
- 
- getProcessor(BeamFnApi.ProcessBundleDescriptor, List<RemoteInputDestination>, StateDelegator, Map<String, Map<String, ProcessBundleDescriptors.TimerSpec>>) - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient
- 
Provides  SdkHarnessClient.BundleProcessor that is capable of processing bundles containing timers and
 state accesses such as:
 
   Side inputs
   User state
   Remote references
  
- getProcessWideContainer() - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
- 
- getProduced(ExecutableT) - Method in interface org.apache.beam.runners.direct.ExecutableGraph
-  
- getProducer(PValue) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
- 
Get the  AppliedPTransform that produced the provided  PValue. 
- getProducer(CollectionT) - Method in interface org.apache.beam.runners.direct.ExecutableGraph
-  
- getProducerConfig() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- getProducerConfigUpdates() - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration
-  
- getProducerFactoryFn() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- getProducersMapCardinality() - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- getProfilingAgentConfiguration() - Method in interface org.apache.beam.runners.dataflow.options.DataflowProfilingOptions
-  
- getProgress() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRangeTracker
- 
Returns the progress made within the restriction so far. 
- getProgress() - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.ReadFiles.BlockTracker
-  
- getProgress() - Method in class org.apache.beam.sdk.transforms.PeriodicSequence.OutputRangeTracker
-  
- getProgress() - Method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
-  
- getProgress() - Method in class org.apache.beam.sdk.transforms.splittabledofn.GrowableOffsetRangeTracker
-  
- getProgress() - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
-  
- getProgress() - Method in interface org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker.HasProgress
- 
A representation for the amount of known completed and known remaining work. 
- getProject() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- getProject() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
- 
Project id to use when launching jobs. 
- getProject() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
- 
Get the project path. 
- getProject() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration
-  
- getProject() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration
-  
- getProjectId() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
- 
Get the project this job exists in. 
- getProjectId() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
-  
- getProjectId() - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText
-  
- getProjectId() - Method in class org.apache.beam.sdk.extensions.ml.DLPInspectText
-  
- getProjectId() - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText
-  
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableConfig
- 
Returns the project id being written to. 
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProvider.BigtableReadSchemaTransformConfiguration
-  
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProvider.BigtableWriteSchemaTransformConfiguration
-  
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
-  
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
-  
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-  
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
-  
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.datastore.DataStoreV1SchemaIOProvider.DataStoreV1SchemaIO
-  
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.SchemaPath
-  
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration
-  
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration
-  
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration
-  
- getProjectId() - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher.TableAndQuery
-  
- getProperties() - Method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
-  
- getProperties() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- getProperties() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table
-  
- getProperties() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
-  
- getProtoBytesToRowFn(Class<T>) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoMessageSchema
-  
- getProtoBytesToRowFromSchemaFunction(String, String) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoByteUtils
-  
- getProtoBytesToRowFunction(String, String) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoByteUtils
-  
- getProvidedSparkContext() - Method in interface org.apache.beam.runners.spark.SparkContextOptions
-  
- getProviderRuntimeValues() - Method in interface org.apache.beam.sdk.testing.TestPipeline.TestValueProviderOptions
-  
- getProvisionInfo(ProvisionApi.GetProvisionInfoRequest, StreamObserver<ProvisionApi.GetProvisionInfoResponse>) - Method in class org.apache.beam.runners.fnexecution.provisioning.StaticGrpcProvisionService
-  
- getProxyConfiguration() - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
- 
- getPTransformForInput() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaCSVTable
-  
- getPTransformForInput() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
-  
- getPTransformForInput() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.PayloadSerializerKafkaTable
-  
- getPTransformForOutput() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaCSVTable
-  
- getPTransformForOutput() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
-  
- getPTransformForOutput() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.PayloadSerializerKafkaTable
-  
- getPTransformId() - Method in class org.apache.beam.runners.fnexecution.data.RemoteInputDestination
-  
- getPublished() - Method in class org.apache.beam.sdk.io.solace.data.Solace.PublishResult
- 
Whether the message was published or not. 
- getPublishedResultsQueue() - Method in class org.apache.beam.sdk.io.solace.broker.JcsmpSessionService
-  
- getPublishedResultsQueue() - Method in class org.apache.beam.sdk.io.solace.broker.SessionService
- 
Returns the  Queue instance associated with this session, with the
 asynchronously received callbacks from Solace for message publications. 
- getPublishLatencyMetric() - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- getPublishMonotonicNanos() - Method in class org.apache.beam.sdk.io.solace.data.Solace.CorrelationKey
-  
- getPublishTimestamp() - Method in class org.apache.beam.sdk.io.pulsar.PulsarMessage
-  
- getPublishTimestampFunction() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- getPubsubRootUrl() - Method in interface org.apache.beam.sdk.io.gcp.pubsub.PubsubOptions
- 
Root URL for use with the Google Cloud Pub/Sub API. 
- getQualifiers() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor
-  
- getQuantifier() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPPattern
-  
- getQueries() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirSearchParameter
-  
- getQuery() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryExportReadSchemaTransformConfiguration
- 
Configures the BigQuery read job with the SQL query. 
- getQuery() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransformConfiguration
-  
- getQuery() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-  
- getQuery() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- getQuery() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration
-  
- getQuery() - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher.TableAndQuery
-  
- getQuery() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadConfiguration
-  
- getQuery() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getQuery() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceConfig
- 
Getting a query which can be source for reading. 
- getQuery() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getQueryLocation() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryExportReadSchemaTransformConfiguration
- 
BigQuery geographic location where the query job will be executed. 
- getQueryName() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- getQueryStartedAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.ChangeStreamResultSetMetadata
- 
- getQueryStartedAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata
- 
The time that the change stream query which produced this record started. 
- getQueryString() - Method in interface org.apache.beam.sdk.extensions.sql.example.BeamSqlDataCatalogExample.DCExamplePipelineOptions
- 
SQL Query. 
- getQueue() - Method in class org.apache.beam.sdk.io.solace.broker.SessionServiceFactory
- 
Getter for the queue. 
- getQueue() - Method in class org.apache.beam.sdk.io.solace.read.UnboundedSolaceSource
-  
- getQuotationMark() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceConfig
- 
Getting a character that will surround Stringin staged CSV files.
 
- getRamMegaBytes() - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- getRange() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
- 
Returns the current range. 
- getRate() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
-  
- getRate() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
-  
- getRate() - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration
-  
- getRateLimitPolicy() - Method in interface org.apache.beam.sdk.io.aws2.kinesis.RateLimitPolicyFactory
-  
- getRateLimitPolicy() - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsV17.RateLimitPolicyFactory
-  
- getRaw(ObjectT) - Method in interface org.apache.beam.sdk.schemas.FieldValueGetter
- 
Returns the raw value of the getter before any further transformations. 
- getRawBytesToRowFunction(Schema) - Static method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider
-  
- getRawBytesToRowFunction(Schema) - Static method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformProvider
-  
- getRawPrivateKey() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getRawPrivateKey() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getRawPrivateKey() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getRawType() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
- 
Returns the raw class type. 
- getRawType() - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
Returns the  Class underlying the  Type represented by this  TypeDescriptor. 
- getRead() - Method in class org.apache.beam.io.requestresponse.Cache.Pair
-  
- getReadCounterPrefix() - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.GcsCountersOptions
-  
- getReaderCacheTimeoutSec() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
The amount of time before UnboundedReaders are considered idle and closed during streaming
 execution. 
- getReadOperation() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.ReadBuilder.Configuration
-  
- getReadQuery() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- getReadResponse() - Method in class org.apache.beam.sdk.io.gcp.healthcare.DicomIO.ReadStudyMetadata.Result
- 
Gets resources. 
- getReadTime() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisRecord
-  
- getReadTime() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-  
- getReadTimePercentage() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- getReason() - Method in exception org.apache.beam.sdk.coders.CannotProvideCoderException
-  
- getReason() - Method in class org.apache.beam.sdk.extensions.ordered.UnprocessedEvent
-  
- getReasons() - Method in exception org.apache.beam.sdk.coders.Coder.NonDeterministicException
-  
- getReceiptHandle() - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsMessage
- 
SQS receipt handle. 
- getReceiver() - Method in class org.apache.beam.runners.fnexecution.control.RemoteOutputReceiver
-  
- getReceiver() - Method in class org.apache.beam.sdk.fn.data.DataEndpoint
-  
- getReceiver() - Method in class org.apache.beam.sdk.fn.data.TimerEndpoint
-  
- getReceiver() - Method in class org.apache.beam.sdk.io.solace.broker.JcsmpSessionService
-  
- getReceiver() - Method in class org.apache.beam.sdk.io.solace.broker.SessionService
- 
Returns a MessageReceiver object for receiving messages from Solace. 
- getReceiverBuilder() - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
- getReceiverClass() - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Gets Spark Receiverclass for a CDAP plugin.
 
- getReceiveTimestamp() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record
- 
Gets the timestamp (in milliseconds since the Unix epoch) when the message was received by
 the Solace broker. 
- getRecommendedSchema() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.DateConversion
-  
- getRecommendedSchema() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.LocalTimestampMicrosConversion
-  
- getRecommendedSchema() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.LocalTimestampMillisConversion
-  
- getRecommendedSchema() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimeMicrosConversion
-  
- getRecommendedSchema() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimeMillisConversion
-  
- getRecommendedSchema() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimestampMicrosConversion
-  
- getRecommendedSchema() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimestampMillisConversion
-  
- getRecord() - Method in class org.apache.beam.sdk.io.gcp.bigquery.SchemaAndRecord
-  
- getRecord() - Method in class org.apache.beam.sdk.io.kudu.TableAndRecord
-  
- getRecord() - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord
- 
Information about the record that failed. 
- getRecordJfrOnGcThrashing() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
If true, save a JFR profile when GC thrashing is first detected. 
- getRecordReadAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.ChangeStreamResultSetMetadata
- 
- getRecordReadAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata
- 
The time at which the record was fully read. 
- getRecordSchema() - Method in class org.apache.beam.io.debezium.DebeziumIO.Read
-  
- getRecordSequence() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChildPartitionsRecord
- 
Indicates the order in which a record was put to the stream. 
- getRecordSequence() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
Indicates the order in which this record was put into the change stream in the scope of a
 partition, commit timestamp and transaction tuple. 
- getRecordStreamEndedAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.ChangeStreamResultSetMetadata
- 
Returns the timestamp at which a record finished to be streamed. 
- getRecordStreamEndedAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata
- 
The time at which the record finished streaming. 
- getRecordStreamStartedAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.ChangeStreamResultSetMetadata
- 
Returns the timestamp at which a record first started to be streamed. 
- getRecordStreamStartedAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata
- 
The time at which the record started to be streamed. 
- getRecordTimestamp() - Method in interface org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecord
- 
The timestamp associated with the record. 
- getRecordTimestamp() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata
- 
The Cloud Spanner timestamp time when this record occurred. 
- getRecordTimestamp() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChildPartitionsRecord
- 
Returns the timestamp that which this partition started being valid in Cloud Spanner. 
- getRecordTimestamp() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
The timestamp at which the modifications within were committed in Cloud Spanner. 
- getRecordTimestamp() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.HeartbeatRecord
- 
Indicates the timestamp for which the change stream query has returned all changes. 
- getRecordType() - Method in class org.apache.beam.sdk.coders.SerializableCoder
-  
- getRedelivered() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record
- 
Indicates whether the message has been redelivered due to a prior delivery failure. 
- getRedistributeNumKeys() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getReferentialConstraints() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
-  
- getRegexFromPattern(RexNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPUtils
- 
Recursively construct a regular expression from a RexNode.
 
- getRegion() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
- 
Get the region this job exists in. 
- getRegion() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
- 
The Google Compute Engine  region for
 creating Dataflow jobs. 
- getRegionFromEnvironment() - Static method in class org.apache.beam.runners.dataflow.options.DefaultGcpRegionFactory
-  
- getRegisteredOptions() - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
-  
- getReidentifyConfig() - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText
-  
- getReidentifyTemplateName() - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText
-  
- getReIterableGroupByKeyResult() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getRelList() - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
-  
- getRelTypes() - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamCalcSplittingRule
-  
- getRelTypes() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.BeamJavaUdfCalcRule
-  
- getRelTypes() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcRule
-  
- getRelTypes() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcSplittingRule
-  
- getRemoteInputDestinations() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
- 
- getRemoteOutputCoders() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
- 
- getRepeatedTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
-  
- getReplacedByJob() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
- 
- getReplacedByJob() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
-  
- getReplacementTransform(AppliedPTransform<PCollection<? extends InputT>, PCollection<OutputT>, ParDo.SingleOutput<InputT, OutputT>>) - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory
-  
- getReplacementTransform(AppliedPTransform<PCollection<? extends InputT>, PCollectionTuple, PTransform<PCollection<? extends InputT>, PCollectionTuple>>) - Method in class org.apache.beam.runners.direct.ParDoMultiOverrideFactory
-  
- getReplicationGroupMessageId() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record
- 
Gets the ID for the message within its replication group (if applicable). 
- getReplyTo() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getReplyTo() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record
- 
Gets the destination to which replies to this message should be sent. 
- getReportCheckpointDuration() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getRequestAsString() - Method in class org.apache.beam.io.requestresponse.ApiIOError
- 
The string representation of the request associated with the error. 
- getRequestTimeStamp() - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsMessage
- 
Timestamp the message was received at (in epoch millis). 
- getRequirements() - Method in class org.apache.beam.sdk.transforms.Contextful
- 
Returns the requirements needed to run the closure. 
- getResidual() - Method in class org.apache.beam.sdk.transforms.splittabledofn.SplitResult
- 
Returns the residual restriction. 
- getResourceHints() - Method in class org.apache.beam.sdk.transforms.PTransform
- 
Returns resource hints set on the transform. 
- getResourceHints() - Method in interface org.apache.beam.sdk.transforms.resourcehints.ResourceHintsOptions
-  
- getResources() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Read.Result
- 
Gets resources. 
- getResources() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Search.Result
- 
Gets resources. 
- getResourceType() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirSearchParameter
-  
- getResponse() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirBundleResponse
- 
HTTP response from the FHIR store after attempting to write the Bundle method. 
- getResponseItemJson() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Document
-  
- getResponses() - Method in class org.apache.beam.io.requestresponse.Result
-  
- getRestrictionCoder() - Method in class org.apache.beam.io.debezium.KafkaSourceConsumerFn
-  
- getRestrictionCoder() - Method in class org.apache.beam.sdk.io.pulsar.ReadFromPulsarDoFn
-  
- getRestrictionCoder() - Method in class org.apache.beam.sdk.transforms.Watch.WatchGrowthFn
-  
- getResult() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao.TransactionResult
- 
Returns the result of the transaction execution. 
- getResult() - Method in class org.apache.beam.sdk.metrics.BoundedTrieResult
-  
- getResultCoder(Pipeline) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler
- 
Provide the result coder. 
- getResultCount() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus
-  
- getResults() - Method in class org.apache.beam.sdk.transforms.JsonToRow.ParseResult
- 
- getRetainDockerContainers() - Method in interface org.apache.beam.sdk.options.ManualDockerEnvironmentOptions
-  
- getRetainExternalizedCheckpointsOnCancellation() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getRetryableCodes() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getReturnType(RelDataTypeFactory) - Method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
-  
- getReturnType(RelDataTypeFactory, SqlOperatorBinding) - Method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
-  
- getReturnType(RelDataTypeFactory) - Method in class org.apache.beam.sdk.extensions.sql.impl.UdafImpl
-  
- getRole() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getRole() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getRole() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getRoot() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- getRootCause() - Method in exception org.apache.beam.sdk.coders.CannotProvideCoderException
- 
- getRootElement() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.XmlConfiguration
-  
- getRootSchema() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- getRootTransforms() - Method in interface org.apache.beam.runners.direct.ExecutableGraph
-  
- getRoutingKey() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getRow() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertError
-  
- getRow() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageApiInsertError
-  
- getRow(String) - Method in class org.apache.beam.sdk.values.Row
- 
Get a  Schema.TypeName.ROW value by field name,  IllegalStateException is thrown if
 schema doesn't match. 
- getRow(int) - Method in class org.apache.beam.sdk.values.Row
- 
Get a  Row value by field index,  IllegalStateException is thrown if schema
 doesn't match. 
- getRowCount() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
-  
- getRowCount() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
-  
- getRowGroupSize() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.ParquetConfiguration
-  
- getRowReceiver(TupleTag<T>) - Method in interface org.apache.beam.sdk.transforms.DoFn.MultiOutputReceiver
- 
- getRowRestriction() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransformConfiguration
-  
- getRows() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
-  
- getRows() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider.TableWithRows
-  
- getRowSchema() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
-  
- getRowSelector(Schema, FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.utils.SelectHelpers
-  
- getRowSelectorOptimized(Schema, FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.utils.SelectHelpers
-  
- getRowsWritten() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteResult
- 
The number of rows written in this batch. 
- getRowToAvroBytesFunction(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
- getRowToGenericRecordFunction(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
- getRowToJsonBytesFunction(Schema) - Static method in class org.apache.beam.sdk.schemas.utils.JsonUtils
- 
- getRowToJsonStringsFunction(Schema) - Static method in class org.apache.beam.sdk.schemas.utils.JsonUtils
- 
- getRowToProtoBytes(String, String) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoByteUtils
-  
- getRowToProtoBytesFn(Class<T>) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoMessageSchema
-  
- getRowToProtoBytesFromSchema(String, String) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoByteUtils
-  
- getRowToRawBytesFunction(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider
-  
- getRowToRawBytesFunction(String) - Static method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider
-  
- getRowType(RelDataTypeFactory) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteTable
-  
- getRowType() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
The type of the primary keys and modified columns within this record. 
- getRpcPriority() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- getRule() - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
-  
- getRuleIndex() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierContext
-  
- getRuleIndex() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionComponentContext
-  
- getRuleIndex() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionContext
-  
- getRuleIndex() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.FieldSpecifierContext
-  
- getRuleIndex() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierContext
-  
- getRuleIndex() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifiedComponentContext
-  
- getRuleIndex() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifierListContext
-  
- getRuleNames() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- getRuleNames() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- getRuleSets() - Static method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRuleSets
-  
- getRunner() - Method in interface org.apache.beam.sdk.options.PipelineOptions
- 
The pipeline runner that will be used to execute the pipeline. 
- getRunningAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata
- 
The time at which the connector started processing this partition. 
- getS3ClientBuilder() - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration
- 
Builder used to create the S3Client.
 
- getS3ClientFactoryClass() - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- getS3StorageClass() - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- getS3StorageClass() - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration
- 
The AWS S3 storage class used for creating S3 objects. 
- getS3ThreadPoolSize() - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- getS3ThreadPoolSize() - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration
- 
Thread pool size, limiting the max concurrent S3 operations. 
- getS3UploadBufferSizeBytes() - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- getS3UploadBufferSizeBytes() - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration
- 
Size of S3 upload chnks. 
- getSafeFilepattern() - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformConfiguration
-  
- getSafeSchema() - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformConfiguration
-  
- getSamplePeriod() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
- 
The length of time sampled request data will be retained. 
- getSamplePeriodBucketSize() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
- 
- getSamplingStrategy() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
-  
- getSasToken() - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- getSaveHeapDumpsToGcsPath() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
CAUTION: This option implies dumpHeapOnOOM, and has similar caveats. 
- getSavepointPath() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getSaveProfilesToGcs() - Method in interface org.apache.beam.runners.dataflow.options.DataflowProfilingOptions
-  
- getSbeFields() - Method in class org.apache.beam.sdk.extensions.sbe.SbeSchema
-  
- getScale() - Method in class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeNumber
-  
- getScan() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
-  
- getScanType() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getScheduledAt() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata
- 
The time at which this partition was scheduled to be queried. 
- getScheduledExecutorService() - Method in interface org.apache.beam.sdk.options.ExecutorOptions
- 
The ScheduledExecutorServiceinstance to use to create threads, can be overridden to
 specify aScheduledExecutorServicethat is compatible with the user's environment.
 
- getSchema() - Method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns the schema used by this coder. 
- getSchema(DestinationT) - Method in class org.apache.beam.sdk.extensions.avro.io.DynamicAvroDestinations
- 
Return an AVRO schema for a given destination. 
- getSchema(Class<T>, Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
-  
- getSchema() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDynamicMessageSchema
-  
- getSchema() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- getSchema() - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTable
- 
Get the schema info of the table. 
- getSchema() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
-  
- getSchema() - Method in class org.apache.beam.sdk.extensions.sql.meta.SchemaBaseBeamTable
-  
- getSchema() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table
-  
- getSchema() - Static method in class org.apache.beam.sdk.io.contextualtextio.RecordWithMetadata
-  
- getSchema() - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformConfiguration
- 
The schema used by sources to deserialize data and create Beam Rows. 
- getSchema() - Method in class org.apache.beam.sdk.io.gcp.bigquery.AvroWriteRequest
-  
- getSchema(DestinationT) - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
- 
Returns the table schema for the destination. 
- getSchema(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.PortableBigQueryDestinations
-  
- getSchema(TableReference, BigQueryServices.DatasetService) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableSchemaCache
-  
- getSchema(PubsubClient.SchemaPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
- 
Return a Beam  Schema from the Pub/Sub schema resource, if exists. 
- getSchema(PubsubClient.SchemaPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
- 
Return a Beam  Schema from the Pub/Sub schema resource, if exists. 
- getSchema(PubsubClient.SchemaPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
- 
Return a Beam  Schema from the Pub/Sub schema resource, if exists. 
- getSchema() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration
-  
- getSchema(PubsubClient.SchemaPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- getSchema() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration
-  
- getSchema() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration
-  
- getSchema() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getSchema() - Method in class org.apache.beam.sdk.io.iceberg.IcebergTableCreateConfig
- 
Schema for the destination, in the event that it must be dynamically created. 
- getSchema() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
-  
- getSchema() - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration
-  
- getSchema() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getSchema() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceConfig
- 
Getting a schema of a Snowflake table. 
- getSchema() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getSchema() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getSchema() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
- 
Returns the schema associated with this type. 
- getSchema(Class<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
Retrieve a  Schema for a given  Class type. 
- getSchema(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
- getSchema() - Method in class org.apache.beam.sdk.schemas.utils.ReflectUtils.ClassWithSchema
-  
- getSchema() - Method in class org.apache.beam.sdk.schemas.utils.ReflectUtils.TypeDescriptorWithSchema
-  
- getSchema() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
-  
- getSchema() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
- 
- getSchema() - Method in class org.apache.beam.sdk.values.PCollection
- 
Returns the attached schema. 
- getSchema() - Method in class org.apache.beam.sdk.values.Row.Builder
- 
Return the schema for the row being built. 
- getSchema() - Method in class org.apache.beam.sdk.values.Row.FieldValueBuilder
-  
- getSchema() - Method in class org.apache.beam.sdk.values.Row
- 
Return  Schema which describes the fields. 
- getSchemaCoder(Class<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
- getSchemaCoder(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
- getSchemaId() - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo
-  
- getSchemaIOProvider() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.avro.AvroTableProvider
-  
- getSchemaIOProvider() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datastore.DataStoreV1TableProvider
-  
- getSchemaIOProvider() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubTableProvider
-  
- getSchemaIOProvider() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.SchemaIOTableProviderWrapper
-  
- getSchemaPath(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
- 
- getSchemaPath(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
- 
Return SchemaPathfromTopicPathif exists.
 
- getSchemaPath(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
- 
Return SchemaPathfromTopicPathif exists.
 
- getSchemaPath(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- getSchemaProvider(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
- getSchemaProvider(Class<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
- getSchemaProviders() - Method in class org.apache.beam.sdk.io.aws2.schemas.AwsSchemaRegistrar
-  
- getSchemaProviders() - Method in class org.apache.beam.sdk.schemas.annotations.DefaultSchema.DefaultSchemaProviderRegistrar
-  
- getSchemaProviders() - Method in interface org.apache.beam.sdk.schemas.SchemaProviderRegistrar
- 
- getSchemaRegistry() - Method in class org.apache.beam.sdk.Pipeline
-  
- getSchematizedData() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2Message
- 
Gets schematized data. 
- getSchemaWithoutAttributes(Schema, List<String>) - Static method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider
-  
- getScheme() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
-  
- getScheme() - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration
- 
The uri scheme used by resources on this filesystem. 
- getScheme() - Method in class org.apache.beam.sdk.io.ClassLoaderFileSystem.ClassLoaderResourceId
-  
- getScheme() - Method in class org.apache.beam.sdk.io.ClassLoaderFileSystem
-  
- getScheme() - Method in class org.apache.beam.sdk.io.FileSystem
- 
Get the URI scheme which defines the namespace of the  FileSystem. 
- getScheme() - Method in interface org.apache.beam.sdk.io.fs.ResourceId
- 
Get the scheme which defines the namespace of the  ResourceId. 
- getSdkComponents() - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
-  
- getSdkContainerImage() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
- 
Container image used to configure SDK execution environment on worker. 
- getSdkHarnessContainerImageOverrides() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
Overrides for SDK harness container images. 
- getSdkHarnessLogLevelOverrides() - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
- 
This option controls the log levels for specifically named loggers. 
- getSdkWorkerId() - Method in interface org.apache.beam.sdk.fn.server.HeaderAccessor
- 
This method should be called from the request method. 
- getSdkWorkerParallelism() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- getSearchEndPoint() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getSeconds() - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration.Rate
-  
- getSelectedFields() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransformConfiguration
-  
- getSemiPersistDir() - Method in interface org.apache.beam.sdk.options.RemoteEnvironmentOptions
-  
- getSempClientFactory() - Method in class org.apache.beam.sdk.io.solace.read.UnboundedSolaceSource
-  
- getSenderTimestamp() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record
- 
Gets the timestamp (in milliseconds since the Unix epoch) when the message was sent by the
 sender. 
- getSendFacility() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2Message
- 
Gets send facility. 
- getSendTime() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2Message
- 
Gets send time. 
- getSequenceNumber() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisRecord
-  
- getSequenceNumber() - Method in class org.apache.beam.sdk.io.gcp.bigquery.RowMutationInformation
- 
Deprecated. 
- getSequenceNumber() - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo
-  
- getSequenceNumber() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record
- 
Gets the sequence number of the message (if applicable). 
- getSerializableFunctionUdfs() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.UdfUdafProvider
- 
- getSerializedATN() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- getSerializedATN() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- getSerializedKey() - Method in class org.apache.beam.runners.flink.adapter.FlinkKey
-  
- getSerializedWindowingStrategy() - Method in class org.apache.beam.sdk.expansion.service.WindowIntoTransformProvider.Configuration
-  
- getSerializer(Schema, Map<String, Object>) - Method in class org.apache.beam.sdk.extensions.avro.schemas.io.payloads.AvroPayloadSerializerProvider
-  
- getSerializer(Schema, Map<String, Object>) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoPayloadSerializerProvider
-  
- getSerializer(Schema, Map<String, Object>) - Method in class org.apache.beam.sdk.io.thrift.ThriftPayloadSerializerProvider
-  
- getSerializer(Schema, Map<String, Object>) - Method in class org.apache.beam.sdk.schemas.io.payloads.JsonPayloadSerializerProvider
-  
- getSerializer(Schema, Map<String, Object>) - Method in interface org.apache.beam.sdk.schemas.io.payloads.PayloadSerializerProvider
- 
Get a PayloadSerializer. 
- getSerializer(String, Schema, Map<String, Object>) - Static method in class org.apache.beam.sdk.schemas.io.payloads.PayloadSerializers
-  
- getServer() - Method in class org.apache.beam.sdk.fn.server.GrpcFnServer
- 
- getServerFactory() - Method in class org.apache.beam.runners.fnexecution.environment.DockerEnvironmentFactory.Provider
-  
- getServerFactory() - Method in class org.apache.beam.runners.fnexecution.environment.EmbeddedEnvironmentFactory.Provider
-  
- getServerFactory() - Method in interface org.apache.beam.runners.fnexecution.environment.EnvironmentFactory.Provider
- 
- getServerFactory() - Method in class org.apache.beam.runners.fnexecution.environment.ExternalEnvironmentFactory.Provider
-  
- getServerName() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getServerName() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getServerName() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getServerTransactionId() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
The unique transaction id in which the modifications occurred. 
- getService() - Method in class org.apache.beam.sdk.fn.server.GrpcFnServer
- 
- getServiceAccount() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
- 
Run the job as a specific service account, instead of the default GCE robot. 
- getServiceURL(String, String) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getServiceURL(String) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getSessionProperties() - Method in class org.apache.beam.sdk.io.solace.broker.JcsmpSessionService
-  
- getSessionProperties() - Method in class org.apache.beam.sdk.io.solace.broker.SessionService
- 
Override this method and provide your specific properties, including all those related to
 authentication, and possibly others too. 
- getSessionServiceFactory() - Method in class org.apache.beam.sdk.io.solace.read.UnboundedSolaceSource
-  
- getSetFieldCreator(TypeDescriptor<T>, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
-  
- getSetters(TypeDescriptor<?>, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
- 
- getSetters(TypeDescriptor<T>, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
-  
- getSha256() - Method in class org.apache.beam.runners.dataflow.util.PackageUtil.StagedFile
- 
The SHA-256 hash of the source file. 
- getShard() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-  
- getShardId() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisRecord
-  
- getShardingFunction() - Method in class org.apache.beam.sdk.io.WriteFiles
-  
- getShardNameTemplate() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration
- 
- getShardNumber() - Method in class org.apache.beam.sdk.values.ShardedKey
-  
- getShortTableUrn() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
- 
Return shortened tablespec in datasets/[dataset]/tables/[table] format. 
- getShortValue() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
- 
Return the optional short value for an item, or null if none is provided. 
- getShortValue() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
- 
The optional short value for an item, or nullif none is provided.
 
- getShutdownSourcesAfterIdleMs() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getSideInput(String) - Method in interface org.apache.beam.runners.fnexecution.translation.BatchSideInputHandlerFactory.SideInputGetter
-  
- getSideInputBroadcast(PCollection<T>, SideInputValues.Loader<T>) - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator.TranslationState
-  
- getSideInputBroadcast(PCollection<T>, SideInputValues.Loader<T>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- getSideInputDataSets() - Method in class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- getSideInputKeys() - Method in class org.apache.beam.runners.twister2.translators.functions.DoFnFunction
- 
get the tag id's of all the keys. 
- getSideInputs() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.ParDoSingle
-  
- getSideInputs(ExecutableStage) - Static method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors
-  
- getSideInputs() - Method in class org.apache.beam.runners.twister2.Twister2PipelineExecutionEnvironment
-  
- getSideInputs() - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
- 
Override to specify that this object needs access to one or more side inputs. 
- getSideInputs() - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
- 
Specifies that this object needs access to one or more side inputs. 
- getSideInputs() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
- 
Returns the side inputs used by this Combine operation. 
- getSideInputs() - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
-  
- getSideInputs() - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
- 
Returns the side inputs used by this Combine operation. 
- getSideInputs() - Method in class org.apache.beam.sdk.transforms.Combine.PerKeyWithHotKeyFanout
- 
Returns the side inputs used by this Combine operation. 
- getSideInputs() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-  
- getSideInputs() - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
-  
- getSideInputs() - Method in class org.apache.beam.sdk.transforms.Requirements
- 
The side inputs that this  Contextful needs access to. 
- getSideInputSpecs() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
- 
Get a mapping from PTransform id to side input id to  side inputs that
 are used during execution. 
- getSideInputWindow(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.WindowMappingFn
- 
Returns the window of the side input corresponding to the given window of the main input. 
- getSingleFileMetadata() - Method in class org.apache.beam.sdk.io.FileBasedSource
- 
Returns the information about the single file that this source is reading from. 
- getSinglePCollection() - Method in class org.apache.beam.sdk.values.PCollectionRowTuple
- 
- getSingleTokenNewPartition(Range.ByteStringRange) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.NewPartition
- 
Return a new NewPartition that only contains one token that matches the parentPartition. 
- getSingleWorkerStatus(String, long, TimeUnit) - Method in class org.apache.beam.runners.fnexecution.status.BeamWorkerStatusGrpcService
- 
Get the latest SDK worker status from the client's corresponding SDK harness. 
- getSink() - Method in interface org.apache.beam.runners.fnexecution.control.ControlClientPool
- 
Sink for control clients. 
- getSink() - Method in class org.apache.beam.runners.fnexecution.control.MapControlClientPool
-  
- getSink() - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
- 
Returns the FileBasedSink for this write operation. 
- getSink() - Method in class org.apache.beam.sdk.io.WriteFiles
-  
- getSinkGroupId() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- getSinks() - Static method in class org.apache.beam.sdk.metrics.Lineage
- 
- getSize() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.FixedBytesField
- 
Get the size. 
- getSize() - Method in class org.apache.beam.sdk.extensions.sql.impl.TVFSlidingWindowFn
- 
Size of the generated windows. 
- getSize() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.DetectNewPartitionsDoFn
-  
- getSize(StreamProgress) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.ReadChangeStreamPartitionDoFn
-  
- getSize(TimestampRange) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.DetectNewPartitionsDoFn
-  
- getSize(PartitionMetadata, TimestampRange) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.ReadChangeStreamPartitionDoFn
-  
- getSize(PulsarSourceDescriptor, OffsetRange) - Method in class org.apache.beam.sdk.io.pulsar.ReadFromPulsarDoFn
-  
- getSize() - Method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeBinary
-  
- getSize() - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-  
- getSize() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-  
- getSketchFromByteBuffer(ByteBuffer) - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount
- 
Converts the passed-in sketch from ByteBuffertobyte[], mappingnull
 ByteBuffers (representing empty sketches) to emptybyte[]s.
 
- getSkipHeaderLines() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
-  
- getSkipKeyClone() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
-  
- getSkipValueClone() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
-  
- getSnapshot() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getSnapshotId() - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo
-  
- getSnapshots() - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteResult
-  
- getSnowPipe() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getSocketTimeout() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getSorterType() - Method in class org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options
- 
Returns the sorter type. 
- getSource() - Method in class org.apache.beam.runners.dataflow.util.PackageUtil.StagedFile
- 
The file to stage. 
- getSource() - Method in interface org.apache.beam.runners.fnexecution.control.ControlClientPool
- 
Source of control clients. 
- getSource() - Method in class org.apache.beam.runners.fnexecution.control.MapControlClientPool
-  
- getSource() - Method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.Read
-  
- getSource() - Method in class org.apache.beam.sdk.io.Read.Bounded
- 
Returns the BoundedSourceused to create thisReadPTransform.
 
- getSource() - Method in class org.apache.beam.sdk.io.Read.Unbounded
- 
Returns the UnboundedSourceused to create thisReadPTransform.
 
- getSource() - Method in class org.apache.beam.sdk.io.TextIO.Read
-  
- getSource() - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
-  
- getSources() - Static method in class org.apache.beam.sdk.metrics.Lineage
- 
Lineage representing sources and optionally side inputs. 
- getSparkMaster() - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
-  
- getSparkReceiverClass() - Method in class org.apache.beam.sdk.io.sparkreceiver.ReceiverBuilder
-  
- getSparkSession() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.EvaluationContext
-  
- getSparkSession() - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator.TranslationState
-  
- getSparkSession() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- getSplit() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.SerializableSplit
-  
- getSplitBacklogBytes() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.UnboundedReaderImpl
-  
- getSplitBacklogBytes() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
- 
Returns the size of the backlog of unread data in the underlying data source represented by
 this split of this source. 
- getSplitPointsConsumed() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
- 
- getSplitPointsConsumed() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-  
- getSplitPointsConsumed() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-  
- getSplitPointsConsumed() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-  
- getSplitPointsProcessed() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
- 
Returns the total number of split points that have been processed. 
- getSplitPointsRemaining() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource.AvroReader
-  
- getSplitPointsRemaining() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
- 
- getSplitPointsRemaining() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
-  
- getSplitPointsRemaining() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-  
- getSSEAlgorithm() - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- getSSEAlgorithm() - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration
- 
Algorithm for SSE-S3 encryption, e.g. 
- getSSECustomerKey() - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- getSSECustomerKey() - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration
- 
SSE key for SSE-C encryption, e.g. 
- getSSEKMSKeyId() - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- getSSEKMSKeyId() - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration
- 
KMS key id for SSE-KMS encyrption, e.g. 
- getSsl() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getStableUniqueNames() - Method in interface org.apache.beam.sdk.options.PipelineOptions
- 
Whether to check for stable unique names on each transform. 
- getStackTrace() - Method in class org.apache.beam.io.requestresponse.ApiIOError
- 
The Exceptionstack trace.
 
- getStackTrace() - Method in class org.apache.beam.sdk.io.csv.CsvIOParseError
- 
The caught Throwable.getStackTrace().
 
- getStackTrace() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HealthcareIOError
-  
- getStageBundleFactory(ExecutableStage) - Method in class org.apache.beam.runners.fnexecution.control.DefaultExecutableStageContext
-  
- getStageBundleFactory(ExecutableStage) - Method in interface org.apache.beam.runners.fnexecution.control.ExecutableStageContext
-  
- getStagedArtifacts(String) - Method in class org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService
- 
Returns the rewritten artifacts associated with this job, keyed by environment. 
- getStageName() - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- getStager() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
The resource stager instance that should be used to stage resources. 
- getStagerClass() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
The class responsible for staging resources to be accessible by workers during job execution. 
- getStagingBucketDir() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceConfig
- 
Getting directory where files are staged. 
- getStagingBucketDir() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeStreamingServiceConfig
- 
Getter for a bucket name with directory where files were staged and waiting for loading. 
- getStagingBucketName() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getStagingBucketName() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getStagingLocation() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
- 
GCS path for staging local files, e.g. 
- getStart() - Method in class org.apache.beam.sdk.extensions.ordered.ContiguousSequenceRange
-  
- getStart() - Method in class org.apache.beam.sdk.io.cassandra.RingRange
-  
- getStart() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient.HL7v2MessagePages
-  
- getStart() - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration
-  
- getStartAtTimestamp() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration
-  
- getStartDate() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-  
- getStartDate() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-  
- getStartDate() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-  
- getStartKey() - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
- 
- getStartOffset() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
- 
Returns the starting offset of the source. 
- getStartPosition() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-  
- getStartPosition() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-  
- getStartPosition() - Method in interface org.apache.beam.sdk.io.range.RangeTracker
- 
Returns the starting position of the current range, inclusive. 
- getStartReadTime() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getStartTime() - Method in class org.apache.beam.sdk.io.CountingSource.CounterMark
- 
Returns the time the reader was started. 
- getStartTime() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.InitialPipelineState
-  
- getStartTime() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.PartitionRecord
-  
- getStartTimestamp() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChildPartitionsRecord
- 
It is the partition_start_time of the child partition token. 
- getStartTimestamp() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata
- 
It is the start time at which the partition started existing in Cloud Spanner. 
- getState() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-  
- getState() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
-  
- getState() - Method in class org.apache.beam.runners.direct.DirectRunner.DirectPipelineResult
-  
- getState() - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
-  
- getState() - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
-  
- getState() - Method in class org.apache.beam.runners.jet.FailedRunningPipelineResults
-  
- getState() - Method in class org.apache.beam.runners.jet.JetPipelineResult
-  
- getState(JobApi.GetJobStateRequest, StreamObserver<JobApi.JobStateEvent>) - Method in class org.apache.beam.runners.jobsubmission.InMemoryJobService
-  
- getState() - Method in class org.apache.beam.runners.jobsubmission.JobInvocation
- 
Retrieve the job's current state. 
- getState(JobApi.GetJobStateRequest, StreamObserver<JobApi.JobStateEvent>) - Method in class org.apache.beam.runners.portability.testing.TestJobService
-  
- getState() - Method in class org.apache.beam.runners.spark.SparkPipelineResult
-  
- getState() - Method in class org.apache.beam.runners.spark.stateful.SparkStateInternals
-  
- getState() - Method in class org.apache.beam.runners.spark.stateful.StateAndTimers
-  
- getState() - Method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineResult
-  
- getState() - Method in class org.apache.beam.runners.twister2.Twister2PipelineResult
-  
- getState() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata
- 
The state in which the current partition is in. 
- getState() - Method in interface org.apache.beam.sdk.PipelineResult
- 
Retrieves the current state of the pipeline execution. 
- getState() - Method in interface org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimator
- 
- getState() - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.Manual
-  
- getState() - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.MonotonicallyIncreasing
-  
- getState() - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.WallTime
-  
- getStateBackend() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getStateBackendFactory() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
- 
- getStateBackendStoragePath() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getStateCoder(Pipeline) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler
- 
Provide the state coder. 
- getStateCoder() - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
- 
- getStateEvent() - Method in class org.apache.beam.runners.jobsubmission.JobInvocation
- 
Retrieve the job's current state. 
- getStateStream(JobApi.GetJobStateRequest, StreamObserver<JobApi.JobStateEvent>) - Method in class org.apache.beam.runners.jobsubmission.InMemoryJobService
-  
- getStaticCreator(TypeDescriptor<?>, Method, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
-  
- getStaticCreator(TypeDescriptor<?>, Method, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
-  
- getStatistic() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteTable
-  
- getStatus() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.WriteFailure
-  
- getStatusCode() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HealthcareIOError
-  
- getStatusDate() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus
-  
- getStatusUpdateFrequency() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler
- 
- getStepNames() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator.JobSpecification
- 
Returns the mapping of AppliedPTransformsto the internal step name
 for thatAppliedPTransform.
 
- getStopPipelineWatermark() - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
-  
- getStopPosition() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-  
- getStopPosition() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-  
- getStopPosition() - Method in interface org.apache.beam.sdk.io.range.RangeTracker
- 
Returns the ending position of the current range, exclusive. 
- getStopReadTime() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getStorageApiAppendThresholdBytes() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getStorageApiAppendThresholdRecordCount() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getStorageClient(BigQueryOptions) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices
- 
- getStorageClient(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl
-  
- getStorageClient(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
-  
- getStorageIntegrationName() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getStorageIntegrationName() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceConfig
- 
Getting Snowflake integration which is used in COPY statement. 
- getStorageIntegrationName() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getStorageLevel() - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
-  
- getStorageWriteApiMaxRequestSize() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getStorageWriteApiMaxRetries() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getStorageWriteApiTriggeringFrequencySec() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getStorageWriteMaxInflightBytes() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getStorageWriteMaxInflightRequests() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getStreamAppendClient(String, DescriptorProtos.DescriptorProto, boolean, AppendRowsRequest.MissingValueInterpretation) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.WriteStreamService
- 
Create an append client for a given Storage API write stream. 
- getStreamAppendClient(String, DescriptorProtos.DescriptorProto, boolean, AppendRowsRequest.MissingValueInterpretation) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.WriteStreamServiceImpl
-  
- getStreamAppendClient(String, DescriptorProtos.DescriptorProto, boolean, AppendRowsRequest.MissingValueInterpretation) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- getStreamingService() - Method in interface org.apache.beam.sdk.io.snowflake.services.SnowflakeServices
-  
- getStreamingService() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeServicesImpl
-  
- getStreamingSideInputCacheExpirationMillis() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getStreamingSideInputCacheMb() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getStreamingTimeoutMs() - Method in interface org.apache.beam.runners.spark.SparkPortableStreamingPipelineOptions
-  
- getStreamName() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisRecord
-  
- getStreamTableDebugString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.DaoFactory
-  
- getString(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getString(Map<String, Object>, String, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getString() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- getString(String) - Method in class org.apache.beam.sdk.values.Row
- 
- getString(int) - Method in class org.apache.beam.sdk.values.Row
- 
Get a Stringvalue by field index,ClassCastExceptionis thrown if schema
 doesn't match.
 
- getStrings(Map<String, Object>, String, List<String>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
-  
- getStringSet(MetricName) - Method in class org.apache.beam.runners.jet.metrics.JetMetricsContainer
-  
- getStringSet(MetricName) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
- 
Return the  StringSet that should be used for implementing the given  metricName
 in this container. 
- getStringSet() - Method in class org.apache.beam.sdk.metrics.StringSetResult.EmptyStringSetResult
- 
Returns an empty immutable set. 
- getStringSet() - Method in class org.apache.beam.sdk.metrics.StringSetResult
-  
- getStringSets() - Method in class org.apache.beam.sdk.metrics.MetricQueryResults
- 
Return the metric results for the sets that matched the filter. 
- getStuckCommitDurationMillis() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getSubmissionMode() - Method in class org.apache.beam.sdk.io.solace.broker.SessionServiceFactory
-  
- getSubmissionMode() - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- getSubnetwork() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
- 
- getSubProvider(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.FullNameTableProvider
-  
- getSubProvider(String) - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.TableProvider
- 
Returns a sub-provider, e.g. 
- getSubProviders() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.TableProvider
- 
Returns all sub-providers, e.g. 
- getSubSchema(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- getSubSchemaNames() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- getSubscription() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration
-  
- getSubscription() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
- 
Get the subscription being read from. 
- getSubscriptionName() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration
-  
- getSubscriptionProvider() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
- 
- getSubSequenceNumber() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisRecord
-  
- getSuccessfulBodies() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.ExecuteBundlesResult
-  
- getSuccessfulBodies() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write.AbstractResult
-  
- getSuccessfulBodies() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write.Result
- 
Gets successful bodies from Write. 
- getSuccessfulBundles() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.ExecuteBundlesResult
- 
Gets successful FhirBundleResponse from execute bundles operation. 
- getSuccessfulInserts() - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
- 
Returns a  PCollection containing the  TableRows that were written to BQ via the
 streaming insert API. 
- getSuccessfulPublish() - Method in class org.apache.beam.sdk.io.solace.write.SolaceOutput
-  
- getSuccessfulStorageApiInserts() - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
- 
Return all rows successfully inserted using one of the storage-api insert methods. 
- getSuccessfulTableLoads() - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
- 
- getSuggestedFilenameSuffix() - Method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
- 
Deprecated. 
- getSuggestedFilenameSuffix() - Method in interface org.apache.beam.sdk.io.FileBasedSink.OutputFileHints
-  
- getSuggestedSuffix() - Method in enum org.apache.beam.sdk.io.Compression
-  
- getSum() - Method in class org.apache.beam.sdk.metrics.DistributionResult
-  
- getSum() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
-  
- getSumAndReset() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
-  
- getSummary() - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo
-  
- getSupertype(Class<? super T>) - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
Returns the generic form of a supertype. 
- getSupported() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryFilter
-  
- getSupported() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableFilter
-  
- getSupportedClass() - Method in interface org.apache.beam.runners.dataflow.util.CloudObjectTranslator
- 
- getSupportedClass() - Method in class org.apache.beam.runners.dataflow.util.RowCoderCloudObjectTranslator
-  
- getSupportedClass() - Method in class org.apache.beam.runners.dataflow.util.SchemaCoderCloudObjectTranslator
-  
- getSynchronizedProcessingOutputWatermark() - Method in interface org.apache.beam.runners.local.Bundle
- 
Returns the processing time output watermark at the time the producing Executablecommitted this bundle.
 
- getSynchronizedProcessingTime() - Method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.SparkWatermarks
-  
- getTable() - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration
-  
- getTable(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- getTable(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
-  
- getTable(String) - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.TableProvider
- 
Get a specific table from this provider it is present, or null if it is not present. 
- getTable(BigQueryOptions, TableReference) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
-  
- getTable() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertError
-  
- getTable() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
- 
Returns the table to read, or nullif reading from a query instead.
 
- getTable() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- getTable() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Returns the table reference, or null.
 
- getTable(TableReference) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
- 
Gets the specified  Table resource by table ID. 
- getTable(TableReference, List<String>) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
-  
- getTable(TableReference, List<String>, BigQueryServices.DatasetService.TableMetadataView) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
-  
- getTable(TableReference) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.DatasetServiceImpl
- 
Gets the specified  Table resource by table ID. 
- getTable(TableReference, List<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.DatasetServiceImpl
-  
- getTable(TableReference, List<String>, BigQueryServices.DatasetService.TableMetadataView) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.DatasetServiceImpl
-  
- getTable(DestinationT) - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
- 
- getTable() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- getTable(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.PortableBigQueryDestinations
-  
- getTable() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration
-  
- getTable() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- getTable(TableReference) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- getTable(TableReference, List<String>) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- getTable(TableReference, List<String>, BigQueryServices.DatasetService.TableMetadataView) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- getTable() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getTable() - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration
-  
- getTable() - Method in class org.apache.beam.sdk.io.iceberg.SchemaTransformConfiguration
-  
- getTable() - Method in class org.apache.beam.sdk.io.kudu.TableAndRecord
-  
- getTable() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadConfiguration
-  
- getTable() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteConfiguration
-  
- getTable() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getTable() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceConfig
- 
Getting a table as a source of reading or destination to write. 
- getTable() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getTableAdminClient() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.BigtableChangeStreamAccessor
-  
- getTableByFullName(TableName) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
-  
- getTableByFullName(TableName) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.FullNameTableProvider
-  
- getTableConstraints(DestinationT) - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
- 
Returns TableConstraints (including primary and foreign key) to be used when creating the
 table. 
- getTableConstraints(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.PortableBigQueryDestinations
-  
- getTableCreateConfig() - Method in class org.apache.beam.sdk.io.iceberg.IcebergDestination
- 
Metadata and constraints for creating a new table, if it must be done dynamically. 
- getTableDescription() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- getTableId() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
Returns the table being read from. 
- getTableId() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProvider.BigtableReadSchemaTransformConfiguration
-  
- getTableId() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProvider.BigtableWriteSchemaTransformConfiguration
-  
- getTableId() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
- 
Return the metadata table name. 
- getTableId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration
-  
- getTableId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration
-  
- getTableId() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
-  
- getTableId() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
-  
- getTableId() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.WriteRowMutations
-  
- getTableIdentifier() - Method in class org.apache.beam.sdk.io.iceberg.IcebergDestination
- 
The iceberg table identifier to write data to. 
- getTableIdentifier() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getTableImpl(TableReference, List<String>, BigQueryServices.DatasetService.TableMetadataView) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- getTableName() - Method in class org.apache.beam.sdk.extensions.sql.impl.TableName
- 
Table name, the last element of the fully-specified table name with path. 
- getTableName() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataTableNames
-  
- getTableName() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
The name of the table in which the modifications within this record occurred. 
- getTableNames() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- getTablePath(Table) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
- 
Returns a full table path (exlucding top-level schema) for a given ZetaSQL Table. 
- getTableProvider() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- getTableProvider() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
- 
Returns the table to read, or nullif reading from a query instead.
 
- getTableProvider() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- getTableReference() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- getTableResource(String, String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
-  
- getTableRow() - Method in class org.apache.beam.sdk.io.gcp.bigquery.RowMutation
-  
- getTables() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
-  
- getTables() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.InMemoryMetaTableProvider
-  
- getTables() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.ReadOnlyTableProvider
-  
- getTables() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.TableProvider
- 
Get all tables from this provider. 
- getTables() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
-  
- getTables() - Method in class org.apache.beam.sdk.extensions.sql.meta.store.InMemoryMetaStore
-  
- getTables() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerSchema
-  
- getTableSchema(String, String) - Static method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO
- 
- getTableSchema() - Method in class org.apache.beam.sdk.io.gcp.bigquery.SchemaAndRecord
-  
- getTableSchema(String, String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogBeamSchema
- 
Gets the table schema, or absent optional if the table doesn't exist in the database. 
- getTableSchema() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.WriteBuilder.Configuration
-  
- getTableSchema() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceConfig
-  
- getTableSpec() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryExportReadSchemaTransformConfiguration
- 
Specifies a table for a BigQuery read job. 
- getTableSpec() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransformConfiguration
-  
- getTableSpec() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
- 
Return the tablespec in [project:].dataset.tableid format. 
- getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.meta.BaseBeamTable
-  
- getTableStatistics(PipelineOptions) - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTable
- 
Estimates the number of rows or the rate for unbounded Tables. 
- getTableStatistics(PipelineOptions, SchemaIO) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datastore.DataStoreV1TableProvider
-  
- getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
-  
- getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable
-  
- getTableStatistics(PipelineOptions, SchemaIO) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.SchemaIOTableProviderWrapper
-  
- getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestBoundedTable
-  
- getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestUnboundedTable
-  
- getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTable
-  
- getTableStringIdentifier(ValueInSingleWindow<Row>) - Method in interface org.apache.beam.sdk.io.iceberg.DynamicDestinations
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.avro.AvroTableProvider
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryTableProvider
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigtable.BigtableTableProvider
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datastore.DataStoreV1TableProvider
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.KafkaTableProvider
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTableProvider
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.parquet.ParquetTableProvider
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubTableProvider
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsublite.PubsubLiteTableProvider
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.ReadOnlyTableProvider
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.SchemaIOTableProviderWrapper
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.seqgen.GenerateSequenceTableProvider
-  
- getTableType() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.TableProvider
- 
Gets the table type this provider handles. 
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTableProvider
-  
- getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.store.InMemoryMetaStore
-  
- getTableUrn(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
- 
Return the tablespec in projects/[project]/datasets/[dataset]/tables/[table] format. 
- getTag() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getTag(int) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
- 
Returns the tuple tag at the given index. 
- getTag() - Method in class org.apache.beam.sdk.values.TaggedPValue
- 
Returns the local tag associated with the  PValue. 
- getTagInternal() - Method in interface org.apache.beam.sdk.values.PCollectionView
- 
- getTagInternal() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
- 
- getTargetDataset() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.TestBigQueryOptions
-  
- getTargetParallelism() - Method in interface org.apache.beam.runners.direct.DirectOptions
-  
- getTargetTable(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageTableSource
-  
- getTargetTableId(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageTableSource
-  
- getTempDatasetId() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getTempDirectory() - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-  
- getTempDirectoryProvider() - Method in class org.apache.beam.sdk.io.FileBasedSink
- 
- getTempFilename() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-  
- getTemplateLocation() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
- 
Where the runner should generate a template file. 
- getTempLocation() - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
- 
Returns the configured temporary location. 
- getTempLocation() - Method in class org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options
- 
Returns the configured temporary location. 
- getTempLocation() - Method in interface org.apache.beam.sdk.options.PipelineOptions
- 
A pipeline level default location for storing temporary files. 
- getTempRoot() - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-  
- GETTER_WITH_NULL_METHOD_ERROR - Static variable in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
-  
- GetterBasedSchemaProvider - Class in org.apache.beam.sdk.schemas
- 
- GetterBasedSchemaProvider() - Constructor for class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
- 
Deprecated. 
- GetterBasedSchemaProviderBenchmark - Class in org.apache.beam.sdk.jmh.schemas
- 
- GetterBasedSchemaProviderBenchmark() - Constructor for class org.apache.beam.sdk.jmh.schemas.GetterBasedSchemaProviderBenchmark
-  
- GetterBasedSchemaProviderV2 - Class in org.apache.beam.sdk.schemas
- 
- GetterBasedSchemaProviderV2() - Constructor for class org.apache.beam.sdk.schemas.GetterBasedSchemaProviderV2
-  
- getTerminateAfterSecondsSinceNewOutput() - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformConfiguration
- 
If no new files are found after this many seconds, this transform will cease to watch for new
 files. 
- GetterTypeSupplier() - Constructor for class org.apache.beam.sdk.schemas.JavaBeanSchema.GetterTypeSupplier
-  
- getTestMode() - Method in interface org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineOptions
- 
Set to true to run the job in test mode. 
- getTestTimeoutSeconds() - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-  
- getThrottleDuration() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
- 
The amount of time an attempt will be throttled if deemed necessary based on previous success
 rate. 
- getThroughputEstimate() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.StreamProgress
-  
- getTimeDomain() - Method in interface org.apache.beam.sdk.state.TimerSpec
-  
- getTimePartitioning() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- getTimerFamilyId() - Method in class org.apache.beam.sdk.fn.data.LogicalEndpoint
-  
- getTimerFamilyId() - Method in class org.apache.beam.sdk.fn.data.TimerEndpoint
-  
- getTimerReceivers() - Method in interface org.apache.beam.runners.fnexecution.control.RemoteBundle
- 
Get a map of (transform id, timer id) to  receivers which consume timers,
 forwarding them to the remote environment. 
- getTimerReceivers() - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor.ActiveBundle
-  
- getTimers(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
-  
- getTimers() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-  
- getTimers() - Method in class org.apache.beam.runners.spark.stateful.StateAndTimers
-  
- getTimers(MetricFilter) - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.WithMetricsSupport
-  
- getTimerSpec() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.TimerSpec
-  
- getTimerSpecs() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
- 
Get a mapping from PTransform id to timer id to  timer specs that are used
 during execution. 
- getTimes() - Method in class org.apache.beam.runners.spark.io.CreateStream
- 
- getTimestamp() - Method in class org.apache.beam.sdk.extensions.ordered.ContiguousSequenceRange
-  
- getTimeStamp() - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsMessage
- 
Timestamp the message was sent at (in epoch millis). 
- getTimestamp() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Document
-  
- getTimestamp() - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWritePayload
-  
- getTimestamp() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.HeartbeatRecord
- 
Indicates the timestamp for which the change stream query has returned all changes. 
- getTimestamp() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getTimestamp(T, Instant) - Method in interface org.apache.beam.sdk.io.kafka.KafkaPublishTimestampFunction
- 
Returns timestamp for element being published to Kafka. 
- getTimestamp() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
-  
- getTimestamp() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getTimestamp() - Method in class org.apache.beam.sdk.metrics.GaugeResult.EmptyGaugeResult
-  
- getTimestamp() - Method in class org.apache.beam.sdk.metrics.GaugeResult
-  
- getTimestamp() - Method in class org.apache.beam.sdk.values.FailsafeValueInSingleWindow
- 
Returns the timestamp of this FailsafeValueInSingleWindow.
 
- getTimestamp() - Method in class org.apache.beam.sdk.values.TimestampedValue
-  
- getTimestamp() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow
- 
Returns the timestamp of this ValueInSingleWindow.
 
- getTimestampAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration
-  
- getTimestampAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
- 
Get the timestamp attribute. 
- getTimestampAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
- 
Get the timestamp attribute. 
- getTimestampAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration
-  
- getTimestampCombiner() - Method in interface org.apache.beam.sdk.state.WatermarkHoldState
- 
Return the  TimestampCombiner which will be used to determine a watermark hold time
 given an element timestamp, and to combine watermarks from windows which are about to be
 merged. 
- getTimestampCombiner() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- getTimestampFn() - Method in class org.apache.beam.sdk.io.solace.read.UnboundedSolaceSource
-  
- getTimestampForRecord(TimestampPolicy.PartitionContext, KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.CustomTimestampPolicyWithLimitedDelay
-  
- getTimestampForRecord(TimestampPolicy.PartitionContext, KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicy
- 
Returns record timestamp (aka event time). 
- getTimestampForRecord(TimestampPolicy.PartitionContext, KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.LogAppendTimePolicy
-  
- getTimestampForRecord(TimestampPolicy.PartitionContext, KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.ProcessingTimePolicy
-  
- getTimestampForRecord(TimestampPolicy.PartitionContext, KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.TimestampFnPolicy
-  
- getTimestampMillis() - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo
-  
- getTimestampMsSinceEpoch() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.OutgoingMessage
- 
Timestamp for element (ms since epoch). 
- getTimestampPolicyFactory() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getTimestampTransforms() - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
- 
The transforms applied to the arrival time of an element to determine when this trigger allows
 output. 
- getTimestampType() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
-  
- getTimeToLive() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record
- 
The number of milliseconds before the message is discarded or moved to Dead Message Queue. 
- getTimeZone() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-  
- getTimeZone() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-  
- getTimeZone() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-  
- getTiming() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
- 
Return the timing of this pane. 
- getTo() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRange
- 
Returns the range end timestamp (exclusive). 
- getTo() - Method in class org.apache.beam.sdk.io.range.OffsetRange
-  
- getToken() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChildPartition
- 
Unique partition identifier, which can be used to perform a change stream query. 
- getTokenNames() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
- 
Deprecated. 
- getTokenNames() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
- 
Deprecated. 
- getTokenWithCorrectPartition(Range.ByteStringRange, ChangeStreamContinuationToken) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamContinuationTokenHelper
- 
Return the continuation token with correct partition. 
- getToKV() - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
-  
- getTopic() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
-  
- getTopic() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration
-  
- getTopic() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
- 
Get the topic being written to. 
- getTopic() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
- 
Get the topic being read from. 
- getTopic() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration
-  
- getTopic() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
-  
- getTopic() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- getTopic() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
- 
Sets the topic from which to read. 
- getTopic() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
-  
- getTopic() - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration
-  
- getTopic() - Method in class org.apache.beam.sdk.io.mqtt.MqttRecord
-  
- getTopic() - Method in class org.apache.beam.sdk.io.pulsar.PulsarMessage
-  
- getTopicName() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration
-  
- getTopicPartition() - Method in class org.apache.beam.sdk.io.kafka.KafkaSourceDescriptor
-  
- getTopicPartitions() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getTopicPattern() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getTopicProvider() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
- 
- getTopicProvider() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
- 
- getTopics() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
-  
- getTopics() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getToRowFunction(Class<T>, Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
-  
- getToRowFunction() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDynamicMessageSchema
-  
- getToRowFunction() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
- 
Returns the fromRow conversion function. 
- getToRowFunction(Class<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
Retrieve the function that converts an object of the specified type to a  Row object. 
- getToRowFunction(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
Retrieve the function that converts an object of the specified type to a  Row object. 
- getToRowFunction() - Method in class org.apache.beam.sdk.values.PCollection
- 
Returns the attached schema's toRowFunction. 
- getToSnapshot() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getToSnapshotRef() - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- getTotalBacklogBytes() - Method in interface org.apache.beam.sdk.io.jms.AutoScaler
- 
Returns the size of the backlog of unread data in the underlying data source represented by all
 splits of this source. 
- getTotalBacklogBytes() - Method in class org.apache.beam.sdk.io.jms.DefaultAutoscaler
-  
- getTotalBacklogBytes() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
- 
Returns the size of the backlog of unread data in the underlying data source represented by
 all splits of this source. 
- getTotalStreamDuration() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.ChangeStreamResultSetMetadata
- 
Returns the total stream duration of change stream records so far. 
- getTotalStreamTimeMillis() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata
- 
The total streaming time (in millis) for this record. 
- getTraitDef() - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-  
- getTransactionIsolation() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- getTransactionTag() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
The transaction tag associated with the given transaction. 
- getTransform(RunnerApi.FunctionSpec, PipelineOptions) - Method in class org.apache.beam.sdk.expansion.service.ExpansionServiceSchemaTransformProvider
-  
- getTransform(RunnerApi.FunctionSpec, PipelineOptions) - Method in interface org.apache.beam.sdk.expansion.service.TransformProvider
-  
- getTransformId() - Method in class org.apache.beam.sdk.fn.data.DataEndpoint
-  
- getTransformId() - Method in class org.apache.beam.sdk.fn.data.LogicalEndpoint
-  
- getTransformId() - Method in class org.apache.beam.sdk.fn.data.TimerEndpoint
-  
- getTransformingMap(Map<K1, V1>, Function<K1, K2>, Function<V1, V2>) - Static method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils
-  
- getTransformNameMapping() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
Mapping of old PTransform names to new ones, specified as JSON {"oldName":"newName",...}
 .
 
- getTransformPayloadTranslators() - Method in class org.apache.beam.runners.dataflow.DataflowRunner.DataflowTransformTranslator
-  
- getTransformPayloadTranslators() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.Registrar
-  
- getTransformPayloadTranslators() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOTranslation.ReadRegistrar
-  
- getTransformPayloadTranslators() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOTranslation.WriteRegistrar
-  
- getTransformPayloadTranslators() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQuerySchemaTransformTranslation.ReadWriteRegistrar
-  
- getTransformPayloadTranslators() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubSubPayloadTranslation.ReadRegistrar
-  
- getTransformPayloadTranslators() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubSubPayloadTranslation.WriteRegistrar
-  
- getTransformPayloadTranslators() - Method in class org.apache.beam.sdk.io.iceberg.IcebergSchemaTransformTranslation.ReadRegistrar
-  
- getTransformPayloadTranslators() - Method in class org.apache.beam.sdk.io.iceberg.IcebergSchemaTransformTranslation.WriteRegistrar
-  
- getTransformPayloadTranslators() - Method in class org.apache.beam.sdk.io.kafka.KafkaSchemaTransformTranslation.ReadRegistrar
-  
- getTransformPayloadTranslators() - Method in class org.apache.beam.sdk.io.kafka.KafkaSchemaTransformTranslation.WriteRegistrar
-  
- getTransformPayloadTranslators() - Method in class org.apache.beam.sdk.io.kafka.upgrade.KafkaIOTranslation.ReadRegistrar
-  
- getTransformPayloadTranslators() - Method in class org.apache.beam.sdk.io.kafka.upgrade.KafkaIOTranslation.WriteRegistrar
-  
- getTransformPayloadTranslators() - Method in class org.apache.beam.sdk.managed.ManagedSchemaTransformTranslation.ManagedTransformRegistrar
-  
- getTransformPayloadTranslators() - Method in class org.apache.beam.sdk.transforms.Redistribute.Registrar
-  
- getTransformStepNames() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-  
- getTransformTranslator(Class<TransformT>) - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator
- 
Returns the  TransformTranslator to use for instances of the specified PTransform class,
 or null if none registered. 
- getTransformTranslator(TransformT) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.PipelineTranslatorBatch
- 
- getTransformTranslator(TransformT) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
- 
- getTransformUniqueID(RunnerApi.FunctionSpec) - Method in interface org.apache.beam.sdk.expansion.service.TransformProvider
-  
- getTranslator() - Method in class org.apache.beam.runners.dataflow.DataflowRunner
- 
Returns the DataflowPipelineTranslator associated with this object. 
- getTransport() - Static method in class org.apache.beam.sdk.extensions.gcp.util.Transport
-  
- getTrigger() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- getTriggeringFrequencySeconds() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- getTriggeringFrequencySeconds() - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration
-  
- getTruncatedRestriction() - Method in class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker.TruncateResult
-  
- getTruncateTimestamps() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogPipelineOptions
- 
Whether to truncate timestamps in tables described by Data Catalog. 
- getTruncateTimestamps() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.ConversionOptions
-  
- getTSetEnvironment() - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- getTSetGraph() - Method in class org.apache.beam.runners.twister2.Twister2PipelineExecutionEnvironment
-  
- getTupleTag() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple.TaggedKeyedPCollection
- 
Returns the TupleTag of this TaggedKeyedPCollection. 
- getTupleTagId(PValue) - Static method in class org.apache.beam.runners.jet.Utils
-  
- getTupleTagList() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
- 
Returns the TupleTagList tuple associated with this schema. 
- getTwister2Home() - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- getType() - Method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns the type this coder encodes/decodes. 
- getType() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory
- 
Returns the type for the datum factory. 
- getType() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.TypeWithNullability
-  
- getType(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- getType() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPMeasure
-  
- getType() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table
- 
type of the table. 
- getType() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getType() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ColumnType
- 
The type of the column. 
- getType() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerSchema.Column
-  
- getType() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getType() - Method in class org.apache.beam.sdk.io.solace.data.Solace.Destination
- 
Gets the type of the destination (TOPIC, QUEUE or UNKNOWN). 
- getType() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
- 
Returns the field type. 
- getType() - Method in class org.apache.beam.sdk.schemas.Schema.Field
- 
- getType(String) - Method in class org.apache.beam.sdk.schemas.Schema.Options
- 
Get the type of an option. 
- getType() - Method in interface org.apache.beam.sdk.testing.TestStream.Event
-  
- getType() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
- 
- getType() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
- 
- getType() - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
- getTypeDescriptor() - Method in class org.apache.beam.sdk.schemas.utils.ReflectUtils.TypeDescriptorWithSchema
-  
- getTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.ViewFn
- 
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollection
- 
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryListFromMultimapViewFn
-  
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryListViewFn
-  
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryMapFromVoidKeyViewFn
-  
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryMapViewFn
-  
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryMultimapFromVoidKeyViewFn
-  
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.InMemoryMultimapViewFn
-  
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.IterableBackedListViewFn
-  
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn
- 
Deprecated. 
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn2
-  
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.ListViewFn
- 
Deprecated. 
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.ListViewFn2
-  
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.MapViewFn
- 
Deprecated. 
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.MapViewFn2
-  
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.MultimapViewFn
- 
Deprecated. 
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.MultimapViewFn2
-  
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn
- 
Deprecated. 
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn2
-  
- getTypeDescriptor() - Method in class org.apache.beam.sdk.values.TupleTag
- 
Returns a TypeDescriptorcapturing what is known statically about the type of thisTupleTaginstance's most-derived class.
 
- getTypeFactory() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- getTypeMap() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- getTypeName() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- getTypeName() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
-  
- getTypeNames() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- getTypeParameter(String) - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
Returns a TypeVariablefor the named type parameter.
 
- getTypePayload() - Method in class org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService.ArtifactDestination
-  
- getTypes() - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
Returns a set of  TypeDescriptor, one for each superclass as well as each
 interface implemented by this class. 
- getTypeUrn() - Method in class org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService.ArtifactDestination
-  
- getUdaf(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct
-  
- getUdafImpl() - Method in class org.apache.beam.sdk.extensions.sql.impl.LazyAggregateCombineFn
-  
- getUdafs() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.UdfUdafProvider
-  
- getUnalignedCheckpointEnabled() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- getUnboundedReaderMaxElements() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
The max elements read from an UnboundedReader before checkpointing. 
- getUnboundedReaderMaxReadTimeMs() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
The max amount of time an UnboundedReader is consumed before checkpointing. 
- getUnboundedReaderMaxReadTimeSec() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
- getUnboundedReaderMaxWaitForElementsMs() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
The max amount of time waiting for elements when reading from UnboundedReader. 
- getUnderlyingDoFn() - Method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides.BatchStatefulDoFn
-  
- getUnderlyingSchemaProvider(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.annotations.DefaultSchema.DefaultSchemaProvider
- 
- getUnderlyingSchemaProvider(Class<T>) - Method in class org.apache.beam.sdk.schemas.annotations.DefaultSchema.DefaultSchemaProvider
- 
- getUnfinishedEndpoints() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
- 
Get all unfinished data and timers endpoints represented as [transform_id]:data and
 [transform_id]:timers:[timer_family_id]. 
- getUnfinishedMinWatermark() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.cache.AsyncWatermarkCache
-  
- getUnfinishedMinWatermark() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.cache.NoOpWatermarkCache
-  
- getUnfinishedMinWatermark() - Method in interface org.apache.beam.sdk.io.gcp.spanner.changestreams.cache.WatermarkCache
- 
- getUnfinishedMinWatermark() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao
- 
- getUnionCoder() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
-  
- getUnionTag() - Method in class org.apache.beam.sdk.transforms.join.RawUnionValue
-  
- getUniqueId() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisRecord
-  
- getUnknownFields() - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWritePayload
-  
- getUnknownFieldsPayload() - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWritePayload
-  
- getUntilTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
- 
The trigger that signals termination of this trigger. 
- getUpdateCompatibilityVersion() - Method in interface org.apache.beam.sdk.options.StreamingOptions
-  
- getUpdatedSchema() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StreamAppendClient
- 
If the table schema has been updated, returns the new schema. 
- getUpdatedSchema(TableSchema, TableSchema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableSchemaUpdateUtils
-  
- getUploadBufferSizeBytes() - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.CreateOptions
- 
If non-null, the upload buffer size to be used. 
- getUrl() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getUrl() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getUrn() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.PayloadTranslator
-  
- getUrn() - Method in class org.apache.beam.sdk.schemas.transforms.SchemaTransformTranslation.SchemaTransformPayloadTranslator
-  
- getUrn() - Method in interface org.apache.beam.sdk.transforms.Materialization
- 
- getUseActiveSparkSession() - Method in interface org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineOptions
-  
- getUseAltsServer() - Method in interface org.apache.beam.sdk.expansion.service.ExpansionServiceOptions
-  
- getUseAtLeastOnceSemantics() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- getUseCdcWrites() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- getUseDataStreamForBatch() - Method in interface org.apache.beam.runners.flink.VersionDependentFlinkPipelineOptions
-  
- getUsePublicIps() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
- 
Specifies whether worker pools should be started with public IP addresses. 
- getUserAgent() - Method in interface org.apache.beam.sdk.options.PipelineOptions
- 
A user agent string as per RFC2616, describing the pipeline to external services. 
- getUserId() - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- getUsername() - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration
-  
- getUsername() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- getUsername() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- getUsername() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration
-  
- getUsername() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getUsername() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getUsername() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getUseSeparateWindmillHeartbeatStreams() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getUsesProvidedSparkContext() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- getUseStandardSql() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryExportReadSchemaTransformConfiguration
- 
Enables BigQuery's Standard SQL dialect when reading from a query. 
- getUseStorageApiConnectionPool() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getUseStorageWriteApi() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getUseStorageWriteApiAtLeastOnce() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- getUseTransformService() - Method in interface org.apache.beam.sdk.extensions.python.PythonExternalTransformOptions
-  
- getUseWindmillIsolatedChannels() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getUsingStandardSql() - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher.TableAndQuery
-  
- getUuid() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.PartitionRecord
-  
- getUUID() - Method in class org.apache.beam.sdk.schemas.Schema
- 
Get this schema's UUID. 
- getUuidFromMessage(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider
-  
- getValidationFailures() - Method in class org.apache.beam.sdk.io.cdap.context.FailureCollectorWrapper
-  
- getValue() - Method in class org.apache.beam.runners.spark.util.ByteArray
-  
- getValue() - Method in class org.apache.beam.runners.spark.util.SideInputBroadcast
-  
- getValue() - Method in class org.apache.beam.sdk.coders.StructuralByteArray
-  
- getValue() - Method in class org.apache.beam.sdk.extensions.timeseries.FillGaps.InterpolateData
-  
- getValue() - Method in class org.apache.beam.sdk.io.range.ByteKey
- 
Returns a read-only  ByteBuffer representing this  ByteKey. 
- getValue() - Method in class org.apache.beam.sdk.metrics.GaugeResult.EmptyGaugeResult
-  
- getValue() - Method in class org.apache.beam.sdk.metrics.GaugeResult
-  
- getValue() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType.Value
- 
Return the integer enum value. 
- getValue(Class<T>) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType.Value
- 
Returns the current value of the OneOf as the destination type. 
- getValue() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType.Value
- 
Returns the current value of the OneOf. 
- getValue(String) - Method in class org.apache.beam.sdk.schemas.Schema.Options
- 
Get the value of an option. 
- getValue(String, Class<T>) - Method in class org.apache.beam.sdk.schemas.Schema.Options
- 
Get the value of an option. 
- getValue() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
- 
Retrieve the value of the display item. 
- getValue() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
- 
The value of the display item. 
- getValue() - Method in class org.apache.beam.sdk.transforms.join.RawUnionValue
-  
- getValue() - Method in class org.apache.beam.sdk.values.FailsafeValueInSingleWindow
- 
Returns the value of this FailsafeValueInSingleWindow.
 
- getValue() - Method in class org.apache.beam.sdk.values.KV
- 
Returns the value of this  KV. 
- getValue(int) - Method in class org.apache.beam.sdk.values.Row
- 
Get value by field index, ClassCastExceptionis thrown if schema doesn't match.
 
- getValue(String) - Method in class org.apache.beam.sdk.values.Row
- 
Get value by field name, ClassCastExceptionis thrown if type doesn't match.
 
- getValue(int) - Method in class org.apache.beam.sdk.values.RowWithGetters
-  
- getValue(int) - Method in class org.apache.beam.sdk.values.RowWithStorage
-  
- getValue() - Method in class org.apache.beam.sdk.values.TaggedPValue
- 
- getValue() - Method in class org.apache.beam.sdk.values.TimestampedValue
-  
- getValue() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow
- 
Returns the value of this ValueInSingleWindow.
 
- getValue() - Method in class org.apache.beam.sdk.values.ValueWithRecordId
-  
- getValueCaptureType() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
The capture type of the change stream that generated this record. 
- getValueClass() - Method in enum org.apache.beam.sdk.io.cdap.PluginConstants.Hadoop
-  
- getValueCoder() - Method in class org.apache.beam.sdk.coders.KvCoder
-  
- getValueCoder() - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
- 
Gets the value coder that will be prefixed by the length. 
- getValueCoder() - Method in class org.apache.beam.sdk.coders.MapCoder
-  
- getValueCoder() - Method in class org.apache.beam.sdk.coders.NullableCoder
- 
- getValueCoder() - Method in class org.apache.beam.sdk.coders.SortedMapCoder
-  
- getValueCoder() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
-  
- getValueCoder() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getValueCoder() - Method in class org.apache.beam.sdk.testing.TestStream
-  
- getValueCoder() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
-  
- getValueCoder() - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
-  
- getValueDeserializerProvider() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getValueOrDefault(String, T) - Method in class org.apache.beam.sdk.schemas.Schema.Options
- 
Get the value of an option. 
- getValues() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
-  
- getValues() - Method in class org.apache.beam.sdk.values.Row
- 
Return the list of raw unmodified data values to enable 0-copy code. 
- getValues() - Method in class org.apache.beam.sdk.values.RowWithGetters
- 
Return the list of raw unmodified data values to enable 0-copy code. 
- getValues() - Method in class org.apache.beam.sdk.values.RowWithStorage
-  
- getValueSerializer() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- getValuesMap() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
-  
- getValueTranslationFunction() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
-  
- getValueTypeDescriptor() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
-  
- getVerifyRowValues() - Method in interface org.apache.beam.sdk.extensions.sql.impl.BeamSqlPipelineOptions
-  
- getView() - Method in class org.apache.beam.runners.dataflow.CreateDataflowView
-  
- getView() - Method in class org.apache.beam.runners.dataflow.DataflowRunner.StreamingPCollectionViewWriterFn
-  
- getView() - Method in class org.apache.beam.sdk.transforms.View.CreatePCollectionView
- 
- getViewFn() - Method in interface org.apache.beam.sdk.values.PCollectionView
- 
- getViewFn() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
-  
- getVocabulary() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- getVocabulary() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- getWarehouse() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- getWarehouse() - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- getWarehouse() - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- getWarnings() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- getWatchInterval() - Method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
-  
- getWatchTopicPartitionDuration() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getWatermark() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-  
- getWatermark() - Method in interface org.apache.beam.sdk.io.aws2.kinesis.WatermarkPolicy
-  
- getWatermark() - Method in class org.apache.beam.sdk.io.aws2.kinesis.WatermarkPolicyFactory.ArrivalTimeWatermarkPolicy
-  
- getWatermark() - Method in class org.apache.beam.sdk.io.aws2.kinesis.WatermarkPolicyFactory.CustomWatermarkPolicy
-  
- getWatermark() - Method in class org.apache.beam.sdk.io.aws2.kinesis.WatermarkPolicyFactory.ProcessingTimeWatermarkPolicy
-  
- getWatermark() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.DetectNewPartitionsState
-  
- getWatermark() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.StreamPartitionWithWatermark
-  
- getWatermark() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.UnboundedReaderImpl
-  
- getWatermark() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata
- 
The time for which all records with a timestamp less than it have been processed. 
- getWatermark(TimestampPolicy.PartitionContext) - Method in class org.apache.beam.sdk.io.kafka.CustomTimestampPolicyWithLimitedDelay
-  
- getWatermark(TimestampPolicy.PartitionContext) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicy
- 
Returns watermark for the partition. 
- getWatermark(TimestampPolicy.PartitionContext) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.LogAppendTimePolicy
-  
- getWatermark(TimestampPolicy.PartitionContext) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.ProcessingTimePolicy
-  
- getWatermark(TimestampPolicy.PartitionContext) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.TimestampFnPolicy
-  
- getWatermark() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
- 
Returns a timestamp before or at the timestamps of all future elements read by this reader. 
- getWatermark() - Method in class org.apache.beam.sdk.testing.TestStream.WatermarkEvent
-  
- getWatermarkAndState() - Method in interface org.apache.beam.sdk.fn.splittabledofn.WatermarkEstimators.WatermarkAndStateObserver
-  
- getWatermarkCache() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.cache.CacheFactory
-  
- getWatermarkFn() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- getWatermarkIdleDurationThreshold() - Method in class org.apache.beam.sdk.io.solace.read.UnboundedSolaceSource
-  
- getWatermarkIndexName() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataTableNames
-  
- getWatermarkLastUpdated() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.DetectNewPartitionsState
-  
- getWatermarkMillis() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
-  
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterAll
-  
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterEach
-  
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterFirst
-  
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
-  
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
-  
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterSynchronizedProcessingTime
-  
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-  
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
-  
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.DefaultTrigger
-  
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.Never.NeverTrigger
-  
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
-  
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
-  
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.ReshuffleTrigger
-  
- getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
- 
For internal use only; no backwards-compatibility guarantees. 
- getWeigher(Coder<InputT>) - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches.BatchingParams
-  
- getWeight() - Method in class org.apache.beam.sdk.fn.data.WeightedList
-  
- getWindmillGetDataStreamCount() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getWindmillHarnessUpdateReportingPeriod() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getWindmillMessagesBetweenIsReadyChecks() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getWindmillRequestBatchedGetWorkResponse() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getWindmillServiceCommitThreads() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getWindmillServiceEndpoint() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
- 
Custom windmill service endpoint. 
- getWindmillServicePort() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getWindmillServiceRpcChannelAliveTimeoutSec() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getWindmillServiceStreamingLogEveryNStreamFailures() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getWindmillServiceStreamingRpcBatchLimit() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getWindmillServiceStreamingRpcHealthCheckPeriodMs() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getWindmillServiceStreamMaxBackoffMillis() - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- getWindow() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
- 
This method returns the number of tuples in each window. 
- getWindow() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-  
- getWindow() - Method in class org.apache.beam.sdk.values.FailsafeValueInSingleWindow
- 
Returns the window of this FailsafeValueInSingleWindow.
 
- getWindow() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow
- 
Returns the window of this ValueInSingleWindow.
 
- getWindowCoder() - Method in class org.apache.beam.sdk.coders.TimestampPrefixingWindowCoder
-  
- getWindowedAggregateDoFnOperator(FlinkStreamingTranslationContext, PTransform<PCollection<KV<K, InputT>>, PCollection<KV<K, OutputT>>>, KvCoder<K, InputAccumT>, Coder<WindowedValue<KV<K, OutputAccumT>>>, SystemReduceFn<K, InputAccumT, ?, OutputAccumT, BoundedWindow>, Map<Integer, PCollectionView<?>>, List<PCollectionView<?>>) - Static method in class org.apache.beam.runners.flink.FlinkStreamingAggregationsTranslators
- 
Create a DoFnOperator instance that group elements per window and apply a combine function on
 them. 
- getWindowedAggregateDoFnOperator(FlinkStreamingTranslationContext, PTransform<PCollection<KV<K, InputT>>, PCollection<KV<K, OutputT>>>, KvCoder<K, InputAccumT>, Coder<WindowedValue<KV<K, OutputAccumT>>>, CombineFnBase.GlobalCombineFn<? super InputAccumT, ?, OutputAccumT>, Map<Integer, PCollectionView<?>>, List<PCollectionView<?>>) - Static method in class org.apache.beam.runners.flink.FlinkStreamingAggregationsTranslators
-  
- getWindowedValueCoder(String, RunnerApi.Components) - Static method in class org.apache.beam.runners.fnexecution.translation.PipelineTranslatorUtils
-  
- getWindowedWrites() - Method in class org.apache.beam.sdk.io.WriteFiles
-  
- getWindowFn() - Method in class org.apache.beam.sdk.transforms.windowing.Window.Assign
-  
- getWindowFn() - Method in class org.apache.beam.sdk.transforms.windowing.Window
-  
- getWindowFn() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- getWindowingStrategy(String, RunnerApi.Components) - Static method in class org.apache.beam.runners.fnexecution.translation.PipelineTranslatorUtils
-  
- getWindowingStrategy() - Method in class org.apache.beam.sdk.values.PCollection
- 
- getWindowingStrategyInternal() - Method in interface org.apache.beam.sdk.values.PCollectionView
- 
- getWindowingStrategyInternal() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
- 
- getWindowMappingFn() - Method in interface org.apache.beam.sdk.values.PCollectionView
- 
For internal use only. 
- getWindowMappingFn() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
-  
- getWindowTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-  
- getWindowTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
- 
Returns a  TypeDescriptor capturing what is known statically about the window type of
 this  WindowFn instance's most-derived class. 
- getWithAutoSharding() - Method in class org.apache.beam.sdk.io.WriteFiles
-  
- getWithPartitions() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadConfiguration
-  
- getWorkCompleted() - Method in class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker.Progress
- 
The known amount of completed work. 
- getWorkerCacheMb() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
- 
The size of the worker's in-memory cache, in megabytes. 
- getWorkerCPUs() - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- getWorkerDiskType() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
- 
Specifies what type of persistent disk is used. 
- getWorkerHarnessContainerImage() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
- 
- getWorkerId() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
- 
The identity of the worker running this pipeline. 
- getWorkerId() - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClient
-  
- getWorkerLogLevelOverrides() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
- 
Deprecated. This option controls the log levels for specifically named loggers. 
- getWorkerMachineType() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
- 
Machine type to create Dataflow worker VMs as. 
- getWorkerPool() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
- 
The identity of the worker pool of this worker. 
- getWorkerRegion() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
- 
The Compute Engine region (https://cloud.google.com/compute/docs/regions-zones/regions-zones)
 in which worker processing should occur, e.g. 
- getWorkerSystemErrMessageLevel() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
- 
Deprecated. Controls the log level given to messages printed to System.err.
 
- getWorkerSystemOutMessageLevel() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
- 
Deprecated. Controls the log level given to messages printed to System.out.
 
- getWorkerZone() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
- 
The Compute Engine zone (https://cloud.google.com/compute/docs/regions-zones/regions-zones) in
 which worker processing should occur, e.g. 
- getWorkRemaining() - Method in class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker.Progress
- 
The known amount of work remaining. 
- getWritableByteChannelFactory() - Method in class org.apache.beam.sdk.io.FileBasedSink
- 
- getWrite() - Method in class org.apache.beam.io.requestresponse.Cache.Pair
-  
- getWrite() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.WriteFailure
-  
- getWriteCounterPrefix() - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.GcsCountersOptions
-  
- getWriteDisposition() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- getWriteDisposition() - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.WriteBuilder.Configuration
-  
- getWriteDisposition() - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceConfig
- 
- getWriteFailures() - Method in exception org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.FailedWritesException
- 
- getWriteOperation() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
- 
Return the WriteOperation that this Writer belongs to. 
- getWriteRecordsTransform() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
-  
- getWriteResult() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.WriteFailure
-  
- getWriteStatement() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration
-  
- getWriteStreamSchema(String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.WriteStreamService
-  
- getWriteStreamSchema(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.WriteStreamServiceImpl
-  
- getWriteStreamSchema(String) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- getWriteStreamService(BigQueryOptions) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices
- 
- getWriteStreamService(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl
-  
- getWriteStreamService(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
-  
- getXmlConfiguration() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration
-  
- getZetaSqlDefaultTimezone() - Method in interface org.apache.beam.sdk.extensions.sql.impl.BeamSqlPipelineOptions
-  
- getZetaSqlRuleSets() - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
-  
- getZetaSqlRuleSets(Collection<RelOptRule>) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
-  
- getZone() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
- 
- global(Map<Integer, GlobalWatermarkHolder.SparkWatermarks>) - Static method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
- 
Build a global TimerInternalsfor all feeding streams.
 
- Global() - Constructor for class org.apache.beam.sdk.schemas.transforms.Group.Global
-  
- GLOBAL_SEQUENCE_TRACKER - Static variable in class org.apache.beam.sdk.extensions.ordered.OrderedEventProcessor
-  
- GlobalConfigRefreshPeriodFactory() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions.GlobalConfigRefreshPeriodFactory
-  
- globalDefault() - Static method in class org.apache.beam.sdk.values.WindowingStrategy
- 
Return a fully specified, default windowing strategy. 
- GlobalDigest() - Constructor for class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.GlobalDigest
-  
- globally() - Static method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct
- 
Computes the approximate number of distinct elements in the input PCollection<InputT>and returns aPCollection<Long>.
 
- globally() - Static method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies
- 
Create the  PTransform that will build a Count-min sketch for keeping track of the
 frequency of the elements in the whole stream. 
- globally() - Static method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles
- 
Compute the stream in order to build a T-Digest structure (MergingDigest) for keeping track of
 the stream distribution and returns a PCollection<MergingDigest>.
 
- globally() - Static method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct
-  
- Globally() - Constructor for class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.Globally
-  
- globally() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Extract
- 
Returns a PTransformthat takes an inputPCollection<byte[]>of HLL++
 sketches and returns aPCollection<Long>of the estimated count of distinct elements
 extracted from each sketch.
 
- globally() - Method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Init.Builder
- 
Returns a Combine.GloballyPTransformthat takes an inputPCollection<InputT>and returns aPCollection<byte[]>which consists of the HLL++
 sketch computed from the elements in the inputPCollection.
 
- globally() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.MergePartial
- 
Returns a Combine.GloballyPTransformthat takes an inputPCollection<byte[]>of HLL++ sketches and returns aPCollection<byte[]>of a new
 sketch merged from the input sketches.
 
- globally() - Static method in class org.apache.beam.sdk.schemas.transforms.Group
- 
Returns a transform that groups all elements in the input  PCollection. 
- globally(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles
- 
Returns a PTransformthat takes aPCollection<T>and returns aPCollection<List<T>>whose single value is aListof the approximateN-tiles
 of the elements of the inputPCollection.
 
- globally(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles
- 
- globally(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateUnique
- 
Deprecated. Returns a PTransformthat takes aPCollection<T>and returns aPCollection<Long>containing a single value that is an estimate of the number of distinct
 elements in the inputPCollection.
 
- globally(double) - Static method in class org.apache.beam.sdk.transforms.ApproximateUnique
- 
Deprecated. 
- Globally(int) - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique.Globally
- 
Deprecated. 
- Globally(double) - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique.Globally
- 
Deprecated. 
- globally(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
- 
Returns a  Combine.GloballyPTransform that uses the given  SerializableFunction to combine all the elements in each window of the input  PCollection into a single value in the output  PCollection. 
- globally(SerializableBiFunction<V, V, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
- 
Returns a  Combine.GloballyPTransform that uses the given  SerializableBiFunction to combine all the elements in each window of the input  PCollection into a single value in the output  PCollection. 
- globally(CombineFnBase.GlobalCombineFn<? super InputT, ?, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Combine
- 
Returns a  Combine.GloballyPTransform that uses the given  GloballyCombineFn to combine all the elements in each window of the input  PCollection
 into a single value in the output  PCollection. 
- globally() - Static method in class org.apache.beam.sdk.transforms.Count
- 
- globally() - Static method in class org.apache.beam.sdk.transforms.Latest
- 
Returns a  PTransform that takes as input a  PCollection<T> and returns a  PCollection<T> whose contents is the latest element according to its event time, or null if there are no elements. 
- globally() - Static method in class org.apache.beam.sdk.transforms.Max
- 
Returns a PTransformthat takes an inputPCollection<T>and returns aPCollection<T>whose contents is the maximum according to the natural ordering ofTof
 the inputPCollection's elements, ornullif there are no elements.
 
- globally(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Max
- 
Returns a PTransformthat takes an inputPCollection<T>and returns aPCollection<T>whose contents is the maximum of the inputPCollection's elements, ornullif there are no elements.
 
- globally() - Static method in class org.apache.beam.sdk.transforms.Mean
- 
Returns a PTransformthat takes an inputPCollection<NumT>and returns aPCollection<Double>whose contents is the mean of the inputPCollection's elements, or0if there are no elements.
 
- globally() - Static method in class org.apache.beam.sdk.transforms.Min
- 
Returns a PTransformthat takes an inputPCollection<T>and returns aPCollection<T>whose contents is the minimum according to the natural ordering ofTof
 the inputPCollection's elements, ornullif there are no elements.
 
- globally(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Min
- 
Returns a PTransformthat takes an inputPCollection<T>and returns aPCollection<T>whose contents is the minimum of the inputPCollection's elements, ornullif there are no elements.
 
- GloballyDistinct() - Constructor for class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.GloballyDistinct
-  
- GlobalSketch() - Constructor for class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.GlobalSketch
-  
- GlobalWatermarkHolder - Class in org.apache.beam.runners.spark.util
- 
A store to hold the global watermarks for a micro-batch. 
- GlobalWatermarkHolder() - Constructor for class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
-  
- GlobalWatermarkHolder.SparkWatermarks - Class in org.apache.beam.runners.spark.util
- 
- GlobalWatermarkHolder.WatermarkAdvancingStreamingListener - Class in org.apache.beam.runners.spark.util
- 
Advance the WMs onBatchCompleted event. 
- GlobalWindow - Class in org.apache.beam.sdk.transforms.windowing
- 
The default window into which all data is placed (via  GlobalWindows). 
- GlobalWindow.Coder - Class in org.apache.beam.sdk.transforms.windowing
- 
- GlobalWindows - Class in org.apache.beam.sdk.transforms.windowing
- 
A  WindowFn that assigns all data to the same window. 
- GlobalWindows() - Constructor for class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-  
- GoogleAdsClientFactory - Interface in org.apache.beam.sdk.io.googleads
- 
Defines how to construct a GoogleAdsClient.
 
- GoogleAdsCredentialsFactory() - Constructor for class org.apache.beam.sdk.io.googleads.GoogleAdsOptions.GoogleAdsCredentialsFactory
-  
- GoogleAdsIO - Class in org.apache.beam.sdk.io.googleads
- 
- GoogleAdsOptions - Interface in org.apache.beam.sdk.io.googleads
- 
Options used to configure Google Ads API specific options. 
- GoogleAdsOptions.GoogleAdsCredentialsFactory - Class in org.apache.beam.sdk.io.googleads
- 
Attempts to load the Google Ads credentials. 
- GoogleAdsUserCredentialFactory - Class in org.apache.beam.sdk.io.googleads
- 
Constructs and returns Credentialsto be used by Google Ads API calls.
 
- GoogleAdsV17 - Class in org.apache.beam.sdk.io.googleads
- 
GoogleAdsV17 provides an API to read Google Ads API v17 reports. 
- GoogleAdsV17.RateLimitPolicy - Interface in org.apache.beam.sdk.io.googleads
- 
This interface can be used to implement custom client-side rate limiting policies. 
- GoogleAdsV17.RateLimitPolicyFactory - Interface in org.apache.beam.sdk.io.googleads
- 
- GoogleAdsV17.Read - Class in org.apache.beam.sdk.io.googleads
- 
A  PTransform that reads the results of a Google Ads query as  GoogleAdsRow
 objects. 
- GoogleAdsV17.ReadAll - Class in org.apache.beam.sdk.io.googleads
- 
A  PTransform that reads the results of many  SearchGoogleAdsStreamRequest
 objects as  GoogleAdsRow objects. 
- GoogleAdsV17.SimpleRateLimitPolicy - Class in org.apache.beam.sdk.io.googleads
- 
This rate limit policy wraps a RateLimiterand can be used in low volume and
 development use cases as a client-side rate limiting policy.
 
- GoogleApiDebugOptions - Interface in org.apache.beam.sdk.extensions.gcp.options
- 
These options configure debug settings for Google API clients created within the Apache Beam SDK. 
- GoogleApiDebugOptions.GoogleApiTracer - Class in org.apache.beam.sdk.extensions.gcp.options
- 
- GoogleApiTracer() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions.GoogleApiTracer
-  
- GraphiteSink - Class in org.apache.beam.runners.spark.metrics.sink
- 
- GraphiteSink(Properties, MetricRegistry, SecurityManager) - Constructor for class org.apache.beam.runners.spark.metrics.sink.GraphiteSink
- 
Constructor for Spark 3.1.x and earlier. 
- GraphiteSink(Properties, MetricRegistry) - Constructor for class org.apache.beam.runners.spark.metrics.sink.GraphiteSink
- 
Constructor for Spark 3.2.x and later. 
- greaterThan(T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- greaterThan(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- greaterThan(T) - Static method in class org.apache.beam.sdk.transforms.Filter
- 
Returns a PTransformthat takes an inputPCollection<T>and returns aPCollection<T>with elements that are greater than a given value, based on the elements'
 natural ordering.
 
- greaterThanEq(T) - Static method in class org.apache.beam.sdk.transforms.Filter
- 
Returns a PTransformthat takes an inputPCollection<T>and returns aPCollection<T>with elements that are greater than or equal to a given value, based on the
 elements' natural ordering.
 
- greaterThanOrEqualTo(T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- greaterThanOrEqualTo(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- Group - Class in org.apache.beam.sdk.schemas.transforms
- 
- Group() - Constructor for class org.apache.beam.sdk.schemas.transforms.Group
-  
- Group.AggregateCombiner<InputT> - Class in org.apache.beam.sdk.schemas.transforms
- 
a  PTransform that does a combine using an aggregation built up by calls to
 aggregateField and aggregateFields. 
- Group.ByFields<InputT> - Class in org.apache.beam.sdk.schemas.transforms
- 
a  PTransform that groups schema elements based on the given fields. 
- Group.CombineFieldsByFields<InputT> - Class in org.apache.beam.sdk.schemas.transforms
- 
a  PTransform that does a per-key combine using an aggregation built up by calls to
 aggregateField and aggregateFields. 
- Group.CombineFieldsByFields.Fanout - Class in org.apache.beam.sdk.schemas.transforms
-  
- Group.CombineFieldsByFields.Fanout.Kind - Enum in org.apache.beam.sdk.schemas.transforms
-  
- Group.CombineFieldsGlobally<InputT> - Class in org.apache.beam.sdk.schemas.transforms
- 
a  PTransform that does a global combine using an aggregation built up by calls to
 aggregateField and aggregateFields. 
- Group.CombineGlobally<InputT,OutputT> - Class in org.apache.beam.sdk.schemas.transforms
- 
- Group.Global<InputT> - Class in org.apache.beam.sdk.schemas.transforms
- 
A  PTransform for doing global aggregations on schema PCollections. 
- GroupAlsoByWindowViaOutputBufferFn<K,InputT,W extends BoundedWindow> - Class in org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions
- 
A FlatMap function that groups by windows in batch mode using ReduceFnRunner.
 
- GroupAlsoByWindowViaOutputBufferFn(WindowingStrategy<?, W>, StateInternalsFactory<K>, SystemReduceFn<K, InputT, Iterable<InputT>, Iterable<InputT>, W>, Supplier<PipelineOptions>) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.GroupAlsoByWindowViaOutputBufferFn
-  
- GroupByKey<K,V> - Class in org.apache.beam.sdk.transforms
- 
GroupByKey<K, V>takes aPCollection<KV<K, V>>, groups the values by key and
 windows, and returns aPCollection<KV<K, Iterable<V>>>representing a map from each
 distinct key and window of the inputPCollectionto anIterableover all the
 values associated with that key in the input per window.
 
- groupByKeyAndWindow(JavaDStream<WindowedValue<KV<K, InputT>>>, Coder<K>, Coder<WindowedValue<InputT>>, WindowingStrategy<?, W>, SerializablePipelineOptions, List<Integer>, String) - Static method in class org.apache.beam.runners.spark.stateful.SparkGroupAlsoByWindowViaWindowSet
-  
- GroupByKeyTranslatorBatch<K,V> - Class in org.apache.beam.runners.twister2.translators.batch
- 
GroupByKey translator. 
- GroupByKeyTranslatorBatch() - Constructor for class org.apache.beam.runners.twister2.translators.batch.GroupByKeyTranslatorBatch
-  
- GroupByWindowFunction<K,V,W extends BoundedWindow> - Class in org.apache.beam.runners.twister2.translators.functions
- 
GroupBy window function. 
- GroupByWindowFunction() - Constructor for class org.apache.beam.runners.twister2.translators.functions.GroupByWindowFunction
-  
- GroupByWindowFunction(WindowingStrategy<?, W>, SystemReduceFn<K, V, Iterable<V>, Iterable<V>, W>, PipelineOptions) - Constructor for class org.apache.beam.runners.twister2.translators.functions.GroupByWindowFunction
-  
- grouped() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
- groupedValues(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
- 
Returns a  Combine.GroupedValuesPTransform that takes a  PCollection of  KVs where a key maps to an  Iterable of values, e.g., the result
 of a  GroupByKey, then uses the given  SerializableFunction to combine all the
 values associated with a key, ignoring the key. 
- groupedValues(SerializableBiFunction<V, V, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
- 
Returns a  Combine.GroupedValuesPTransform that takes a  PCollection of  KVs where a key maps to an  Iterable of values, e.g., the result
 of a  GroupByKey, then uses the given  SerializableFunction to combine all the
 values associated with a key, ignoring the key. 
- groupedValues(CombineFnBase.GlobalCombineFn<? super InputT, ?, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Combine
- 
Returns a  Combine.GroupedValuesPTransform that takes a  PCollection of  KVs where a key maps to an  Iterable of values, e.g., the result
 of a  GroupByKey, then uses the given  CombineFn to combine all the values
 associated with a key, ignoring the key. 
- GroupingState<InputT,OutputT> - Interface in org.apache.beam.sdk.state
- 
A  ReadableState cell that combines multiple input values and outputs a single value of a
 different type. 
- GroupIntoBatches<K,InputT> - Class in org.apache.beam.sdk.transforms
- 
A  PTransform that batches inputs to a desired batch size. 
- GroupIntoBatches.BatchingParams<InputT> - Class in org.apache.beam.sdk.transforms
- 
Wrapper class for batching parameters supplied by users. 
- GroupIntoBatches.WithShardedKey - Class in org.apache.beam.sdk.transforms
-  
- GroupIntoBatchesOverride - Class in org.apache.beam.runners.dataflow
-  
- GroupIntoBatchesOverride() - Constructor for class org.apache.beam.runners.dataflow.GroupIntoBatchesOverride
-  
- GrowableOffsetRangeTracker - Class in org.apache.beam.sdk.transforms.splittabledofn
- 
- GrowableOffsetRangeTracker(long, GrowableOffsetRangeTracker.RangeEndEstimator) - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.GrowableOffsetRangeTracker
-  
- GrowableOffsetRangeTracker.RangeEndEstimator - Interface in org.apache.beam.sdk.transforms.splittabledofn
- 
Provides the estimated end offset of the range. 
- Growth() - Constructor for class org.apache.beam.sdk.transforms.Watch.Growth
-  
- growthOf(Watch.Growth.PollFn<InputT, OutputT>, Requirements) - Static method in class org.apache.beam.sdk.transforms.Watch
- 
Watches the growth of the given poll function. 
- growthOf(Watch.Growth.PollFn<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Watch
- 
Watches the growth of the given poll function. 
- growthOf(Contextful<Watch.Growth.PollFn<InputT, OutputT>>, SerializableFunction<OutputT, KeyT>) - Static method in class org.apache.beam.sdk.transforms.Watch
- 
Watches the growth of the given poll function, using the given "key function" to deduplicate
 outputs. 
- GrpcContextHeaderAccessorProvider - Class in org.apache.beam.sdk.fn.server
- 
A HeaderAccessorProvider which intercept the header in a GRPC request and expose the relevant
 fields. 
- GrpcContextHeaderAccessorProvider() - Constructor for class org.apache.beam.sdk.fn.server.GrpcContextHeaderAccessorProvider
-  
- GrpcDataService - Class in org.apache.beam.runners.fnexecution.data
- 
- GrpcDataService() - Constructor for class org.apache.beam.runners.fnexecution.data.GrpcDataService
- 
- GrpcFnServer<ServiceT extends FnService> - Class in org.apache.beam.sdk.fn.server
- 
A  gRPC Server which manages a single  FnService. 
- GrpcLoggingService - Class in org.apache.beam.runners.fnexecution.logging
- 
An implementation of the Beam Fn Logging Service over gRPC. 
- GrpcStateService - Class in org.apache.beam.runners.fnexecution.state
- 
An implementation of the Beam Fn State service. 
- guessExpressionType(String, Map<String, Type>) - Static method in class org.apache.beam.sdk.schemas.transforms.providers.StringCompiler
-  
- ICEBERG - Static variable in class org.apache.beam.sdk.managed.Managed
-  
- IcebergCatalogConfig - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergCatalogConfig() - Constructor for class org.apache.beam.sdk.io.iceberg.IcebergCatalogConfig
-  
- IcebergCatalogConfig.Builder - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergDestination - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergDestination() - Constructor for class org.apache.beam.sdk.io.iceberg.IcebergDestination
-  
- IcebergDestination.Builder - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergIO - Class in org.apache.beam.sdk.io.iceberg
- 
- IcebergIO() - Constructor for class org.apache.beam.sdk.io.iceberg.IcebergIO
-  
- IcebergIO.ReadRows - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergIO.WriteRows - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergReadSchemaTransformProvider - Class in org.apache.beam.sdk.io.iceberg
- 
- IcebergReadSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.iceberg.IcebergReadSchemaTransformProvider
-  
- icebergRecordToBeamRow(Schema, Record) - Static method in class org.apache.beam.sdk.io.iceberg.IcebergUtils
- 
Converts an Iceberg  Record to a Beam  Row. 
- IcebergScanConfig - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergScanConfig() - Constructor for class org.apache.beam.sdk.io.iceberg.IcebergScanConfig
-  
- IcebergScanConfig.Builder - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergScanConfig.ScanType - Enum in org.apache.beam.sdk.io.iceberg
-  
- icebergSchemaToBeamSchema(Schema) - Static method in class org.apache.beam.sdk.io.iceberg.IcebergUtils
- 
Converts an Iceberg  Schema to a Beam  Schema. 
- IcebergSchemaTransformTranslation - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergSchemaTransformTranslation() - Constructor for class org.apache.beam.sdk.io.iceberg.IcebergSchemaTransformTranslation
-  
- IcebergSchemaTransformTranslation.ReadRegistrar - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergSchemaTransformTranslation.WriteRegistrar - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergTableCreateConfig - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergTableCreateConfig() - Constructor for class org.apache.beam.sdk.io.iceberg.IcebergTableCreateConfig
-  
- IcebergTableCreateConfig.Builder - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergUtils - Class in org.apache.beam.sdk.io.iceberg
- 
Utilities for converting between Beam and Iceberg types, made public for user's convenience. 
- IcebergWriteResult - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergWriteSchemaTransformProvider - Class in org.apache.beam.sdk.io.iceberg
- 
- IcebergWriteSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider
-  
- IcebergWriteSchemaTransformProvider.Configuration - Class in org.apache.beam.sdk.io.iceberg
-  
- IcebergWriteSchemaTransformProvider.Configuration.Builder - Class in org.apache.beam.sdk.io.iceberg
-  
- id() - Method in class org.apache.beam.runners.jobsubmission.JobPreparation
-  
- id - Variable in enum org.apache.beam.sdk.io.kafka.KafkaTimestampType
-  
- identifier() - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.expansion.service.WindowIntoTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSchemaIOProvider
- 
Returns an id that uniquely represents this IO. 
- identifier() - Method in class org.apache.beam.sdk.extensions.avro.schemas.io.payloads.AvroPayloadSerializerProvider
-  
- identifier() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoPayloadSerializerProvider
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.Fixed32
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.Fixed64
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.SFixed32
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.SFixed64
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.SInt32
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.SInt64
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.UInt32
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.UInt64
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.LocalMktDate
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.TZTimeOnly
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.TZTimestamp
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.Uint16
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.Uint32
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.Uint64
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.Uint8
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.UTCDateOnly
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.UTCTimeOnly
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.UTCTimestamp
-  
- identifier() - Method in class org.apache.beam.sdk.extensions.sql.expansion.SqlTransformSchemaTransformProvider
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.TimeWithLocalTzType
-  
- identifier() - Method in class org.apache.beam.sdk.io.csv.providers.CsvWriteTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.fileschematransform.AvroReadSchemaTransformFormatProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.fileschematransform.AvroWriteSchemaTransformFormatProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.fileschematransform.CsvWriteSchemaTransformFormatProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformProvider
- 
- identifier() - Method in class org.apache.beam.sdk.io.fileschematransform.JsonReadSchemaTransformFormatProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.fileschematransform.JsonWriteSchemaTransformFormatProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.fileschematransform.LineReadSchemaTransformFormatProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.fileschematransform.ParquetReadSchemaTransformFormatProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.fileschematransform.ParquetWriteSchemaTransformFormatProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.fileschematransform.XmlWriteSchemaTransformFormatProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryDlqProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryExportReadSchemaTransformProvider
- 
- identifier() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySchemaIOProvider
- 
Returns an id that uniquely represents this IO. 
- identifier() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryFileLoadsSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.datastore.DataStoreV1SchemaIOProvider
- 
Returns an id that uniquely represents this IO. 
- identifier() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubDlqProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubSchemaIOProvider
- 
Returns an id that uniquely represents this IO. 
- identifier() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.DlqProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.iceberg.IcebergReadSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.jdbc.JdbcSchemaIOProvider
- 
Returns an id that uniquely represents this IO. 
- identifier() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.jdbc.providers.ReadFromMySqlSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.jdbc.providers.ReadFromOracleSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.jdbc.providers.ReadFromPostgresSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.jdbc.providers.ReadFromSqlServerSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.jdbc.providers.WriteToMySqlSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.jdbc.providers.WriteToOracleSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.jdbc.providers.WriteToPostgresSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.jdbc.providers.WriteToSqlServerSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.json.providers.JsonWriteTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadProvider
- 
- identifier() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteProvider
- 
- identifier() - Method in class org.apache.beam.sdk.io.thrift.ThriftPayloadSerializerProvider
-  
- identifier() - Method in class org.apache.beam.sdk.managed.ManagedSchemaTransformProvider
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.managed.testing.TestSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.managed.testing.TestSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.schemas.io.payloads.JsonPayloadSerializerProvider
-  
- identifier() - Method in interface org.apache.beam.sdk.schemas.io.Providers.Identifyable
- 
Returns an id that uniquely represents this among others implementing its derived interface. 
- identifier() - Method in interface org.apache.beam.sdk.schemas.io.SchemaIOProvider
- 
Returns an id that uniquely represents this IO. 
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.Date
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.DateTime
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.FixedPrecisionNumeric
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.FixedString
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.MicrosInstant
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.NanosDuration
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.NanosInstant
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.PythonCallable
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.SchemaLogicalType
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.Time
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.UuidLogicalType
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.VariableBytes
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.VariableString
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- IDENTIFIER() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifiedComponentContext
-  
- IDENTIFIER() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.SimpleIdentifierContext
-  
- identifier() - Method in class org.apache.beam.sdk.schemas.transforms.providers.FlattenTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaExplodeTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaFilterTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider
-  
- identifier() - Method in class org.apache.beam.sdk.schemas.transforms.providers.LoggingTransformProvider
-  
- identifier() - Method in interface org.apache.beam.sdk.schemas.transforms.SchemaTransformProvider
- 
Returns an id that uniquely represents this transform. 
- Identifier() - Constructor for class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
-  
- identity() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
- 
Returns the identity element of this operation, i.e. 
- identity() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
- 
Returns the value that should be used for the combine of the empty set. 
- identity() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
- 
Returns the identity element of this operation, i.e. 
- identity() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
- 
Returns the identity element of this operation, i.e. 
- identity() - Static method in class org.apache.beam.sdk.transforms.SerializableFunctions
-  
- IDENTITY_ELEMENT - Static variable in class org.apache.beam.sdk.metrics.DistributionResult
- 
The IDENTITY_ELEMENT is used to start accumulating distributions. 
- IdGenerator - Interface in org.apache.beam.sdk.fn
- 
A generator of unique IDs. 
- IdGenerators - Class in org.apache.beam.sdk.fn
- 
- IdGenerators() - Constructor for class org.apache.beam.sdk.fn.IdGenerators
-  
- ignored() - Static method in interface org.apache.beam.runners.fnexecution.control.BundleProgressHandler
- 
Returns a handler that ignores metrics. 
- ignoreInput(Watch.Growth.TerminationCondition<?, StateT>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
- 
Wraps a given input-independent  Watch.Growth.TerminationCondition as an equivalent condition with
 a given input type, passing  null to the original condition as input. 
- ignoreInsertIds() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Setting this option to true disables insertId based data deduplication offered by BigQuery. 
- ignoreUnknownValues() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Accept rows that contain values that do not match the schema. 
- immediate(T) - Static method in class org.apache.beam.sdk.state.ReadableStates
- 
A  ReadableState constructed from a constant value, hence immediately available. 
- immutableNames() - Method in class org.apache.beam.sdk.metrics.MetricsFilter
-  
- immutableNamesBuilder() - Method in class org.apache.beam.sdk.metrics.MetricsFilter.Builder
-  
- immutableSteps() - Method in class org.apache.beam.sdk.metrics.MetricsFilter
-  
- immutableStepsBuilder() - Method in class org.apache.beam.sdk.metrics.MetricsFilter.Builder
-  
- implement(EnumerableRelImplementor, EnumerableRel.Prefer) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter
-  
- implementor() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamSqlUnparseContext
-  
- importCatalogItems() - Static method in class org.apache.beam.sdk.extensions.ml.RecommendationAIIO
-  
- importFhirResource(String, String, String) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Importing a FHIR resource from GCS. 
- importFhirResource(String, String, String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- importResources(String, String, String, FhirIO.Import.ContentStructure) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO
- 
Import resources. 
- importResources(ValueProvider<String>, ValueProvider<String>, ValueProvider<String>, FhirIO.Import.ContentStructure) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO
- 
Import resources. 
- importUserEvents() - Static method in class org.apache.beam.sdk.extensions.ml.RecommendationAIIO
-  
- Impulse - Class in org.apache.beam.sdk.transforms
- 
For internal use only; no backwards-compatibility guarantees. 
- IMPULSE_ELEMENT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- ImpulseP - Class in org.apache.beam.runners.jet.processors
- 
/** * Jet Processorimplementation for Beam's Impulse primitive.
 
- ImpulseSource - Class in org.apache.beam.runners.twister2.translators.functions
- 
A SourceFuncwhich executes the impulse transform contract.
 
- ImpulseSource() - Constructor for class org.apache.beam.runners.twister2.translators.functions.ImpulseSource
-  
- ImpulseTranslatorBatch - Class in org.apache.beam.runners.twister2.translators.batch
- 
Impulse translator. 
- ImpulseTranslatorBatch() - Constructor for class org.apache.beam.runners.twister2.translators.batch.ImpulseTranslatorBatch
-  
- in(Pipeline, PCollection<FhirBundleResponse>, PCollection<HealthcareIOError<FhirBundleParameter>>) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.ExecuteBundlesResult
- 
Entry point for the ExecuteBundlesResult, storing the successful and failed bundles and their
 metadata. 
- in(Pipeline, PCollection<Solace.PublishResult>) - Static method in class org.apache.beam.sdk.io.solace.write.SolaceOutput
-  
- in(Pipeline) - Static method in class org.apache.beam.sdk.values.PBegin
- 
- in(Pipeline) - Static method in class org.apache.beam.sdk.values.PDone
- 
- IN_ARRAY_OPERATOR - Static variable in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamBigQuerySqlDialect
-  
- inc() - Method in class org.apache.beam.runners.jet.metrics.CounterImpl
-  
- inc(long) - Method in class org.apache.beam.runners.jet.metrics.CounterImpl
-  
- inc() - Method in interface org.apache.beam.sdk.metrics.Counter
- 
Increment the counter. 
- inc(long) - Method in interface org.apache.beam.sdk.metrics.Counter
- 
Increment the counter by the given amount. 
- inc() - Method in class org.apache.beam.sdk.metrics.DelegatingCounter
- 
Increment the counter. 
- inc(long) - Method in class org.apache.beam.sdk.metrics.DelegatingCounter
- 
Increment the counter by the given amount. 
- inc() - Method in class org.apache.beam.sdk.metrics.NoOpCounter
-  
- inc(long) - Method in class org.apache.beam.sdk.metrics.NoOpCounter
-  
- incActivePartitionReadCounter() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
- incChangeStreamMutationGcCounter() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
- incChangeStreamMutationUserCounter() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
- incClosestreamCount() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
- incDataRecordCounter() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
- incHeartbeatCount() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
- incHeartbeatRecordCount() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
- incListPartitionsCount() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
- include(String, HasDisplayData) - Method in interface org.apache.beam.sdk.transforms.display.DisplayData.Builder
- 
Register display data from the specified subcomponent at the given path. 
- inCombinedNonLatePanes(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
- 
Creates a new  PAssert.IterableAssert like this one, but with the assertion restricted to only
 run on the provided window across all panes that were not produced by the arrival of late
 data. 
- inCombinedNonLatePanes(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-  
- IncomingMessage() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.IncomingMessage
-  
- IncompatibleWindowException - Exception in org.apache.beam.sdk.transforms.windowing
- 
- IncompatibleWindowException(WindowFn<?, ?>, String) - Constructor for exception org.apache.beam.sdk.transforms.windowing.IncompatibleWindowException
-  
- incomplete(List<TimestampedValue<OutputT>>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
- 
Constructs a  Watch.Growth.PollResult with the given outputs and declares that new outputs might
 appear for the current input. 
- incomplete(Instant, List<OutputT>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
- 
- incOrphanedNewPartitionCleanedCount() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
- incPartitionMergeCount() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
- incPartitionReconciledWithoutTokenCount() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
- incPartitionReconciledWithTokenCount() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
- incPartitionRecordCount() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
- incPartitionRecordMergeCount() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
- incPartitionRecordSplitCount() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
- incPartitionSplitCount() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
- incPartitionStreamCount() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
- incQueryCounter() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
- increment() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
- 
Returns a RandomAccessData that is the smallest value of same length which is strictly greater
 than this. 
- increment(Long) - Method in class org.apache.beam.sdk.extensions.sql.provider.UdfTestProvider.IncrementFn
-  
- incrementAll(Date) - Method in class org.apache.beam.sdk.extensions.sql.provider.UdfTestProvider.DateIncrementAllFn
-  
- IncrementFn() - Constructor for class org.apache.beam.sdk.extensions.sql.provider.UdfTestProvider.IncrementFn
-  
- incrementingLongs() - Static method in class org.apache.beam.sdk.fn.IdGenerators
- 
Returns an  IdGenerator which provides successive incrementing longs. 
- index() - Method in class org.apache.beam.sdk.io.splunk.SplunkEvent
-  
- INDEX_OF_MAX - Static variable in class org.apache.beam.sdk.io.ShardNameTemplate
- 
Shard name containing the index and max. 
- indexOf(String) - Method in class org.apache.beam.sdk.schemas.Schema
- 
Find the index of a given field. 
- indexOfProjectionColumnRef(long, List<ResolvedColumn>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
- 
Return an index of the projection column reference. 
- inEarlyGlobalWindowPanes() - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
- 
- inEarlyGlobalWindowPanes() - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-  
- inEarlyPane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
- 
Creates a new  PAssert.IterableAssert like this one, but with the assertion restricted to only
 run on the provided window across all panes that were produced by the arrival of early data. 
- inEarlyPane(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-  
- inEarlyPane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
- 
Creates a new  PAssert.SingletonAssert like this one, but with the assertion restricted to
 only run on the provided window, running the checker only on early panes for each key. 
- InferableFunction<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
- 
- InferableFunction() - Constructor for class org.apache.beam.sdk.transforms.InferableFunction
-  
- InferableFunction(ProcessFunction<InputT, OutputT>) - Constructor for class org.apache.beam.sdk.transforms.InferableFunction
-  
- inferBeamSchema(DataSource, String) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
-  
- inferType(Object) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
- 
- inFinalPane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
- 
Creates a new  PAssert.IterableAssert like this one, but with the assertion restricted to only
 run on the provided window, running the checker only on the final pane for each key. 
- inFinalPane(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-  
- inFinalPane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
- 
Creates a new  PAssert.SingletonAssert like this one, but with the assertion restricted to
 only run on the provided window, running the checker only on the final pane for each key. 
- InfluxDbIO - Class in org.apache.beam.sdk.io.influxdb
- 
IO to read and write from InfluxDB. 
- InfluxDbIO.DataSourceConfiguration - Class in org.apache.beam.sdk.io.influxdb
- 
A POJO describing a DataSourceConfiguration such as URL, userName and password. 
- InfluxDbIO.Read - Class in org.apache.beam.sdk.io.influxdb
- 
A  PTransform to read from InfluxDB metric or data related to query. 
- InfluxDbIO.Write - Class in org.apache.beam.sdk.io.influxdb
- 
- ingestHL7v2Message(String, Message) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Ingest an HL7v2 message. 
- ingestHL7v2Message(String, Message) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- ingestMessages(String) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO
- 
Write with Messages.Ingest method. 
- INHERIT_IO_FILE - Static variable in class org.apache.beam.runners.fnexecution.environment.ProcessManager
- 
A symbolic file to indicate that we want to inherit I/O of parent process. 
- inheritedDescription(String, String, String, int) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider
-  
- inheritedDescription(String, String, String, int) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider
-  
- init(Processor.Context) - Method in class org.apache.beam.runners.jet.processors.BoundedSourceP
-  
- init(Processor.Context) - Method in class org.apache.beam.runners.jet.processors.UnboundedSourceP
-  
- init(SparkPipelineOptions, JavaSparkContext) - Static method in class org.apache.beam.runners.spark.metrics.MetricsAccumulator
- 
Init metrics accumulator if it has not been initiated. 
- init(ResultSetMetaData) - Method in interface org.apache.beam.sdk.io.singlestore.SingleStoreIO.RowMapperWithInit
-  
- initAccumulators(SparkPipelineOptions, JavaSparkContext) - Static method in class org.apache.beam.runners.spark.SparkRunner
- 
Init Metrics/Aggregators accumulators. 
- initClient() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Deidentify.DeidentifyFn
-  
- initClient() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Export.ExportResourcesFn
-  
- initContext(Class<?>) - Static method in class org.apache.beam.sdk.io.cdap.Plugin
- 
- initialBackoff() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
-  
- initialize(HttpRequest) - Method in class org.apache.beam.sdk.extensions.gcp.auth.NullCredentialInitializer
-  
- initialize(AbstractGoogleClientRequest<?>) - Method in class org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions.GoogleApiTracer
-  
- initialize(HttpRequest) - Method in class org.apache.beam.sdk.extensions.gcp.util.LatencyRecordingHttpRequestInitializer
-  
- initialize(HttpRequest) - Method in class org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer
-  
- initialize(HttpRequest) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient.AuthenticatedRetryInitializer
-  
- InitializeDoFn - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn
- 
A DoFn responsible to initialize the metadata table and prepare it for managing the state of the
 pipeline. 
- InitializeDoFn(DaoFactory, Instant, BigtableIO.ExistingPipelineOptions) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.InitializeDoFn
-  
- InitializeDoFn - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn
- 
A DoFn responsible for initializing the change stream Connector. 
- InitializeDoFn(DaoFactory, MapperFactory, Timestamp, Timestamp) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.InitializeDoFn
-  
- initializeWriteSessionProperties(SolaceIO.SubmissionMode) - Method in class org.apache.beam.sdk.io.solace.broker.SessionService
- 
This method will be called by the write connector when a new session is started. 
- InitialPartition - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.model
- 
Utility class to determine initial partition constants and methods. 
- InitialPartition() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.InitialPartition
-  
- InitialPipelineState - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.model
- 
- InitialPipelineState(Instant, boolean) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.InitialPipelineState
-  
- initialRestriction() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.DetectNewPartitionsDoFn
-  
- initialRestriction() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.ReadChangeStreamPartitionDoFn
-  
- initialRestriction(PartitionMetadata) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.DetectNewPartitionsDoFn
- 
- initialRestriction(PartitionMetadata) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.ReadChangeStreamPartitionDoFn
- 
The restriction for a partition will be defined from the start and end timestamp to query the
 partition for. 
- initialSystemTimeAt(Instant) - Method in class org.apache.beam.runners.spark.io.CreateStream
- 
Set the initial synchronized processing time. 
- initPluginType(Class<?>) - Static method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Gets value of a plugin type. 
- initPulsarClients() - Method in class org.apache.beam.sdk.io.pulsar.ReadFromPulsarDoFn
-  
- InjectPackageStrategy(Class<?>) - Constructor for class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.InjectPackageStrategy
-  
- inLatePane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
- 
Creates a new  PAssert.IterableAssert with the assertion restricted to only run on the
 provided window across all panes that were produced by the arrival of late data. 
- inLatePane(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-  
- inLatePane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
- 
Creates a new  PAssert.SingletonAssert with the assertion restricted to only run on the
 provided window, running the checker only on late panes for each key. 
- inMemory(TableProvider...) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
- 
This method creates  BeamSqlEnv using empty *
 Pipeline Options. 
- inMemory() - Method in class org.apache.beam.sdk.transforms.View.AsList
- 
Returns a PCollection view like this one, but whose resulting list will be entirely cached in
 memory. 
- inMemory(boolean) - Method in class org.apache.beam.sdk.transforms.View.AsList
- 
Returns a PCollection view like this one, but whose resulting list will be entirely cached in
 memory according to the input parameter. 
- inMemory() - Method in class org.apache.beam.sdk.transforms.View.AsMap
- 
Returns a PCollection view like this one, but whose resulting map will be entirely cached in
 memory. 
- inMemory(boolean) - Method in class org.apache.beam.sdk.transforms.View.AsMap
- 
Returns a PCollection view like this one, but whose resulting map will be entirely cached in
 memory according to the input parameter. 
- inMemory() - Method in class org.apache.beam.sdk.transforms.View.AsMultimap
- 
Returns a PCollection view like this one, but whose resulting map will be entirely cached in
 memory. 
- inMemory(boolean) - Method in class org.apache.beam.sdk.transforms.View.AsMultimap
- 
Returns a PCollection view like this one, but whose resulting map will be entirely cached in
 memory according to the input parameter. 
- InMemoryBagUserStateFactory<K,V,W extends BoundedWindow> - Class in org.apache.beam.runners.fnexecution.state
- 
Holds user state in memory. 
- InMemoryBagUserStateFactory() - Constructor for class org.apache.beam.runners.fnexecution.state.InMemoryBagUserStateFactory
-  
- inMemoryFinalizer(InstructionRequestHandler) - Static method in class org.apache.beam.runners.fnexecution.control.BundleFinalizationHandlers
- 
A bundle finalizer that stores all bundle finalization requests in memory. 
- InMemoryJobService - Class in org.apache.beam.runners.jobsubmission
- 
A InMemoryJobService that prepares and runs jobs on behalf of a client using a  JobInvoker. 
- InMemoryListFromMultimapViewFn(PCollectionViews.TypeDescriptorSupplier<T>) - Constructor for class org.apache.beam.sdk.values.PCollectionViews.InMemoryListFromMultimapViewFn
-  
- inMemoryListView(PCollection<T>, PCollectionViews.TypeDescriptorSupplier<T>, WindowingStrategy<?, W>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
- 
Returns a  PCollectionView<List<T>> capable of processing elements windowed using the
 provided  WindowingStrategy. 
- InMemoryListViewFn(PCollectionViews.TypeDescriptorSupplier<T>) - Constructor for class org.apache.beam.sdk.values.PCollectionViews.InMemoryListViewFn
-  
- inMemoryListViewUsingVoidKey(PCollection<KV<Void, T>>, PCollectionViews.TypeDescriptorSupplier<T>, WindowingStrategy<?, W>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
- 
Returns a  PCollectionView<List<T>> capable of processing elements windowed using the
 provided  WindowingStrategy. 
- InMemoryMapFromVoidKeyViewFn(Coder<K>, Coder<V>) - Constructor for class org.apache.beam.sdk.values.PCollectionViews.InMemoryMapFromVoidKeyViewFn
-  
- inMemoryMapView(PCollection<KV<K, V>>, Coder<K>, Coder<V>, WindowingStrategy<?, W>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
- 
Returns a  PCollectionView<Map<K, V>> capable of processing elements windowed using the
 provided  WindowingStrategy. 
- InMemoryMapViewFn(Coder<K>, Coder<V>) - Constructor for class org.apache.beam.sdk.values.PCollectionViews.InMemoryMapViewFn
-  
- inMemoryMapViewUsingVoidKey(PCollection<KV<Void, KV<K, V>>>, Coder<K>, Coder<V>, WindowingStrategy<?, W>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
- 
Returns a  PCollectionView<Map<K, V>> capable of processing elements windowed using the
 provided  WindowingStrategy. 
- InMemoryMetaStore - Class in org.apache.beam.sdk.extensions.sql.meta.store
- 
A  MetaStore which stores the meta info in memory. 
- InMemoryMetaStore() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.store.InMemoryMetaStore
-  
- InMemoryMetaTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider
- 
A InMemoryMetaTableProvideris an abstractTableProviderfor in-memory types.
 
- InMemoryMetaTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.InMemoryMetaTableProvider
-  
- InMemoryMultimapFromVoidKeyViewFn(Coder<K>, Coder<V>) - Constructor for class org.apache.beam.sdk.values.PCollectionViews.InMemoryMultimapFromVoidKeyViewFn
-  
- inMemoryMultimapView(PCollection<KV<K, V>>, Coder<K>, Coder<V>, WindowingStrategy<?, W>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
- 
Returns a  PCollectionView<Map<K, Iterable<V>>> capable of processing elements windowed
 using the provided  WindowingStrategy. 
- InMemoryMultimapViewFn(Coder<K>, Coder<V>) - Constructor for class org.apache.beam.sdk.values.PCollectionViews.InMemoryMultimapViewFn
-  
- inMemoryMultimapViewUsingVoidKey(PCollection<KV<Void, KV<K, V>>>, Coder<K>, Coder<V>, WindowingStrategy<?, W>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
- 
Returns a  PCollectionView<Map<K, Iterable<V>>> capable of processing elements windowed
 using the provided  WindowingStrategy. 
- inNamespace(String) - Static method in class org.apache.beam.sdk.metrics.MetricNameFilter
-  
- inNamespace(Class<?>) - Static method in class org.apache.beam.sdk.metrics.MetricNameFilter
-  
- Inner() - Constructor for class org.apache.beam.sdk.schemas.transforms.Filter.Inner
-  
- innerBroadcastJoin(PCollection<RhsT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Join
- 
Perform an inner join, broadcasting the right side. 
- innerJoin(PCollection<KV<K, V1>>, PCollection<KV<K, V2>>) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
- 
Inner join of two collections of KV elements. 
- innerJoin(String, PCollection<KV<K, V1>>, PCollection<KV<K, V2>>) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
- 
Inner join of two collections of KV elements. 
- innerJoin(PCollection<RhsT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Join
- 
Perform an inner join. 
- inOnlyPane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
- 
Creates a new  PAssert.IterableAssert like this one, but with the assertion restricted to only
 run on the provided window. 
- inOnlyPane(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-  
- inOnlyPane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
- 
Creates a new  PAssert.SingletonAssert like this one, but with the assertion restricted to
 only run on the provided window. 
- inOnTimePane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
- 
Creates a new  PAssert.IterableAssert like this one, but with the assertion restricted to only
 run on the provided window. 
- inOnTimePane(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-  
- inOnTimePane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
- 
Creates a new  PAssert.SingletonAssert like this one, but with the assertion restricted to
 only run on the provided window, running the checker only on the on-time pane for each key. 
- inOrder(Trigger...) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterEach
- 
Returns an AfterEachTriggerwith the given subtriggers.
 
- inOrder(List<Trigger>) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterEach
- 
Returns an AfterEachTriggerwith the given subtriggers.
 
- InProcessServerFactory - Class in org.apache.beam.sdk.fn.server
- 
A  ServerFactory which creates  servers with the  InProcessServerBuilder. 
- INPUT - Static variable in class org.apache.beam.sdk.managed.ManagedTransformConstants
-  
- INPUT_ROWS_TAG - Static variable in class org.apache.beam.sdk.schemas.transforms.providers.JavaExplodeTransformProvider
-  
- INPUT_ROWS_TAG - Static variable in class org.apache.beam.sdk.schemas.transforms.providers.JavaFilterTransformProvider
-  
- INPUT_ROWS_TAG - Static variable in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider
-  
- INPUT_ROWS_TAG - Static variable in class org.apache.beam.sdk.schemas.transforms.providers.LoggingTransformProvider
-  
- INPUT_TAG - Static variable in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.expansion.service.WindowIntoTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.extensions.sql.expansion.SqlTransformSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.csv.providers.CsvWriteTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformProvider
- 
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryExportReadSchemaTransformProvider
- 
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryFileLoadsSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.json.providers.JsonWriteTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadProvider
- 
- inputCollectionNames() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteProvider
- 
- inputCollectionNames() - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.schemas.transforms.providers.FlattenTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaExplodeTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaFilterTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider
-  
- inputCollectionNames() - Method in class org.apache.beam.sdk.schemas.transforms.providers.LoggingTransformProvider
-  
- inputCollectionNames() - Method in interface org.apache.beam.sdk.schemas.transforms.SchemaTransformProvider
- 
Returns the input collection names of this transform. 
- inputFormatProvider - Variable in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
- 
- inputOf(ProcessFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
- 
Returns a type descriptor for the input of the given  ProcessFunction, subject to Java
 type erasure: may contain unresolved type variables if the type was erased. 
- inputOf(SerializableFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
- 
- inputOf(Contextful.Fn<InputT, OutputT>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
- 
- INPUTS - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- insert(PartitionMetadata) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao
- 
Inserts the partition metadata. 
- insert(PartitionMetadata) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao.InTransactionContext
- 
Inserts the partition metadata. 
- INSERT_OR_UPDATE_URN - Static variable in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar
-  
- INSERT_URN - Static variable in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar
-  
- insertAll(TableReference, List<FailsafeValueInSingleWindow<TableRow, TableRow>>, List<String>, InsertRetryPolicy, List<ValueInSingleWindow<T>>, ErrorContainer<T>, boolean, boolean, boolean, List<ValueInSingleWindow<TableRow>>) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
- 
Inserts  TableRows with the specified insertIds if not null. 
- insertAll(TableReference, List<FailsafeValueInSingleWindow<TableRow, TableRow>>, List<String>, InsertRetryPolicy, List<ValueInSingleWindow<T>>, ErrorContainer<T>, boolean, boolean, boolean, List<ValueInSingleWindow<TableRow>>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.DatasetServiceImpl
-  
- insertAll(TableReference, List<TableRow>, List<String>) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- insertAll(TableReference, List<FailsafeValueInSingleWindow<TableRow, TableRow>>, List<String>, InsertRetryPolicy, List<ValueInSingleWindow<T>>, ErrorContainer<T>, boolean, boolean, boolean, List<ValueInSingleWindow<TableRow>>) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- InsertBuilder() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.InsertBuilder
-  
- insertDataToTable(String, String, String, List<Map<String, Object>>) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
- 
Inserts rows to a table using a BigQuery streaming write. 
- insertDeduplicate() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
-  
- insertDistributedSync() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
-  
- InsertOrUpdateBuilder() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.InsertOrUpdateBuilder
-  
- insertQuorum() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
-  
- InsertRetryPolicy - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
A retry policy for streaming BigQuery inserts. 
- InsertRetryPolicy() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy
-  
- InsertRetryPolicy.Context - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
Contains information about a failed insert. 
- insertRows(Schema, Row...) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery
-  
- INSTANCE - Static variable in interface org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderProvider.Factory
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory.GenericDatumFactory
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.JdbcDriver
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamJavaTypeFactory
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelDataTypeSystem
-  
- instance() - Static method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelMetadataQuery
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamAggregateProjectMergeRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamAggregationRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamBasicAggregationRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamCalcMergeRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamCalcRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamCoGBKJoinRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamEnumerableConverterRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIntersectRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIOPushDownRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIOSinkRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamJoinAssociateRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamMatchRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamMinusRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSideInputJoinRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSideInputLookupJoinRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSortRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamTableFunctionScanRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamUncollectRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamUnionRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamUnnestRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamValuesRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamWindowRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamJavaUdfCalcRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcMergeRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcSplittingRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.unnest.BeamZetaSqlUncollectRule
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.unnest.BeamZetaSqlUnnestRule
-  
- INSTANCE - Static variable in exception org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver.CloseException
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow.Coder
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.schemas.AutoValueSchema.AbstractGetterTypeSupplier
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.schemas.JavaBeanSchema.GetterTypeSupplier
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.schemas.JavaFieldSchema.JavaFieldTypeSupplier
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
-  
- INSTANCE - Static variable in class org.apache.beam.sdk.transforms.windowing.GlobalWindow
- 
- INSTANCE - Static variable in class org.apache.beam.sdk.transforms.windowing.PaneInfo.PaneInfoCoder
-  
- instanceId - Variable in class org.apache.beam.sdk.io.gcp.healthcare.WebPathParser.DicomWebPath
-  
- InstantCoder - Class in org.apache.beam.sdk.coders
- 
A  Coder for joda  Instant that encodes it as a big endian  Long shifted
 such that lexicographic ordering of the bytes corresponds to chronological order. 
- InstantDeserializer - Class in org.apache.beam.sdk.io.kafka.serialization
- 
- InstantDeserializer() - Constructor for class org.apache.beam.sdk.io.kafka.serialization.InstantDeserializer
-  
- instantiateCoder(String, RunnerApi.Components) - Static method in class org.apache.beam.runners.fnexecution.translation.PipelineTranslatorUtils
- 
Creates a coder for a given PCollection id from the Proto definition. 
- instantiateDestination(String) - Method in interface org.apache.beam.sdk.io.iceberg.DynamicDestinations
-  
- instantiateHealthcareClient() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.HL7v2Read.FetchHL7v2Message.HL7v2MessageGetFn
- 
Instantiate healthcare client (version v1). 
- instantiateHealthcareClient() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.Read.FetchHL7v2Message.HL7v2MessageGetFn
- 
Instantiate healthcare client (version v1). 
- instantiateRunnerWireCoder(PipelineNode.PCollectionNode, RunnerApi.Components) - Static method in class org.apache.beam.runners.fnexecution.wire.WireCoders
- 
Instantiates a runner-side wire coder for the given PCollection. 
- instantiateRunnerWireCoder(PipelineNode.PCollectionNode, RunnerApi.Components, RunnerApi.ExecutableStagePayload.WireCoderSetting) - Static method in class org.apache.beam.runners.fnexecution.wire.WireCoders
- 
Instantiates a runner-side wire coder for the given PCollection. 
- InstantSerializer - Class in org.apache.beam.sdk.io.kafka.serialization
- 
- InstantSerializer() - Constructor for class org.apache.beam.sdk.io.kafka.serialization.InstantSerializer
-  
- InstructionRequestHandler - Interface in org.apache.beam.runners.fnexecution.control
- 
Interface for any function that can handle a Fn API BeamFnApi.InstructionRequest.
 
- INT16 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- INT16 - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
The type of int16 fields. 
- INT32 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- INT32 - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
The type of int32 fields. 
- INT64 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- INT64 - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
The type of int64 fields. 
- INT8 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- IntBundle() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.IntBundle
-  
- INTEGER - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- integers() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
- 
- integersGlobally() - Static method in class org.apache.beam.sdk.transforms.Max
- 
Returns a PTransformthat takes an inputPCollection<Integer>and returns aPCollection<Integer>whose contents is the maximum of the inputPCollection's
 elements, orInteger.MIN_VALUEif there are no elements.
 
- integersGlobally() - Static method in class org.apache.beam.sdk.transforms.Min
- 
Returns a PTransformthat takes an inputPCollection<Integer>and returns aPCollection<Integer>whose contents is a single value that is the minimum of the inputPCollection's elements, orInteger.MAX_VALUEif there are no elements.
 
- integersGlobally() - Static method in class org.apache.beam.sdk.transforms.Sum
- 
Returns a PTransformthat takes an inputPCollection<Integer>and returns aPCollection<Integer>whose contents is the sum of the inputPCollection's
 elements, or0if there are no elements.
 
- integersPerKey() - Static method in class org.apache.beam.sdk.transforms.Max
- 
Returns a PTransformthat takes an inputPCollection<KV<K, Integer>>and
 returns aPCollection<KV<K, Integer>>that contains an output element mapping each
 distinct key in the inputPCollectionto the maximum of the values associated with that
 key in the inputPCollection.
 
- integersPerKey() - Static method in class org.apache.beam.sdk.transforms.Min
- 
Returns a PTransformthat takes an inputPCollection<KV<K, Integer>>and
 returns aPCollection<KV<K, Integer>>that contains an output element mapping each
 distinct key in the inputPCollectionto the minimum of the values associated with that
 key in the inputPCollection.
 
- integersPerKey() - Static method in class org.apache.beam.sdk.transforms.Sum
- 
Returns a PTransformthat takes an inputPCollection<KV<K, Integer>>and
 returns aPCollection<KV<K, Integer>>that contains an output element mapping each
 distinct key in the inputPCollectionto the sum of the values associated with that key
 in the inputPCollection.
 
- interceptor() - Static method in class org.apache.beam.sdk.fn.server.GrpcContextHeaderAccessorProvider
-  
- interceptResponse(HttpResponse) - Method in class org.apache.beam.sdk.extensions.gcp.util.UploadIdResponseInterceptor
-  
- Internal - Annotation Type in org.apache.beam.sdk.annotations
- 
Signifies that a publicly accessible API (public class, method or field) is intended for internal
 use only and not for public consumption. 
- InterpolateData() - Constructor for class org.apache.beam.sdk.extensions.timeseries.FillGaps.InterpolateData
-  
- interpolateKey(double) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
- 
Returns a  ByteKeykey such that  [startKey, key) represents
 approximately the specified fraction of the range  [startKey, endKey). 
- intersectAll(PCollection<T>) - Static method in class org.apache.beam.sdk.transforms.Sets
- 
Returns a new PTransformtransform that follows SET ALL semantics to compute the
 intersection with providedPCollection<T>.
 
- intersectAll() - Static method in class org.apache.beam.sdk.transforms.Sets
- 
Returns a new PTransformtransform that follows SET ALL semantics which takes aPCollectionList<PCollection<T>>and returns aPCollection<T>containing the
 intersection all of collections done in order for all collections inPCollectionList<T>.
 
- intersectDistinct(PCollection<T>) - Static method in class org.apache.beam.sdk.transforms.Sets
- 
Returns a new PTransformtransform that follows SET DISTINCT semantics to compute the
 intersection with providedPCollection<T>.
 
- intersectDistinct() - Static method in class org.apache.beam.sdk.transforms.Sets
- 
Returns a PTransformthat takes aPCollectionList<PCollection<T>>and returns aPCollection<T>containing the intersection of collections done in order for all
 collections inPCollectionList<T>.
 
- intersects(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
- 
Returns whether this window intersects the given window. 
- IntervalWindow - Class in org.apache.beam.sdk.transforms.windowing
- 
- IntervalWindow(Instant, Instant) - Constructor for class org.apache.beam.sdk.transforms.windowing.IntervalWindow
- 
Creates a new IntervalWindow that represents the half-open time interval [start, end). 
- IntervalWindow(Instant, ReadableDuration) - Constructor for class org.apache.beam.sdk.transforms.windowing.IntervalWindow
-  
- IntervalWindow.IntervalWindowCoder - Class in org.apache.beam.sdk.transforms.windowing
- 
- IntervalWindowCoder() - Constructor for class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
-  
- into(TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.transforms.FlatMapElements
- 
- into(TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.transforms.MapElements
- 
- into(TypeDescriptor<K2>) - Static method in class org.apache.beam.sdk.transforms.MapKeys
- 
- into(TypeDescriptor<V2>) - Static method in class org.apache.beam.sdk.transforms.MapValues
- 
- into(WindowFn<? super T, ?>) - Static method in class org.apache.beam.sdk.transforms.windowing.Window
- 
Creates a  WindowPTransform that uses the given  WindowFn to window the
 data. 
- InTransactionContext(String, TransactionContext, Dialect) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao.InTransactionContext
- 
Constructs a context to execute a user defined function transactionally. 
- InvalidConfigurationException - Exception in org.apache.beam.sdk.schemas.io
- 
Exception thrown when the configuration for a  SchemaIO is invalid. 
- InvalidConfigurationException(String) - Constructor for exception org.apache.beam.sdk.schemas.io.InvalidConfigurationException
-  
- InvalidConfigurationException(String, Throwable) - Constructor for exception org.apache.beam.sdk.schemas.io.InvalidConfigurationException
-  
- InvalidConfigurationException(Throwable) - Constructor for exception org.apache.beam.sdk.schemas.io.InvalidConfigurationException
-  
- InvalidLocationException - Exception in org.apache.beam.sdk.schemas.io
- 
Exception thrown when the configuration for a  SchemaIO is invalid. 
- InvalidLocationException(String) - Constructor for exception org.apache.beam.sdk.schemas.io.InvalidLocationException
-  
- InvalidLocationException(String, Throwable) - Constructor for exception org.apache.beam.sdk.schemas.io.InvalidLocationException
-  
- InvalidLocationException(Throwable) - Constructor for exception org.apache.beam.sdk.schemas.io.InvalidLocationException
-  
- InvalidSchemaException - Exception in org.apache.beam.sdk.schemas.io
- 
Exception thrown when the schema for a  SchemaIO is invalid. 
- InvalidSchemaException(String) - Constructor for exception org.apache.beam.sdk.schemas.io.InvalidSchemaException
-  
- InvalidSchemaException(String, Throwable) - Constructor for exception org.apache.beam.sdk.schemas.io.InvalidSchemaException
-  
- InvalidSchemaException(Throwable) - Constructor for exception org.apache.beam.sdk.schemas.io.InvalidSchemaException
-  
- InvalidTableException - Exception in org.apache.beam.sdk.extensions.sql.meta.provider
- 
Exception thrown when the request for a table is invalid, such as invalid metadata. 
- InvalidTableException(String) - Constructor for exception org.apache.beam.sdk.extensions.sql.meta.provider.InvalidTableException
-  
- InvalidTableException(String, Throwable) - Constructor for exception org.apache.beam.sdk.extensions.sql.meta.provider.InvalidTableException
-  
- InvalidTableException(Throwable) - Constructor for exception org.apache.beam.sdk.extensions.sql.meta.provider.InvalidTableException
-  
- invokeAdvance(ReaderT) - Method in class org.apache.beam.runners.flink.metrics.ReaderInvocationUtil
-  
- invokeStart(ReaderT) - Method in class org.apache.beam.runners.flink.metrics.ReaderInvocationUtil
-  
- invokeWithExecutor(RunnerApi.Pipeline, Struct, String, ListeningExecutorService) - Method in class org.apache.beam.runners.flink.FlinkJobInvoker
-  
- invokeWithExecutor(RunnerApi.Pipeline, Struct, String, ListeningExecutorService) - Method in class org.apache.beam.runners.jobsubmission.JobInvoker
- 
Start running a job, abstracting its state as a  JobInvocation instance. 
- invokeWithExecutor(RunnerApi.Pipeline, Struct, String, ListeningExecutorService) - Method in class org.apache.beam.runners.spark.SparkJobInvoker
-  
- inWindow(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
- 
Creates a new  PAssert.IterableAssert like this one, but with the assertion restricted to only
 run on the provided window. 
- inWindow(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-  
- inWindow(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
- 
Creates a new  PAssert.SingletonAssert like this one, but with the assertion restricted to
 only run on the provided window. 
- ioException() - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.StorageObjectOrIOException
- 
Returns the IOException.
 
- ir() - Method in class org.apache.beam.sdk.extensions.sbe.SerializableIr
- 
Returns the underlying  Ir. 
- IrOptions() - Constructor for class org.apache.beam.sdk.extensions.sbe.SbeSchema.IrOptions
-  
- IS_MERGING_WINDOW_FN - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- IS_PAIR_LIKE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- IS_STREAM_LIKE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- IS_WRAPPER - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- isAbsolute() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- isAccessible() - Method in interface org.apache.beam.sdk.options.ValueProvider
- 
- isAccessible() - Method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
-  
- isAccessible() - Method in class org.apache.beam.sdk.options.ValueProvider.RuntimeValueProvider
-  
- isAccessible() - Method in class org.apache.beam.sdk.options.ValueProvider.StaticValueProvider
-  
- isAliveOrThrow() - Method in class org.apache.beam.runners.fnexecution.environment.ProcessManager.RunningProcess
- 
Checks if the underlying process is still running. 
- isAllowDuplicates() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- isAllowedLatenessSpecified() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- isAlreadyMerged() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- isAppProfileSingleClusterAndTransactional(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
- 
Verify the app profile is for single cluster routing with allow single-row transactions
 enabled. 
- isArray() - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
Returns true if this type is known to be an array type. 
- isAtSplitPoint() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
- 
Returns true if the reader is at a split point. 
- isAtSplitPoint() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
- 
Returns true only for the first record; compressed sources cannot be split. 
- isAtSplitPoint() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
- 
- isAutoBalanceWriteFilesShardingEnabled() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- isBlockOnRun() - Method in interface org.apache.beam.runners.direct.DirectOptions
-  
- isBlockOnRun() - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-  
- isBounded() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSchemaIOProvider
-  
- isBounded() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
-  
- isBounded() - Method in interface org.apache.beam.sdk.extensions.sql.impl.rel.BeamRelNode
- 
Whether the collection of rows represented by this relational expression is bounded (known to
 be finite) or unbounded (may or may not be finite). 
- isBounded() - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamPCollectionTable
-  
- isBounded() - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTable
- 
Whether this table is bounded (known to be finite) or unbounded (may or may not be finite). 
- isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigtable.BigtableTable
-  
- isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
-  
- isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
-  
- isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable
-  
- isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestBoundedTable
-  
- isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestUnboundedTable
-  
- isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTable
-  
- isBounded() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySchemaIOProvider
- 
Indicates whether the PCollections produced by this transform will contain a bounded or
 unbounded number of elements. 
- isBounded() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.ReadChangeStreamPartitionProgressTracker
- 
This restriction tracker is for unbounded streams. 
- isBounded() - Method in class org.apache.beam.sdk.io.gcp.datastore.DataStoreV1SchemaIOProvider
-  
- isBounded() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubSchemaIOProvider
-  
- isBounded() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRangeTracker
-  
- isBounded() - Method in class org.apache.beam.sdk.io.jdbc.JdbcSchemaIOProvider
-  
- isBounded() - Method in interface org.apache.beam.sdk.schemas.io.SchemaIOProvider
-  
- isBounded() - Method in class org.apache.beam.sdk.transforms.PeriodicSequence.OutputRangeTracker
-  
- isBounded() - Method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
-  
- isBounded() - Method in class org.apache.beam.sdk.transforms.splittabledofn.GrowableOffsetRangeTracker
-  
- isBounded() - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
-  
- isBounded() - Method in class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker
- 
Return the boundedness of the current restriction. 
- isBounded() - Method in class org.apache.beam.sdk.values.PCollection
-  
- isBoundedCollection(Collection<PCollection<?>>) - Method in class org.apache.beam.runners.spark.SparkRunner.Evaluator
-  
- isCacheDisabled() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- isCleanArtifactsPerJob() - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver.ServerConfiguration
-  
- isClosed() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- isClosed() - Method in interface org.apache.beam.sdk.io.solace.broker.MessageProducer
- 
Returns true if the message producer is closed, false otherwise. 
- isClosed() - Method in class org.apache.beam.sdk.io.solace.broker.SolaceMessageProducer
-  
- isClosed() - Method in class org.apache.beam.sdk.transforms.errorhandling.ErrorHandler.DefaultErrorHandler
-  
- isClosed() - Method in interface org.apache.beam.sdk.transforms.errorhandling.ErrorHandler
-  
- isClosed() - Method in class org.apache.beam.sdk.transforms.errorhandling.ErrorHandler.PTransformErrorHandler
-  
- isCollectionType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
-  
- isCommitOffsetsInFinalizeEnabled() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.extensions.sql.impl.TVFSlidingWindowFn
-  
- isCompatible(Trigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
-  
- isCompatible(Trigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
-  
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-  
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-  
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-  
- isCompatible(Trigger) - Method in class org.apache.beam.sdk.transforms.windowing.DefaultTrigger
-  
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-  
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-  
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-  
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-  
- isCompatible(Trigger) - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
- 
For internal use only; no backwards-compatibility guarantees. 
- isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
- 
- isCompositeType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
-  
- isCompound() - Method in class org.apache.beam.sdk.extensions.sql.impl.TableName
- 
Whether it's a compound table name (with multiple path components). 
- isCompressed(String) - Static method in enum org.apache.beam.sdk.io.CompressedSource.CompressionMode
- 
Deprecated. Returns whether the file's extension matches of one of the known compression formats. 
- isCompressed(String) - Method in enum org.apache.beam.sdk.io.Compression
-  
- isCompressionEnabled() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- isConsumingReceivedData() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
-  
- isCooperative() - Method in class org.apache.beam.runners.jet.processors.BoundedSourceP
-  
- isCooperative() - Method in class org.apache.beam.runners.jet.processors.UnboundedSourceP
-  
- isDateTimeType(Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
- 
Returns true if the type is any of the various date time types. 
- isDateType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
-  
- isDdl(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
-  
- isDecimal(Schema.TypeName) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
- 
Checks if type is decimal. 
- isDirectory() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
-  
- isDirectory() - Method in class org.apache.beam.sdk.io.ClassLoaderFileSystem.ClassLoaderResourceId
-  
- isDirectory() - Method in interface org.apache.beam.sdk.io.fs.ResourceId
- 
Returns  true if this  ResourceId represents a directory, false otherwise. 
- isDisjoint(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
- 
Returns whether this window is disjoint from the given window. 
- isDone() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
- 
- isDone() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-  
- isDone() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-  
- isDynamicRead() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- isEmpty() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.CachedSideInputReader
-  
- isEmpty() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.SparkSideInputReader
-  
- isEmpty() - Method in class org.apache.beam.runners.spark.util.CachedSideInputReader
-  
- isEmpty() - Method in class org.apache.beam.runners.spark.util.SparkSideInputReader
-  
- isEmpty(StateAccessor<K>) - Method in class org.apache.beam.runners.twister2.translators.functions.internal.SystemReduceFnBuffering
-  
- isEmpty() - Method in class org.apache.beam.runners.twister2.utils.Twister2SideInputReader
-  
- isEmpty() - Method in class org.apache.beam.sdk.fn.data.WeightedList
-  
- isEmpty() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.StreamProgress
-  
- isEmpty() - Method in class org.apache.beam.sdk.io.range.ByteKey
- 
Returns  true if the  byte[] backing this  ByteKey is of length 0. 
- isEmpty() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
-  
- isEmpty() - Method in interface org.apache.beam.sdk.state.GroupingState
- 
- isEmpty() - Method in interface org.apache.beam.sdk.state.MapState
- 
- isEmpty() - Method in interface org.apache.beam.sdk.state.MultimapState
- 
- isEmpty() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
-  
- isEmpty() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
-  
- isEmpty() - Method in class org.apache.beam.sdk.transforms.Requirements
- 
Whether this is an empty set of requirements. 
- isEnableStreamingEngine() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-  
- isEncodingPositionsOverridden() - Method in class org.apache.beam.sdk.schemas.Schema
- 
Returns whether encoding positions have been explicitly overridden. 
- isEnforceEncodability() - Method in interface org.apache.beam.runners.direct.DirectOptions
-  
- isEnforceImmutability() - Method in interface org.apache.beam.runners.direct.DirectOptions
-  
- isEOF() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
- 
- isEOF() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
-  
- isEOF() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
-  
- isEOF() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- isEOF() - Method in interface org.apache.beam.sdk.io.solace.broker.MessageReceiver
- 
Test clients may return true to signal that all expected messages have been pulled
 and the test may complete. 
- isEOS() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- isEqualTo(T) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
- 
Asserts that the value in question is equal to the provided value, according to Object.equals(java.lang.Object).
 
- isEqWithEpsilon(RelOptCost) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
-  
- isExternalizedCheckpointsEnabled() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
- 
Enables or disables externalized checkpoints. 
- isFailToLock() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.StreamProgress
-  
- isFirst() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
- 
Return true if this is the first pane produced for the associated window. 
- IsFlinkNativeTransform() - Constructor for class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.IsFlinkNativeTransform
-  
- IsFlinkNativeTransform() - Constructor for class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.IsFlinkNativeTransform
-  
- isForceStreaming() - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
-  
- isForceWatermarkSync() - Method in class org.apache.beam.runners.spark.io.CreateStream
-  
- isGetter(Method) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
-  
- isHeartbeat() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.StreamProgress
-  
- isHotKeyLoggingEnabled() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
- 
If enabled then the literal key will be logged to Cloud Logging if a hot key is detected. 
- isIn(Collection<T>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- isIn(Coder<T>, Collection<T>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- isIn(T[]) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- isIn(Coder<T>, T[]) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- isInboundEdgeOfVertex(Edge, String, String, String) - Method in interface org.apache.beam.runners.jet.DAGBuilder.WiringListener
-  
- isInboundEdgeOfVertex(Edge, String, String, String) - Method in class org.apache.beam.runners.jet.processors.FlattenP.Supplier
-  
- IsInf - Class in org.apache.beam.sdk.extensions.sql.impl.udf
- 
IS_INF(X) 
- IsInf() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.udf.IsInf
-  
- isInf(Double) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.IsInf
-  
- isInf(Float) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.IsInf
-  
- isInfinite() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
-  
- isInitialEvent(long, EventT) - Method in interface org.apache.beam.sdk.extensions.ordered.EventExaminer
- 
Is this event the first expected event for the given key and window if the per key sequence is
 used? In case of global sequence it determines the first global sequence event. 
- isInitialPartition(String) - Static method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.InitialPartition
- 
Verifies if the given partition token is the initial partition. 
- isInputSortRelAndLimitOnly() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.AbstractBeamCalcRel
-  
- isInsertDefault() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
- 
Returns whether or not this transformation applies a default value. 
- isIntegral(Schema.TypeName) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
- 
Checks if type is integral. 
- isJoinLegal(Join) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
- 
This method checks if a join is legal and can be converted into Beam SQL. 
- isKey(ImmutableBitSet) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
-  
- isLast() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
- 
Return true if this is the last pane that will be produced in the associated window. 
- isLastEvent(long, EventT) - Method in interface org.apache.beam.sdk.extensions.ordered.EventExaminer
- 
Is this event the last expected event for a given key and window? 
- isLastEventReceived() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus
-  
- isLastRecordInTransactionInPartition() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
Indicates whether this record is the last emitted for the given transaction in the given
 partition. 
- isLe(RelOptCost) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
-  
- isLeaf(PCollection<?>) - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator.TranslationState
-  
- isLeaf(PCollection<?>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- isLimitOnly() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSortRel
-  
- isLogicalType(String) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
-  
- isLogicalType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
-  
- isLt(RelOptCost) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
-  
- isMapType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
-  
- isMetadata() - Method in class org.apache.beam.sdk.values.PCollectionViews.ValueOrMetadata
-  
- isMetricsSupported() - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
- 
Indicates whether metrics reporting is supported. 
- isModeSpecified() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- isMutable() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- IsNan - Class in org.apache.beam.sdk.extensions.sql.impl.udf
- 
IS_NAN(X) 
- IsNan() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.udf.IsNan
-  
- isNan(Float) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.IsNan
-  
- isNan(Double) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.IsNan
-  
- isNonMerging() - Method in class org.apache.beam.sdk.transforms.windowing.NonMergingWindowFn
-  
- isNonMerging() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
- 
Returns true if this WindowFnnever needs to merge any windows.
 
- isNull(String) - Method in class org.apache.beam.sdk.extensions.sql.provider.UdfTestProvider.IsNullFn
-  
- isNullable() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.TypeWithNullability
-  
- isNullable() - Method in class org.apache.beam.sdk.io.snowflake.data.SnowflakeColumn
-  
- isNullable() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
- 
Returns whether the field is nullable. 
- IsNullFn() - Constructor for class org.apache.beam.sdk.extensions.sql.provider.UdfTestProvider.IsNullFn
-  
- isNumericType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
-  
- isOneOf(T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- isOneOf(Coder<T>, T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- isOpen() - Method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingReadableByteChannel
-  
- isOpen() - Method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingSeekableByteChannel
-  
- isOpen() - Method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingWritableByteChannel
-  
- isOutboundEdgeOfVertex(Edge, String, String, String) - Method in interface org.apache.beam.runners.jet.DAGBuilder.WiringListener
-  
- isOutboundEdgeOfVertex(Edge, String, String, String) - Method in class org.apache.beam.runners.jet.processors.FlattenP.Supplier
-  
- isPreviewEnabled() - Method in class org.apache.beam.sdk.io.cdap.context.BatchSinkContextImpl
-  
- isPreviewEnabled() - Method in class org.apache.beam.sdk.io.cdap.context.BatchSourceContextImpl
-  
- isPreviewEnabled() - Method in class org.apache.beam.sdk.io.cdap.context.StreamingSourceContextImpl
-  
- isPrimaryKey() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ColumnType
- 
True if the column is part of the primary key, false otherwise. 
- isPrimitiveType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
-  
- isProduceStatusUpdateOnEveryEvent() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler
- 
Indicates if the status update needs to be sent after each event's processing. 
- isQueueNonExclusive(String) - Method in class org.apache.beam.sdk.io.solace.broker.BasicAuthSempClient
-  
- isQueueNonExclusive(String) - Method in class org.apache.beam.sdk.io.solace.broker.SempBasicAuthClientExecutor
-  
- isQueueNonExclusive(String) - Method in interface org.apache.beam.sdk.io.solace.broker.SempClient
- 
Determines if the specified queue is non-exclusive. 
- isReadOnly() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- isReadSeekEfficient() - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata
-  
- isReady() - Method in class org.apache.beam.sdk.fn.stream.DataStreams.DataStreamDecoder
-  
- isReady() - Method in interface org.apache.beam.sdk.fn.stream.PrefetchableIterator
- 
Returns trueif and only ifIterator.hasNext()andIterator.next()will not require an
 expensive operation.
 
- isReceiverStopped() - Method in class org.apache.beam.sdk.io.sparkreceiver.WrappedSupervisor
-  
- isRedistributed() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- isRegisterByteSizeObserverCheap(RandomAccessData) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
-  
- isRegisterByteSizeObserverCheap(ByteString) - Method in class org.apache.beam.runners.fnexecution.wire.ByteStringCoder
- 
- isRegisterByteSizeObserverCheap(BigDecimal) - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
- 
- isRegisterByteSizeObserverCheap(Integer) - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
- 
- isRegisterByteSizeObserverCheap(Long) - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
- 
- isRegisterByteSizeObserverCheap(Short) - Method in class org.apache.beam.sdk.coders.BigEndianShortCoder
- 
- isRegisterByteSizeObserverCheap(BigInteger) - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
- 
- isRegisterByteSizeObserverCheap(Boolean) - Method in class org.apache.beam.sdk.coders.BooleanCoder
-  
- isRegisterByteSizeObserverCheap(byte[]) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
- 
- isRegisterByteSizeObserverCheap(Byte) - Method in class org.apache.beam.sdk.coders.ByteCoder
- 
- isRegisterByteSizeObserverCheap(T) - Method in class org.apache.beam.sdk.coders.Coder
- 
- isRegisterByteSizeObserverCheap(Double) - Method in class org.apache.beam.sdk.coders.DoubleCoder
- 
- isRegisterByteSizeObserverCheap(ReadableDuration) - Method in class org.apache.beam.sdk.coders.DurationCoder
- 
- isRegisterByteSizeObserverCheap(Float) - Method in class org.apache.beam.sdk.coders.FloatCoder
- 
- isRegisterByteSizeObserverCheap(Instant) - Method in class org.apache.beam.sdk.coders.InstantCoder
- 
- isRegisterByteSizeObserverCheap(IterableT) - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
- 
- isRegisterByteSizeObserverCheap(KV<K, V>) - Method in class org.apache.beam.sdk.coders.KvCoder
- 
Returns whether both keyCoder and valueCoder are considered not expensive. 
- isRegisterByteSizeObserverCheap(T) - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
- 
LengthPrefixCoderis cheap ifvalueCoderis cheap.
 
- isRegisterByteSizeObserverCheap(T) - Method in class org.apache.beam.sdk.coders.NullableCoder
- 
NullableCoderis cheap ifvalueCoderis cheap.
 
- isRegisterByteSizeObserverCheap(T) - Method in class org.apache.beam.sdk.coders.TimestampPrefixingWindowCoder
-  
- isRegisterByteSizeObserverCheap(Integer) - Method in class org.apache.beam.sdk.coders.VarIntCoder
- 
- isRegisterByteSizeObserverCheap(Long) - Method in class org.apache.beam.sdk.coders.VarLongCoder
- 
- isRegisterByteSizeObserverCheap(Void) - Method in class org.apache.beam.sdk.coders.VoidCoder
- 
- isRegisterByteSizeObserverCheap(ByteString) - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
- 
- isRegisterByteSizeObserverCheap(HyperLogLogPlus) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.HyperLogLogPlusCoder
-  
- isRegisterByteSizeObserverCheap(EncodedBoundedWindow) - Method in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow.Coder
-  
- isRegisterByteSizeObserverCheap(KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
-  
- isRegisterByteSizeObserverCheap(ProducerRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
-  
- isRegisterByteSizeObserverCheap(OffsetRange) - Method in class org.apache.beam.sdk.io.range.OffsetRange.Coder
-  
- isRegisterByteSizeObserverCheap(RawUnionValue) - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
- 
Since this coder uses elementCoders.get(index) and coders that are known to run in constant
 time, we defer the return value to that coder. 
- isRegisterByteSizeObserverCheap(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
-  
- isResume() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.InitialPipelineState
-  
- isRowLocked(Row) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.encoder.MetadataTableEncoder
- 
Returns true if row is locked. 
- isRunnerDeterminedSharding() - Method in interface org.apache.beam.runners.direct.DirectTestOptions
-  
- isSdfTimer(String) - Static method in class org.apache.beam.runners.fnexecution.control.BundleCheckpointHandlers.StateAndTimerBundleCheckpointHandler
- 
A helper function to help check whether the given timer is the timer which is set for
 rescheduling BeamFnApi.DelayedBundleApplication.
 
- isSetter(Method) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
-  
- isShouldReportDiagnosticMetrics() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
- 
Whether additional diagnostic metrics should be reported for a Transform. 
- isSideInputLookupJoin() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
-  
- isSimple() - Method in class org.apache.beam.sdk.extensions.sql.impl.TableName
- 
Whether it's a simple name, with a single name component. 
- isSplittable() - Method in class org.apache.beam.sdk.io.CompressedSource
- 
Determines whether a single file represented by this source is splittable. 
- isSplittable() - Method in class org.apache.beam.sdk.io.FileBasedSource
- 
Determines whether a file represented by this source is can be split into bundles. 
- isStart() - Method in class org.apache.beam.sdk.io.redis.RedisCursor
-  
- isStarted() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
- 
- isStarted() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-  
- isStreaming() - Method in interface org.apache.beam.sdk.options.StreamingOptions
- 
Set to true if running a streaming pipeline. 
- isStreamingEngine() - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
-  
- isStringType(Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- isStringType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
-  
- isSubtypeOf(Schema.TypeName) - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
-  
- isSubtypeOf(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
Return true if this type is a subtype of the given type. 
- isSuccess() - Method in class org.apache.beam.sdk.io.tika.ParseResult
- 
Returns whether this file was parsed successfully. 
- isSuccess() - Method in class org.apache.beam.sdk.testing.SuccessOrFailure
-  
- isSupertypeOf(Schema.TypeName) - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
- 
Whether this is a super type of the another type. 
- isSupertypeOf(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
Returns true if this type is assignable from the given type. 
- isSupported() - Method in enum org.apache.beam.sdk.extensions.sql.meta.ProjectSupport
-  
- isSystemTransaction() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
- 
Whether the given transaction is Spanner system transaction. 
- isTableEmpty(TableReference) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
- 
Returns true if the table is empty. 
- isTableEmpty(TableReference) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.DatasetServiceImpl
-  
- isTableEmpty(TableReference) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- isTableResolved(Table) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
- 
True if the table was resolved using the Calcite schema. 
- isTerminal() - Method in enum org.apache.beam.runners.local.ExecutionDriver.DriverState
-  
- isTerminal() - Method in enum org.apache.beam.sdk.PipelineResult.State
-  
- isTerminated(JobApi.JobState.Enum) - Static method in class org.apache.beam.runners.jobsubmission.JobInvocation
-  
- isTimer() - Method in class org.apache.beam.sdk.fn.data.LogicalEndpoint
-  
- isTimestampCombinerSpecified() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- isTopicExists(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
- 
Return true if topicexists.
 
- isTopicExists(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
-  
- isTopicExists(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
-  
- isTopicExists(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- isTriggerSpecified() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- isTrue() - Method in interface org.apache.beam.io.requestresponse.CallShouldBackoff
- 
Report whether to backoff. 
- isTrustSelfSignedCerts() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
-  
- isUnbounded() - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Gets value of a plugin type. 
- isUnknown() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
-  
- isUnknown() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
- 
If any of the values for rowCount, rate or window is infinite, it returns true. 
- isUnknown() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
- 
Return true if there is no timing information for the current  PaneInfo. 
- isUpdate() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
- 
Whether to update the currently running pipeline with the same name as this one. 
- isValid(int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- isValidPartition(Range.ByteStringRange) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ByteStringRangeHelper
- 
Checks if the partition's start key is before its end key. 
- isWholeStream - Variable in class org.apache.beam.sdk.coders.Coder.Context
- 
Deprecated. Whether the encoded or decoded value fills the remainder of the output or input (resp.)
 record/stream contents. 
- isWildcard(GcsPath) - Static method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
- 
Returns true if the given speccontains wildcard.
 
- isWrapperFor(Class<?>) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- isWrapping() - Method in class org.apache.beam.sdk.io.cassandra.RingRange
-  
- isZero() - Method in class org.apache.beam.runners.spark.metrics.MetricsContainerStepMapAccumulator
-  
- isZero() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsAccumulator
-  
- item(String, String) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
- 
Create a display item for the specified key and string value. 
- item(String, ValueProvider<?>) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
- 
- item(String, Integer) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
- 
Create a display item for the specified key and integer value. 
- item(String, Long) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
- 
Create a display item for the specified key and integer value. 
- item(String, Float) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
- 
Create a display item for the specified key and floating point value. 
- item(String, Double) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
- 
Create a display item for the specified key and floating point value. 
- item(String, Boolean) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
- 
Create a display item for the specified key and boolean value. 
- item(String, Instant) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
- 
Create a display item for the specified key and timestamp value. 
- item(String, Duration) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
- 
Create a display item for the specified key and duration value. 
- item(String, Class<T>) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
- 
Create a display item for the specified key and class value. 
- item(String, DisplayData.Type, T) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
- 
Create a display item for the specified key, type, and value. 
- Item() - Constructor for class org.apache.beam.sdk.transforms.display.DisplayData.Item
-  
- items() - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
-  
- items() - Method in class org.apache.beam.sdk.transforms.display.DisplayData
-  
- ItemSpec() - Constructor for class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
-  
- iterable(Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
-  
- iterable() - Static method in class org.apache.beam.sdk.transforms.Materializations
- 
For internal use only; no backwards-compatibility guarantees. 
- ITERABLE_MATERIALIZATION_URN - Static variable in class org.apache.beam.sdk.transforms.Materializations
- 
The URN for a  Materialization where the primitive view type is an iterable of fully
 specified windowed values. 
- IterableBackedListViewFn(PCollectionViews.TypeDescriptorSupplier<T>) - Constructor for class org.apache.beam.sdk.values.PCollectionViews.IterableBackedListViewFn
-  
- IterableCoder<T> - Class in org.apache.beam.sdk.coders
- 
- IterableCoder(Coder<T>) - Constructor for class org.apache.beam.sdk.coders.IterableCoder
-  
- IterableLikeCoder<T,IterableT extends java.lang.Iterable<T>> - Class in org.apache.beam.sdk.coders
- 
An abstract base class with functionality for assembling a  Coder for a class that
 implements  Iterable. 
- IterableLikeCoder(Coder<T>, String) - Constructor for class org.apache.beam.sdk.coders.IterableLikeCoder
-  
- iterables() - Static method in class org.apache.beam.sdk.transforms.Flatten
- 
Returns a PTransformthat takes aPCollection<Iterable<T>>and returns aPCollection<T>containing all the elements from all theIterables.
 
- iterables() - Static method in class org.apache.beam.sdk.transforms.ToString
- 
Transforms each item in the iterable of the input  PCollection to a  String using
 the  Object.toString() method followed by a "," until the last element in the iterable. 
- iterables(String) - Static method in class org.apache.beam.sdk.transforms.ToString
- 
Transforms each item in the iterable of the input  PCollection to a  String using
 the  Object.toString() method followed by the specified delimiter until the last element
 in the iterable. 
- iterables(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
- 
- iterableView(PCollection<T>, PCollectionViews.TypeDescriptorSupplier<T>, WindowingStrategy<?, W>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
- 
Returns a  PCollectionView<Iterable<T>> capable of processing elements windowed using
 the provided  WindowingStrategy. 
- IterableViewFn(PCollectionViews.TypeDescriptorSupplier<T>) - Constructor for class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn
- 
Deprecated. 
- IterableViewFn2(PCollectionViews.TypeDescriptorSupplier<T>) - Constructor for class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn2
-  
- iterableViewUsingVoidKey(TupleTag<Materializations.MultimapView<Void, T>>, PCollection<KV<Void, T>>, PCollectionViews.TypeDescriptorSupplier<T>, WindowingStrategy<?, W>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
- 
- iterableWithSize(int) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- iterableWithSize(SerializableMatcher<? super Integer>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- iterator() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- iterator() - Method in interface org.apache.beam.sdk.fn.stream.PrefetchableIterable
-  
- iterator() - Method in class org.apache.beam.sdk.fn.stream.PrefetchableIterables.Default
-  
- iterator() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient.HL7v2MessagePages
-  
- iterator() - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
-  
- iterator() - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices.FakeBigQueryServerStream
-  
- OBJECT_TYPE_NAME - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- ObjectPool<KeyT,ObjectT> - Class in org.apache.beam.sdk.io.aws2.common
- 
Reference counting object pool to easily share & destroy objects. 
- ObjectPool(Function<KeyT, ObjectT>) - Constructor for class org.apache.beam.sdk.io.aws2.common.ObjectPool
-  
- ObjectPool(Function<KeyT, ObjectT>, ThrowingConsumer<Exception, ObjectT>) - Constructor for class org.apache.beam.sdk.io.aws2.common.ObjectPool
-  
- ObjectPool.ClientPool<ClientT extends SdkClient> - Class in org.apache.beam.sdk.io.aws2.common
- 
Client pool to easily share AWS clients per configuration. 
- observe(RestrictionTracker<RestrictionT, PositionT>, RestrictionTrackers.ClaimObserver<PositionT>) - Static method in class org.apache.beam.sdk.fn.splittabledofn.RestrictionTrackers
- 
- observeTimestamp(Instant) - Method in interface org.apache.beam.sdk.transforms.splittabledofn.TimestampObservingWatermarkEstimator
- 
Update watermark estimate with latest output timestamp. 
- observeTimestamp(Instant) - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.MonotonicallyIncreasing
-  
- of(PTransform<PCollection<RequestT>, Result<KV<RequestT, ResponseT>>>, PTransform<PCollection<KV<RequestT, ResponseT>>, Result<KV<RequestT, ResponseT>>>) - Static method in class org.apache.beam.io.requestresponse.Cache.Pair
-  
- of(Caller<RequestT, ResponseT>, Coder<ResponseT>) - Static method in class org.apache.beam.io.requestresponse.RequestResponseIO
- 
- of(String, String, String) - Static method in class org.apache.beam.runners.dataflow.util.PackageUtil.StagedFile
-  
- of() - Static method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
-  
- of(ByteBuffer) - Static method in class org.apache.beam.runners.flink.adapter.FlinkKey
-  
- of(K, Coder<K>) - Static method in class org.apache.beam.runners.flink.adapter.FlinkKey
-  
- of(BeamFnApi.ProcessBundleDescriptor, List<RemoteInputDestination>, Map<String, Coder>, Map<String, Map<String, ProcessBundleDescriptors.SideInputSpec>>, Map<String, Map<String, ProcessBundleDescriptors.BagUserStateSpec>>, Map<String, Map<String, ProcessBundleDescriptors.TimerSpec>>) - Static method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
-  
- of(String, String, RunnerApi.FunctionSpec, Coder<T>, Coder<W>) - Static method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.SideInputSpec
-  
- of(Coder<T>, FnDataReceiver<T>) - Static method in class org.apache.beam.runners.fnexecution.control.RemoteOutputReceiver
-  
- of(Coder<T>, String) - Static method in class org.apache.beam.runners.fnexecution.data.RemoteInputDestination
-  
- of() - Static method in class org.apache.beam.runners.fnexecution.wire.ByteStringCoder
-  
- of(K, Coder<K>) - Static method in class org.apache.beam.runners.local.StructuralKey
- 
Create a new Structural Key of the provided key that can be encoded by the provided coder. 
- of(T, CloseableResource.Closer<T>) - Static method in class org.apache.beam.runners.portability.CloseableResource
- 
- of(JobApi.MetricResults) - Static method in class org.apache.beam.runners.portability.PortableMetrics
-  
- of(Coder<T>, Duration, boolean) - Static method in class org.apache.beam.runners.spark.io.CreateStream
- 
Creates a new Spark based stream intended for test purposes. 
- of(Coder<T>, Duration) - Static method in class org.apache.beam.runners.spark.io.CreateStream
- 
Creates a new Spark based stream without forced watermark sync, intended for test purposes. 
- of(Table<String, String, byte[]>, Collection<byte[]>) - Static method in class org.apache.beam.runners.spark.stateful.StateAndTimers
-  
- of(SideInputReader, Collection<PCollectionView<?>>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.CachedSideInputReader
- 
Creates a SideInputReader that caches results for costly  Materializations if present, otherwise the SideInputReader is returned as is. 
- of(SideInputReader) - Static method in class org.apache.beam.runners.spark.util.CachedSideInputReader
- 
Create a new cached SideInputReader.
 
- of() - Static method in class org.apache.beam.sdk.coders.BigDecimalCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.BigEndianLongCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.BigEndianShortCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.BigIntegerCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.BitSetCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.BooleanCoder
- 
- of() - Static method in class org.apache.beam.sdk.coders.ByteArrayCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.ByteCoder
-  
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.CollectionCoder
-  
- of(Coder<IntermediateT>, DelegateCoder.CodingFunction<T, IntermediateT>, DelegateCoder.CodingFunction<IntermediateT, T>) - Static method in class org.apache.beam.sdk.coders.DelegateCoder
-  
- of(Coder<IntermediateT>, DelegateCoder.CodingFunction<T, IntermediateT>, DelegateCoder.CodingFunction<IntermediateT, T>, TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.coders.DelegateCoder
-  
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.DequeCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.DoubleCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.DurationCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.FloatCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.InstantCoder
-  
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.IterableCoder
-  
- of(Coder<K>, Coder<V>) - Static method in class org.apache.beam.sdk.coders.KvCoder
-  
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.LengthPrefixCoder
-  
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.ListCoder
-  
- of(Coder<K>, Coder<V>) - Static method in class org.apache.beam.sdk.coders.MapCoder
- 
Produces a MapCoder with the given keyCoder and valueCoder. 
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.NullableCoder
-  
- of(Schema) - Static method in class org.apache.beam.sdk.coders.RowCoder
-  
- of(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.coders.SerializableCoder
- 
- of(Class<T>) - Static method in class org.apache.beam.sdk.coders.SerializableCoder
- 
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.SetCoder
- 
Produces a  SetCoder with the given  elementCoder. 
- of(Coder<KeyT>) - Static method in class org.apache.beam.sdk.coders.ShardedKeyCoder
-  
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.SnappyCoder
- 
- of(Coder<K>, Coder<V>) - Static method in class org.apache.beam.sdk.coders.SortedMapCoder
- 
Produces a MapCoder with the given keyCoder and valueCoder. 
- of(Class<T>) - Static method in class org.apache.beam.sdk.coders.StringDelegateCoder
-  
- of(Class<T>, TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.coders.StringDelegateCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.StringUtf8Coder
-  
- of() - Static method in class org.apache.beam.sdk.coders.TextualIntegerCoder
-  
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.TimestampPrefixingWindowCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.VarIntCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.VarLongCoder
-  
- of() - Static method in class org.apache.beam.sdk.coders.VoidCoder
-  
- of(Coder<T>, byte[], int) - Static method in class org.apache.beam.sdk.coders.ZstdCoder
- 
- of(Coder<T>, byte[]) - Static method in class org.apache.beam.sdk.coders.ZstdCoder
- 
- of(Coder<T>, int) - Static method in class org.apache.beam.sdk.coders.ZstdCoder
- 
- of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.ZstdCoder
- 
- of() - Static method in class org.apache.beam.sdk.expansion.service.ExpansionServiceSchemaTransformProvider
-  
- of(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
- of(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns an AvroCoderinstance for the provided element type.
 
- of(TypeDescriptor<T>, boolean) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns an AvroCoderinstance for the provided element type, respecting whether to use
 Avro's Reflect* or Specific* suite for encoding and decoding.
 
- of(Class<T>) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns an AvroCoderinstance for the provided element class.
 
- of(Class<T>, boolean) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns an AvroCoderinstance for the given class, respecting whether to use Avro's
 Reflect* or Specific* suite for encoding and decoding.
 
- of(Class<T>, Schema) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns an AvroCoderinstance for the provided element type using the provided Avro
 schema
 
- of(AvroDatumFactory<T>, Schema) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns an  AvroCoder instance for the provided  AvroDatumFactory using the
 provided Avro schema. 
- of(Class<T>, Schema, boolean) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns an AvroCoderinstance for the provided element type using the provided Avro
 schema, respecting whether to use Avro's Reflect* or Specific* suite for encoding and decoding.
 
- of(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroGenericCoder
-  
- of(Class<T>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory
- 
- of(Class<T>, boolean) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory
- 
Returns an  AvroDatumFactory instance for the provided element type respecting Avro's
 Reflect* or Specific* suite for encoding and decoding. 
- of(Class<T>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory.ReflectDatumFactory
-  
- of(Class<T>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory.SpecificDatumFactory
-  
- of(Class<? extends InputT>) - Static method in class org.apache.beam.sdk.extensions.jackson.AsJsons
- 
- of(Class<? extends OutputT>) - Static method in class org.apache.beam.sdk.extensions.jackson.ParseJsons
- 
- of() - Static method in class org.apache.beam.sdk.extensions.ordered.combiner.SequenceRangeAccumulator.SequenceRangeAccumulatorCoder
-  
- of(long, long, Instant) - Static method in class org.apache.beam.sdk.extensions.ordered.ContiguousSequenceRange
-  
- of() - Static method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
-  
- of(Descriptors.Descriptor) - Static method in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
- 
Returns a  DynamicProtoCoder for the Protocol Buffers  DynamicMessage for the
 given  Descriptors.Descriptor. 
- of(ProtoDomain, Descriptors.Descriptor) - Static method in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
- 
Returns a  DynamicProtoCoder for the Protocol Buffers  DynamicMessage for the
 given  Descriptors.Descriptor. 
- of(ProtoDomain, String) - Static method in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
- 
- of(Class<T>) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
- 
Returns a  ProtoCoder for the given Protocol Buffers  Message. 
- of(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
- 
- of(String) - Static method in class org.apache.beam.sdk.extensions.python.transforms.DataframeTransform
- 
Instantiates a multi-language wrapper for a Python DataframeTransform with a given lambda
 function. 
- of(String, Schema.FieldType, Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.python.transforms.RunInference
- 
Instantiates a multi-language wrapper for a Python RunInference with a given model loader. 
- of(String, Schema) - Static method in class org.apache.beam.sdk.extensions.python.transforms.RunInference
- 
Instantiates a multi-language wrapper for a Python RunInference with a given model loader. 
- of() - Static method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.HyperLogLogPlusCoder
-  
- of(BeamSqlTable) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteTable
-  
- of(RexCall) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPCall
-  
- of(RexPatternFieldRef) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPFieldRef
-  
- of(RexLiteral) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- of(Byte) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- of(Short) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- of(Integer) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- of(Long) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- of(BigDecimal) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- of(Float) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- of(Double) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- of(ReadableDateTime) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- of(Boolean) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- of(String) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPLiteral
-  
- of(RexNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPOperation
-  
- of(SqlOperator) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPOperator
-  
- of(Schema, String, RexCall, Quantifier) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPPattern
-  
- of(RelFieldCollation) - Static method in class org.apache.beam.sdk.extensions.sql.impl.cep.OrderKey
-  
- of(Duration, Duration) - Static method in class org.apache.beam.sdk.extensions.sql.impl.TVFSlidingWindowFn
-  
- of(Object...) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestBoundedTable
- 
Convenient way to build a mocked bounded table. 
- of(Schema) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestBoundedTable
- 
Build a mocked bounded table with the specified type. 
- of(Object...) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestUnboundedTable
- 
Convenient way to build a mocked unbounded table. 
- of(FrameworkConfig, ExpressionConverter, RelOptCluster, QueryTrait) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ConversionContext
-  
- of(Duration, String...) - Static method in class org.apache.beam.sdk.extensions.timeseries.FillGaps
- 
Construct the transform for the given duration and key fields. 
- of(Duration, FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.extensions.timeseries.FillGaps
- 
Construct the transform for the given duration and key fields. 
- of() - Static method in class org.apache.beam.sdk.io.aws2.dynamodb.AttributeValueCoder
-  
- of(BigInteger, BigInteger) - Static method in class org.apache.beam.sdk.io.cassandra.RingRange
-  
- of(String, TableSchema.ColumnType) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.Column
-  
- of(String, TableSchema.ColumnType, TableSchema.DefaultType, Object) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.Column
-  
- of(TableSchema.TypeName) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- of(TableSchema.Column...) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema
-  
- of() - Static method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.ParamsCoder
-  
- of() - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.DocumentCoder
-  
- of(Coder<BoundedWindow>, Coder<DestinationT>) - Static method in class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
-  
- of() - Static method in class org.apache.beam.sdk.io.fs.MetadataCoder
- 
- of() - Static method in class org.apache.beam.sdk.io.fs.MetadataCoderV2
- 
- of() - Static method in class org.apache.beam.sdk.io.fs.ResourceIdCoder
- 
- of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertErrorCoder
-  
- of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageApiInsertErrorCoder
-  
- of(TableRow, RowMutationInformation) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.RowMutation
-  
- of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.RowMutation.RowMutationCoder
-  
- of(RowMutationInformation.MutationType, long) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.RowMutationInformation
- 
Deprecated. 
- of(RowMutationInformation.MutationType, String) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.RowMutationInformation
- 
- of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoder
-  
- of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV2
-  
- of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV3
-  
- of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
-  
- of() - Static method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteResultCoder
-  
- of(String, String) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirBundleParameter
-  
- of(String) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirBundleParameter
-  
- of(FhirBundleParameter, String) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirBundleResponse
-  
- of(String, String, Map<String, T>) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirSearchParameter
- 
Creates a FhirSearchParameter of type T. 
- of(String, Map<String, T>) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirSearchParameter
- 
Creates a FhirSearchParameter of type T, without a key. 
- of(Coder<T>) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirSearchParameterCoder
-  
- of(Coder<T>) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HealthcareIOErrorCoder
-  
- of(PCollectionTuple) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.HL7v2Read.Result
-  
- of(PCollectionTuple) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.Read.Result
-  
- of() - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2MessageCoder
-  
- of(Class<HL7v2Message>) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2MessageCoder
-  
- of(String, String) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2ReadParameter
-  
- of(String, HL7v2Message) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2ReadResponse
- 
- of() - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2ReadResponseCoder
-  
- of(Class<HL7v2ReadResponse>) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2ReadResponseCoder
-  
- of() - Static method in class org.apache.beam.sdk.io.gcp.healthcare.JsonArrayCoder
-  
- of(PubsubMessage, long, long, String, String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.IncomingMessage
-  
- of(PubsubMessage, long, String, String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.OutgoingMessage
-  
- of(PubsubMessage, long, String, String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.OutgoingMessage
-  
- of() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessagePayloadOnlyCoder
-  
- of(TypeDescriptor<PubsubMessage>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesAndMessageIdAndOrderingKeyCoder
-  
- of() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesAndMessageIdAndOrderingKeyCoder
-  
- of(TypeDescriptor<PubsubMessage>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesAndMessageIdCoder
-  
- of() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesAndMessageIdCoder
-  
- of(TypeDescriptor<PubsubMessage>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
-  
- of() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
-  
- of() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithMessageIdCoder
-  
- of(TypeDescriptor<PubsubMessage>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithTopicCoder
-  
- of() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithTopicCoder
-  
- of(ByteString) - Static method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.Uuid
-  
- of(Timestamp, Timestamp) - Static method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRange
- 
Constructs a timestamp range. 
- of(Class<T>) - Static method in class org.apache.beam.sdk.io.hadoop.WritableCoder
- 
Returns a WritableCoderinstance for the provided element class.
 
- of(JdbcIO.DataSourceConfiguration) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceProviderFromDataSourceConfiguration
-  
- of(JdbcIO.DataSourceConfiguration) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO.PoolableDataSourceProvider
-  
- of(Schema) - Static method in class org.apache.beam.sdk.io.jdbc.SchemaUtil.BeamRowMapper
-  
- of(SerializableFunction<TopicPartition, Boolean>) - Static method in class org.apache.beam.sdk.io.kafka.CheckStopReadingFnWrapper
-  
- of(String, String) - Static method in class org.apache.beam.sdk.io.kafka.ConfluentSchemaRegistryDeserializerProvider
-  
- of(String, String, Integer) - Static method in class org.apache.beam.sdk.io.kafka.ConfluentSchemaRegistryDeserializerProvider
-  
- of(String, String, Integer, Map<String, ?>) - Static method in class org.apache.beam.sdk.io.kafka.ConfluentSchemaRegistryDeserializerProvider
-  
- of(String, int, String) - Static method in class org.apache.beam.sdk.io.kafka.ConfluentSchemaRegistryDeserializerProvider
-  
- of(String, int, String, Integer) - Static method in class org.apache.beam.sdk.io.kafka.ConfluentSchemaRegistryDeserializerProvider
-  
- of(String, int, String, Integer, Map<String, ?>) - Static method in class org.apache.beam.sdk.io.kafka.ConfluentSchemaRegistryDeserializerProvider
-  
- of(Coder<K>, Coder<V>) - Static method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
-  
- of(TopicPartition, Long, Instant, Long, Instant, List<String>) - Static method in class org.apache.beam.sdk.io.kafka.KafkaSourceDescriptor
-  
- of(Coder<K>, Coder<V>) - Static method in class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
-  
- of(Neo4jIO.DriverConfiguration) - Static method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverProviderFromDriverConfiguration
-  
- of() - Static method in class org.apache.beam.sdk.io.pulsar.PulsarMessageCoder
-  
- of(String, Long, Long, MessageId, String, String) - Static method in class org.apache.beam.sdk.io.pulsar.PulsarSourceDescriptor
-  
- of(int...) - Static method in class org.apache.beam.sdk.io.range.ByteKey
- 
Creates a new  ByteKey backed by a copy of the specified  int[]. 
- of(ByteKey, ByteKey) - Static method in class org.apache.beam.sdk.io.range.ByteKeyRange
- 
Creates a new  ByteKeyRange with the given start and end keys. 
- of(ByteKeyRange) - Static method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
- 
- of() - Static method in class org.apache.beam.sdk.io.range.OffsetRange.Coder
-  
- of(Coder<MatchResult.Metadata>) - Static method in class org.apache.beam.sdk.io.ReadableFileCoder
-  
- of() - Static method in class org.apache.beam.sdk.io.ReadableFileCoder
-  
- of(String, long, boolean) - Static method in class org.apache.beam.sdk.io.redis.RedisCursor
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeDate
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeDateTime
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeTime
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeTimestamp
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeTimestampLTZ
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeTimestampNTZ
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeTimestampTZ
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.geospatial.SnowflakeGeography
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.logical.SnowflakeBoolean
-  
- of(int, int) - Static method in class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeDecimal
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeDouble
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeFloat
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeInteger
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeNumber
-  
- of(int, int) - Static method in class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeNumber
-  
- of(int, int) - Static method in class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeNumeric
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeReal
-  
- of(String, SnowflakeDataType) - Static method in class org.apache.beam.sdk.io.snowflake.data.SnowflakeColumn
-  
- of(String, SnowflakeDataType, boolean) - Static method in class org.apache.beam.sdk.io.snowflake.data.SnowflakeColumn
-  
- of(SnowflakeColumn...) - Static method in class org.apache.beam.sdk.io.snowflake.data.SnowflakeTableSchema
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.structured.SnowflakeArray
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.structured.SnowflakeObject
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.structured.SnowflakeVariant
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeBinary
-  
- of(long) - Static method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeBinary
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeChar
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeString
-  
- of(long) - Static method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeString
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeText
-  
- of(long) - Static method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeText
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeVarBinary
-  
- of() - Static method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeVarchar
-  
- of(long) - Static method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeVarchar
-  
- of(SnowflakeIO.DataSourceConfiguration) - Static method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceProviderFromDataSourceConfiguration
-  
- of() - Static method in class org.apache.beam.sdk.io.splunk.SplunkEventCoder
-  
- of(Class<T>, TProtocolFactory) - Static method in class org.apache.beam.sdk.io.thrift.ThriftCoder
- 
Returns an  ThriftCoder instance for the provided  clazz and  protocolFactory. 
- of(Class<T>) - Static method in class org.apache.beam.sdk.io.xml.JAXBCoder
- 
Create a coder for a given type of JAXB annotated objects. 
- of(ValueProvider<X>, SerializableFunction<X, T>) - Static method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
- 
- of(T) - Static method in class org.apache.beam.sdk.options.ValueProvider.StaticValueProvider
- 
- of(FieldAccessDescriptor.FieldDescriptor.ListQualifier) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier
-  
- of(FieldAccessDescriptor.FieldDescriptor.MapQualifier) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier
-  
- of(SerializableFunction<Row, byte[]>, SerializableFunction<byte[], Row>) - Static method in interface org.apache.beam.sdk.schemas.io.payloads.PayloadSerializer
-  
- of(String, int) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
- 
Return an instance of FixedBytes with specified byte array length. 
- of(int) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
- 
Return an instance of FixedBytes with specified byte array length. 
- of(int, int) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.FixedPrecisionNumeric
- 
Create a FixedPrecisionNumeric instance with specified precision and scale. 
- of(int) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.FixedPrecisionNumeric
- 
Create a FixedPrecisionNumeric instance with specified scale and unspecified precision. 
- of(Row) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.FixedPrecisionNumeric
- 
Create a FixedPrecisionNumeric instance with specified argument row. 
- of(String, int) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.FixedString
- 
Return an instance of FixedString with specified string length. 
- of(int) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.FixedString
- 
Return an instance of FixedString with specified string length. 
- of(String, int) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.VariableBytes
- 
Return an instance of VariableBytes with specified max byte array length. 
- of(int) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.VariableBytes
- 
Return an instance of VariableBytes with specified max byte array length. 
- of(String, int) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.VariableString
- 
Return an instance of VariableString with specified max string length. 
- of(int) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.VariableString
- 
Return an instance of VariableString with specified max string length. 
- of(String, Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.Schema.Field
- 
Return's a field with the give name and type. 
- of(Schema.TypeName) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
- of(Schema.Field...) - Static method in class org.apache.beam.sdk.schemas.Schema
-  
- of(Schema, TypeDescriptor<T>, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Static method in class org.apache.beam.sdk.schemas.SchemaCoder
- 
- of(Schema) - Static method in class org.apache.beam.sdk.schemas.SchemaCoder
- 
- of() - Static method in class org.apache.beam.sdk.schemas.transforms.Cast.Narrowing
-  
- of(Schema, Cast.Validator) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
-  
- of() - Static method in class org.apache.beam.sdk.schemas.transforms.Cast.Widening
-  
- of(int) - Static method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields.Fanout
-  
- of(SerializableFunction<Row, Integer>) - Static method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields.Fanout
-  
- of(FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.transforms.WithKeys
-  
- of(Coder<T>) - Static method in class org.apache.beam.sdk.testing.TestStream.TestStreamCoder
-  
- of(SerializableBiFunction<V, V, V>) - Static method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
- 
Returns a CombineFnthat uses the givenSerializableBiFunctionto combine
 values.
 
- of(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
- 
Returns a CombineFnthat uses the givenSerializableFunctionto combine
 values.
 
- of(SerializableFunction<Iterable<V>, V>, int) - Static method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
- 
Returns a CombineFnthat uses the givenSerializableFunctionto combine
 values, attempting to buffer at leastbufferSizevalues between invocations.
 
- of(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine.SimpleCombineFn
- 
Deprecated. Returns a CombineFnthat uses the givenSerializableFunctionto combine
 values.
 
- of(ClosureT, Requirements) - Static method in class org.apache.beam.sdk.transforms.Contextful
- 
Constructs a pair of the given closure and its requirements. 
- of(Iterable<T>) - Static method in class org.apache.beam.sdk.transforms.Create
- 
Returns a new  Create.Values transform that produces a  PCollection containing
 elements of the provided  Iterable. 
- of(T, T...) - Static method in class org.apache.beam.sdk.transforms.Create
- 
Returns a new  Create.Values transform that produces a  PCollection containing
 the specified elements. 
- of(Map<K, V>) - Static method in class org.apache.beam.sdk.transforms.Create
- 
Returns a new  Create.Values transform that produces a  PCollection of  KVs corresponding to the keys and values of the specified  Map. 
- of(DisplayData.Path, Class<?>, String) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
-  
- of(DoFn<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- of(CoGbkResultSchema, UnionCoder) - Static method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
- 
- of(TupleTag<V>, List<V>) - Static method in class org.apache.beam.sdk.transforms.join.CoGbkResult
- 
Returns a new CoGbkResult that contains just the given tag and given data. 
- of(List<TupleTag<?>>) - Static method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
-  
- of(TupleTag<InputT>, PCollection<KV<K, InputT>>) - Static method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
- 
Returns a new KeyedPCollectionTuple<K>with the given tag and initial PCollection.
 
- of(String, PCollection<KV<K, InputT>>) - Static method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
- 
- of(List<Coder<?>>) - Static method in class org.apache.beam.sdk.transforms.join.UnionCoder
- 
Builds a union coder with the given list of element coders. 
- of(T, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Max
- 
- of(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Max
- 
- of() - Static method in class org.apache.beam.sdk.transforms.Mean
- 
A Combine.CombineFnthat computes the arithmetic mean (a.k.a.
 
- of(T, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Min
- 
- of(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Min
- 
- of(DoFn<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.ParDo
- 
- of(int, Partition.PartitionWithSideInputsFn<? super T>, Requirements) - Static method in class org.apache.beam.sdk.transforms.Partition
- 
Returns a new PartitionPTransformthat divides its inputPCollectioninto the given number of partitions, using the given partitioning function.
 
- of(int, Partition.PartitionFn<? super T>) - Static method in class org.apache.beam.sdk.transforms.Partition
- 
Returns a new PartitionPTransformthat divides its inputPCollectioninto the given number of partitions, using the given partitioning function.
 
- of() - Static method in class org.apache.beam.sdk.transforms.Reshuffle
-  
- of(ByteKeyRange) - Static method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
-  
- of(RestrictionT) - Static method in class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker.TruncateResult
- 
- of(RestrictionT, RestrictionT) - Static method in class org.apache.beam.sdk.transforms.splittabledofn.SplitResult
- 
Returns a  SplitResult for the specified primary and residual restrictions. 
- of(PTransform<PCollection<T>, ?>) - Static method in class org.apache.beam.sdk.transforms.Tee
- 
Returns a new Tee PTransform that will apply an auxilary transform to the input as well as pass
 it on. 
- of(Consumer<PCollection<T>>) - Static method in class org.apache.beam.sdk.transforms.Tee
- 
Returns a new Tee PTransform that will apply an auxilary transform to the input as well as pass
 it on. 
- of() - Static method in class org.apache.beam.sdk.transforms.ToJson
-  
- of(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Top
- 
Returns a PTransformthat takes an inputPCollection<T>and returns aPCollection<List<T>>with a single element containing the largestcountelements of
 the inputPCollection<T>, in decreasing order, sorted using the givenComparator<T>.
 
- of(PCollectionView<ViewT>) - Static method in class org.apache.beam.sdk.transforms.View.CreatePCollectionView
-  
- of(Trigger.OnceTrigger...) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterAll
- 
Returns an AfterAllTriggerwith the given subtriggers.
 
- of(List<Trigger>) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterAll
- 
Returns an AfterAllTriggerwith the given subtriggers.
 
- of(Trigger.OnceTrigger...) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterFirst
- 
Returns an AfterFirstTriggerwith the given subtriggers.
 
- of(List<Trigger>) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterFirst
- 
Returns an AfterFirstTriggerwith the given subtriggers.
 
- of() - Static method in class org.apache.beam.sdk.transforms.windowing.DefaultTrigger
- 
Returns the default trigger. 
- of(Duration) - Static method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
- 
Partitions the timestamp space into half-open intervals of the form [N * size, (N + 1) * size),
 where 0 is the epoch. 
- of() - Static method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
-  
- of() - Static method in class org.apache.beam.sdk.transforms.windowing.PaneInfo.PaneInfoCoder
-  
- of(Duration) - Static method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
- 
Assigns timestamps into half-open intervals of the form [N * period, N * period + size), where
 0 is the epoch. 
- of(T, Exception) - Static method in class org.apache.beam.sdk.transforms.WithFailures.ExceptionElement
-  
- of(OutputT, PCollection<FailureElementT>) - Static method in class org.apache.beam.sdk.transforms.WithFailures.Result
-  
- of(PCollection<OutputElementT>, PCollection<FailureElementT>) - Static method in class org.apache.beam.sdk.transforms.WithFailures.Result
-  
- of(PCollectionTuple, TupleTag<OutputElementT>, TupleTag<FailureElementT>) - Static method in class org.apache.beam.sdk.transforms.WithFailures.Result
-  
- of(SerializableFunction<V, K>) - Static method in class org.apache.beam.sdk.transforms.WithKeys
- 
Returns a PTransformthat takes aPCollection<V>and returns aPCollection<KV<K, V>>, where each of the values in the inputPCollectionhas been
 paired with a key computed from the value by invoking the givenSerializableFunction.
 
- of(K) - Static method in class org.apache.beam.sdk.transforms.WithKeys
- 
Returns a PTransformthat takes aPCollection<V>and returns aPCollection<KV<K, V>>, where each of the values in the inputPCollectionhas been
 paired with the given key.
 
- of(SerializableFunction<T, Instant>) - Static method in class org.apache.beam.sdk.transforms.WithTimestamps
- 
- of(Coder<T>, Coder<ErrorT>, Coder<? extends BoundedWindow>) - Static method in class org.apache.beam.sdk.values.FailsafeValueInSingleWindow.Coder
-  
- of(T, Instant, BoundedWindow, PaneInfo, ErrorT) - Static method in class org.apache.beam.sdk.values.FailsafeValueInSingleWindow
-  
- of(K, V) - Static method in class org.apache.beam.sdk.values.KV
- 
Returns a  KV with the given key and value. 
- of(PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionList
- 
- of(Iterable<PCollection<T>>) - Static method in class org.apache.beam.sdk.values.PCollectionList
- 
- of(String, PCollection<Row>) - Static method in class org.apache.beam.sdk.values.PCollectionRowTuple
- 
- of(String, PCollection<Row>, String, PCollection<Row>) - Static method in class org.apache.beam.sdk.values.PCollectionRowTuple
- 
- of(String, PCollection<Row>, String, PCollection<Row>, String, PCollection<Row>) - Static method in class org.apache.beam.sdk.values.PCollectionRowTuple
- 
- of(String, PCollection<Row>, String, PCollection<Row>, String, PCollection<Row>, String, PCollection<Row>) - Static method in class org.apache.beam.sdk.values.PCollectionRowTuple
- 
- of(String, PCollection<Row>, String, PCollection<Row>, String, PCollection<Row>, String, PCollection<Row>, String, PCollection<Row>) - Static method in class org.apache.beam.sdk.values.PCollectionRowTuple
- 
- of(TupleTag<T>, PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
- 
- of(String, PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
- 
- of(String, PCollection<T>, String, PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
- 
- of(String, PCollection<T>, String, PCollection<T>, String, PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
- 
- of(String, PCollection<T>, String, PCollection<T>, String, PCollection<T>, String, PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
- 
- of(String, PCollection<T>, String, PCollection<T>, String, PCollection<T>, String, PCollection<T>, String, PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
- 
- of(K, int) - Static method in class org.apache.beam.sdk.values.ShardedKey
-  
- of(TupleTag<?>, PCollection<?>) - Static method in class org.apache.beam.sdk.values.TaggedPValue
-  
- of(V, Instant) - Static method in class org.apache.beam.sdk.values.TimestampedValue
- 
Returns a new TimestampedValuewith the given value and timestamp.
 
- of(Coder<T>) - Static method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
-  
- of(TupleTag<?>) - Static method in class org.apache.beam.sdk.values.TupleTagList
- 
- of(List<TupleTag<?>>) - Static method in class org.apache.beam.sdk.values.TupleTagList
- 
- of(Class<T>) - Static method in class org.apache.beam.sdk.values.TypeDescriptor
- 
- of(Type) - Static method in class org.apache.beam.sdk.values.TypeDescriptor
- 
- of(Coder<T>, Coder<? extends BoundedWindow>) - Static method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
-  
- of(T, Instant, BoundedWindow, PaneInfo) - Static method in class org.apache.beam.sdk.values.ValueInSingleWindow
-  
- of(Coder<ValueT>) - Static method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
-  
- of(WindowFn<T, W>) - Static method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- ofByteSize(long) - Static method in class org.apache.beam.sdk.transforms.GroupIntoBatches
- 
Aim to create batches each with the specified byte size. 
- ofByteSize(long, SerializableFunction<InputT, Long>) - Static method in class org.apache.beam.sdk.transforms.GroupIntoBatches
- 
Aim to create batches each with the specified byte size. 
- ofCallerAndSetupTeardown(CallerSetupTeardownT, Coder<ResponseT>) - Static method in class org.apache.beam.io.requestresponse.RequestResponseIO
- 
- ofDoubles() - Static method in class org.apache.beam.sdk.transforms.Max
- 
- ofDoubles() - Static method in class org.apache.beam.sdk.transforms.Min
- 
- ofDoubles() - Static method in class org.apache.beam.sdk.transforms.Sum
- 
- ofExpandedValue(PCollection<?>) - Static method in class org.apache.beam.sdk.values.TaggedPValue
-  
- offer(ArtifactRetrievalService, ArtifactStagingServiceGrpc.ArtifactStagingServiceStub, String) - Static method in class org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService
- 
Lazily stages artifacts by letting an ArtifactStagingService resolve and request artifacts. 
- offerCoders(Coder[]) - Method in interface org.apache.beam.sdk.state.StateSpec
- 
For internal use only; no backwards-compatibility guarantees. 
- offeringClientsToPool(ControlClientPool.Sink, HeaderAccessor) - Static method in class org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService
- 
- ofFirstElement() - Static method in class org.apache.beam.sdk.transforms.windowing.AfterSynchronizedProcessingTime
-  
- offset(Duration) - Method in interface org.apache.beam.sdk.state.Timer
- 
- OFFSET_INFINITY - Static variable in class org.apache.beam.sdk.io.range.OffsetRangeTracker
- 
Offset corresponding to infinity. 
- offsetBasedDeduplicationSupported() - Method in class org.apache.beam.sdk.io.UnboundedSource
- 
If offsetBasedDeduplicationSupported returns true, then the UnboundedSource needs to provide
 the following:
 
   UnboundedReader which provides offsets that are unique for each element and
       lexicographically ordered. 
- OffsetBasedReader(OffsetBasedSource<T>) - Constructor for class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-  
- OffsetBasedSource<T> - Class in org.apache.beam.sdk.io
- 
A  BoundedSource that uses offsets to define starting and ending positions. 
- OffsetBasedSource(long, long, long) - Constructor for class org.apache.beam.sdk.io.OffsetBasedSource
-  
- OffsetBasedSource.OffsetBasedReader<T> - Class in org.apache.beam.sdk.io
- 
- OffsetByteRangeCoder - Class in org.apache.beam.sdk.io.gcp.pubsublite.internal
-  
- OffsetByteRangeCoder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.internal.OffsetByteRangeCoder
-  
- OffsetRange - Class in org.apache.beam.sdk.io.range
- 
A restriction represented by a range of integers [from, to). 
- OffsetRange(long, long) - Constructor for class org.apache.beam.sdk.io.range.OffsetRange
-  
- OffsetRange.Coder - Class in org.apache.beam.sdk.io.range
- 
- OffsetRangeTracker - Class in org.apache.beam.sdk.io.range
- 
- OffsetRangeTracker(long, long) - Constructor for class org.apache.beam.sdk.io.range.OffsetRangeTracker
- 
Creates an OffsetRangeTrackerfor the specified range.
 
- OffsetRangeTracker - Class in org.apache.beam.sdk.transforms.splittabledofn
- 
- OffsetRangeTracker(OffsetRange) - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
-  
- ofIntegers() - Static method in class org.apache.beam.sdk.transforms.Max
- 
- ofIntegers() - Static method in class org.apache.beam.sdk.transforms.Min
- 
- ofIntegers() - Static method in class org.apache.beam.sdk.transforms.Sum
- 
- ofKVs(String, Schema.FieldType, Schema.FieldType, Coder<KeyT>) - Static method in class org.apache.beam.sdk.extensions.python.transforms.RunInference
- 
Similar to  RunInference#of(String, FieldType, FieldType) but the input is a  PCollection of  KVs. 
- ofKVs(String, Schema, Coder<KeyT>) - Static method in class org.apache.beam.sdk.extensions.python.transforms.RunInference
- 
- ofLongs() - Static method in class org.apache.beam.sdk.transforms.Max
- 
- ofLongs() - Static method in class org.apache.beam.sdk.transforms.Min
- 
- ofLongs() - Static method in class org.apache.beam.sdk.transforms.Sum
- 
- ofNamed(Map<String, ?>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters
-  
- ofNone() - Static method in class org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters
-  
- ofPatientEverything(HealthcareApiClient, String, Map<String, Object>) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient.FhirResourcePagesIterator
- 
Instantiates a new GetPatientEverything FHIR resource pages iterator. 
- ofPositional(List) - Static method in class org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters
-  
- ofPrimitiveOutputsInternal(Pipeline, TupleTagList, Map<TupleTag<?>, Coder<?>>, WindowingStrategy<?, ?>, PCollection.IsBounded) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
- 
For internal use only; no backwards-compatibility guarantees. 
- ofProvider(ValueProvider<T>, Coder<T>) - Static method in class org.apache.beam.sdk.transforms.Create
- 
- ofSearch(HealthcareApiClient, String, String, Map<String, Object>) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient.FhirResourcePagesIterator
- 
Instantiates a new search FHIR resource pages iterator. 
- ofSize(long) - Static method in class org.apache.beam.sdk.transforms.GroupIntoBatches
- 
Aim to create batches each with the specified element count. 
- OK - Static variable in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySinkMetrics
-  
- on(Join.FieldsEqual.Impl) - Method in class org.apache.beam.sdk.schemas.transforms.Join.Impl
- 
Join the PCollections using the provided predicate. 
- on(PCollection<?>...) - Static method in class org.apache.beam.sdk.transforms.Wait
- 
Waits on the given signal collections. 
- on(List<PCollection<?>>) - Static method in class org.apache.beam.sdk.transforms.Wait
- 
Waits on the given signal collections. 
- ON_TIME_AND_ONLY_FIRING - Static variable in class org.apache.beam.sdk.transforms.windowing.PaneInfo
- 
PaneInfoto use when there will be exactly one firing and it is on time.
 
- onAdvance(int, int) - Method in class org.apache.beam.sdk.fn.stream.AdvancingPhaser
-  
- onBatchCompleted(JavaStreamingListenerBatchCompleted) - Method in class org.apache.beam.runners.spark.metrics.MetricsAccumulator.AccumulatorCheckpointingSparkListener
-  
- onBatchCompleted(JavaStreamingListenerBatchCompleted) - Method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.WatermarkAdvancingStreamingListener
-  
- onBeforeRequest(String, String, Message) - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsV17.RateLimitPolicy
- 
Called before a request is sent. 
- onBeforeRequest(String, String, Message) - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17.SimpleRateLimitPolicy
-  
- onBundleSuccess() - Method in interface org.apache.beam.sdk.transforms.DoFn.BundleFinalizer.Callback
-  
- OnceTrigger(List<Trigger>) - Constructor for class org.apache.beam.sdk.transforms.windowing.Trigger.OnceTrigger
-  
- onCheckpoint(BeamFnApi.ProcessBundleResponse) - Method in interface org.apache.beam.runners.fnexecution.control.BundleCheckpointHandler
-  
- onCheckpoint(BeamFnApi.ProcessBundleResponse) - Method in class org.apache.beam.runners.fnexecution.control.BundleCheckpointHandlers.StateAndTimerBundleCheckpointHandler
-  
- onClaimed(PositionT) - Method in interface org.apache.beam.sdk.fn.splittabledofn.RestrictionTrackers.ClaimObserver
- 
- onClaimFailed(PositionT) - Method in interface org.apache.beam.sdk.fn.splittabledofn.RestrictionTrackers.ClaimObserver
- 
- onClose(Consumer<FnApiControlClient>) - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClient
-  
- onCompleted(BeamFnApi.ProcessBundleResponse) - Method in interface org.apache.beam.runners.fnexecution.control.BundleProgressHandler
- 
Handles the bundle's completion report. 
- onCompleted() - Method in class org.apache.beam.sdk.fn.stream.BufferingStreamObserver
-  
- onCompleted() - Method in class org.apache.beam.sdk.fn.stream.DirectStreamObserver
-  
- onCompleted() - Method in class org.apache.beam.sdk.fn.stream.ForwardingClientResponseObserver
-  
- onCompleted() - Method in class org.apache.beam.sdk.fn.stream.SynchronizedStreamObserver
-  
- oneOfEncoder(List<Encoder<T>>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers
- 
Creates a one-of Spark EncoderofStructTypewhere each alternative is
 represented as colum / field named by its index with a separateEncodereach.
 
- OneOfType - Class in org.apache.beam.sdk.schemas.logicaltypes
- 
A logical type representing a union of fields. 
- OneOfType.Value - Class in org.apache.beam.sdk.schemas.logicaltypes
- 
Represents a single OneOf value. 
- onError(Throwable) - Method in class org.apache.beam.sdk.fn.stream.BufferingStreamObserver
-  
- onError(Throwable) - Method in class org.apache.beam.sdk.fn.stream.DirectStreamObserver
-  
- onError(Throwable) - Method in class org.apache.beam.sdk.fn.stream.ForwardingClientResponseObserver
-  
- onError(Throwable) - Method in class org.apache.beam.sdk.fn.stream.SynchronizedStreamObserver
-  
- onError(String, String, Message, GoogleAdsError) - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsV17.RateLimitPolicy
- 
Called after a request fails with a retryable error. 
- onError(String, String, Message, GoogleAdsError) - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17.SimpleRateLimitPolicy
-  
- onGcTimer(Instant, ValueState<SortedMap<Instant, TimestampedValue<ValueT>>>, ValueState<SortedMap<Instant, Long>>) - Method in class org.apache.beam.sdk.extensions.timeseries.FillGaps.FillGapsDoFn
-  
- onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamAggregateProjectMergeRule
-  
- onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamAggregationRule
-  
- onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamBasicAggregationRule
-  
- onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamCalcMergeRule
-  
- onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamCalcSplittingRule
-  
- onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamCoGBKJoinRule
-  
- onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIOPushDownRule
-  
- onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamJoinAssociateRule
-  
- onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamJoinPushThroughJoinRule
-  
- onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSideInputJoinRule
-  
- onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamUnnestRule
-  
- onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcMergeRule
-  
- onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.unnest.BeamZetaSqlUnnestRule
-  
- onMerge(ReduceFn<K, T, Iterable<T>, W>.OnMergeContext) - Method in class org.apache.beam.runners.twister2.translators.functions.internal.SystemReduceFnBuffering
-  
- onNext(T) - Method in class org.apache.beam.sdk.fn.stream.BufferingStreamObserver
-  
- onNext(T) - Method in class org.apache.beam.sdk.fn.stream.DirectStreamObserver
-  
- onNext(ReqT) - Method in class org.apache.beam.sdk.fn.stream.ForwardingClientResponseObserver
-  
- onNext(V) - Method in class org.apache.beam.sdk.fn.stream.SynchronizedStreamObserver
-  
- onPollComplete(StateT) - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
- 
Called by the  Watch transform to compute a new termination state after every poll
 completion. 
- onProgress(BeamFnApi.ProcessBundleProgressResponse) - Method in interface org.apache.beam.runners.fnexecution.control.BundleProgressHandler
- 
Handles a progress report from the bundle while it is executing. 
- onReceiverStart() - Method in class org.apache.beam.sdk.io.sparkreceiver.WrappedSupervisor
-  
- onSeenNewOutput(Instant, StateT) - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
- 
- onStartup() - Method in interface org.apache.beam.sdk.harness.JvmInitializer
- 
Implement onStartup to run some custom initialization immediately after the JVM is launched for
 pipeline execution. 
- onSuccess(List<KinesisRecord>) - Method in interface org.apache.beam.sdk.io.aws2.kinesis.RateLimitPolicy
- 
Called after Kinesis records are successfully retrieved. 
- onSuccess(List<KinesisRecord>) - Method in class org.apache.beam.sdk.io.aws2.kinesis.RateLimitPolicyFactory.DefaultRateLimiter
-  
- onSuccess(List<KinesisRecord>) - Method in class org.apache.beam.sdk.io.aws2.kinesis.RateLimitPolicyFactory.DelayIntervalRateLimiter
-  
- onSuccess(String, String, Message) - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsV17.RateLimitPolicy
- 
Called after a request succeeds. 
- onSuccess(String, String, Message) - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17.SimpleRateLimitPolicy
-  
- onTeardown() - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiConvertMessages.ConvertMessagesDoFn
-  
- onTeardown() - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiFlushAndFinalizeDoFn
-  
- onTeardown() - Method in class org.apache.beam.sdk.io.gcp.bigquery.UpdateSchemaDestination
-  
- onThrottle(KinesisClientThrottledException) - Method in interface org.apache.beam.sdk.io.aws2.kinesis.RateLimitPolicy
- 
Called after the Kinesis client is throttled. 
- onThrottle(KinesisClientThrottledException) - Method in class org.apache.beam.sdk.io.aws2.kinesis.RateLimitPolicyFactory.DefaultRateLimiter
-  
- onTimer(String, String, KeyT, BoundedWindow, Instant, Instant, TimeDomain) - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
-  
- onTimer(String, Instant, TimerMap, TimerMap, ValueState<SortedMap<Instant, TimestampedValue<ValueT>>>, ValueState<SortedMap<Instant, Long>>, DoFn.OutputReceiver<ValueT>) - Method in class org.apache.beam.sdk.extensions.timeseries.FillGaps.FillGapsDoFn
-  
- OnTimerContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.OnTimerContext
-  
- onTrigger(ReduceFn<K, T, Iterable<T>, W>.OnTriggerContext) - Method in class org.apache.beam.runners.twister2.translators.functions.internal.SystemReduceFnBuffering
-  
- onWindowExpiration(BoundedWindow, Instant, KeyT) - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
-  
- OnWindowExpirationContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.OnWindowExpirationContext
-  
- open(MetricConfig) - Method in class org.apache.beam.runners.flink.metrics.FileReporter
-  
- open(WritableByteChannel) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Sink
-  
- open(GcsPath) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
- 
Opens an object in GCS. 
- open(ClassLoaderFileSystem.ClassLoaderResourceId) - Method in class org.apache.beam.sdk.io.ClassLoaderFileSystem
-  
- open(String) - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
- 
- open() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
- 
- open(WritableByteChannel) - Method in interface org.apache.beam.sdk.io.FileIO.Sink
- 
Initializes writing to the given channel. 
- open(ResourceIdT) - Method in class org.apache.beam.sdk.io.FileSystem
- 
Returns a read channel for the given ResourceIdT.
 
- open(ResourceId) - Static method in class org.apache.beam.sdk.io.FileSystems
- 
- open(WritableByteChannel) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Sink
-  
- open(WritableByteChannel) - Method in class org.apache.beam.sdk.io.TextIO.Sink
-  
- open(WritableByteChannel) - Method in class org.apache.beam.sdk.io.TFRecordIO.Sink
-  
- open(WritableByteChannel) - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.Sink
-  
- open(WritableByteChannel) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Sink
-  
- openSeekable() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
- 
- optimizedWrites() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
If true, enables new codepaths that are expected to use less resources while writing to
 BigQuery. 
- Options() - Constructor for class org.apache.beam.runners.dataflow.DataflowPipelineRegistrar.Options
-  
- Options() - Constructor for class org.apache.beam.runners.direct.DirectRegistrar.Options
-  
- Options() - Constructor for class org.apache.beam.runners.flink.FlinkRunnerRegistrar.Options
-  
- Options() - Constructor for class org.apache.beam.runners.jet.JetRunnerRegistrar.Options
-  
- options() - Method in class org.apache.beam.runners.jobsubmission.JobPreparation
-  
- Options() - Constructor for class org.apache.beam.runners.prism.PrismRegistrar.Options
-  
- Options() - Constructor for class org.apache.beam.runners.spark.SparkRunnerRegistrar.Options
-  
- Options() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingRunnerRegistrar.Options
-  
- Options() - Constructor for class org.apache.beam.runners.twister2.Twister2RunnerRegistrar.Options
-  
- options() - Static method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter
-  
- options - Variable in class org.apache.beam.sdk.extensions.sorter.ExternalSorter
-  
- Options() - Constructor for class org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options
-  
- Options() - Constructor for class org.apache.beam.sdk.options.ManualDockerEnvironmentOptions.Options
-  
- Options() - Constructor for class org.apache.beam.sdk.options.RemoteEnvironmentOptions.Options
-  
- Options() - Constructor for class org.apache.beam.sdk.transforms.resourcehints.ResourceHintsOptions.Options
-  
- OptionsRegistrar() - Constructor for class org.apache.beam.runners.portability.testing.TestUniversalRunner.OptionsRegistrar
-  
- ORACLE - Static variable in class org.apache.beam.sdk.io.jdbc.JdbcUtil
-  
- Order - Class in org.apache.beam.sdk.extensions.sql.example.model
- 
Describes an order. 
- Order(int, int) - Constructor for class org.apache.beam.sdk.extensions.sql.example.model.Order
-  
- Order() - Constructor for class org.apache.beam.sdk.extensions.sql.example.model.Order
-  
- OrderByKey() - Constructor for class org.apache.beam.sdk.values.KV.OrderByKey
-  
- OrderByValue() - Constructor for class org.apache.beam.sdk.values.KV.OrderByValue
-  
- OrderedEventProcessor<EventT,EventKeyT,ResultT,StateT extends MutableState<EventT,ResultT>> - Class in org.apache.beam.sdk.extensions.ordered
- 
Transform for processing ordered events. 
- OrderedEventProcessor() - Constructor for class org.apache.beam.sdk.extensions.ordered.OrderedEventProcessor
-  
- OrderedEventProcessorResult<KeyT,ResultT,EventT> - Class in org.apache.beam.sdk.extensions.ordered
- 
The result of the ordered processing. 
- orderedList(Coder<T>) - Static method in class org.apache.beam.sdk.state.StateSpecs
-  
- OrderedListState<T> - Interface in org.apache.beam.sdk.state
- 
A  ReadableState cell containing a list of values sorted by timestamp. 
- OrderedProcessingGlobalSequenceHandler(Class<EventT>, Class<KeyT>, Class<StateT>, Class<ResultT>) - Constructor for class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler.OrderedProcessingGlobalSequenceHandler
-  
- OrderedProcessingHandler<EventT,KeyT,StateT extends MutableState<EventT,?>,ResultT> - Class in org.apache.beam.sdk.extensions.ordered
- 
Parent class for Ordered Processing configuration handlers. 
- OrderedProcessingHandler(Class<EventT>, Class<KeyT>, Class<StateT>, Class<ResultT>) - Constructor for class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler
- 
Provide concrete classes which will be used by the ordered processing transform. 
- OrderedProcessingHandler.OrderedProcessingGlobalSequenceHandler<EventT,KeyT,StateT extends MutableState<EventT,?>,ResultT> - Class in org.apache.beam.sdk.extensions.ordered
- 
Parent class for Ordered Processing configuration handlers to handle processing of the events
 where global sequence is used. 
- OrderedProcessingStatus - Class in org.apache.beam.sdk.extensions.ordered
- 
Indicates the status of ordered processing for a particular key. 
- OrderedProcessingStatus() - Constructor for class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus
-  
- OrderedProcessingStatus.Builder - Class in org.apache.beam.sdk.extensions.ordered
-  
- OrderKey - Class in org.apache.beam.sdk.extensions.sql.impl.cep
- 
The OrderKeyclass stores the information to sort a column.
 
- orFinally(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
- 
Specify an ending condition for this trigger. 
- OrFinallyTrigger - Class in org.apache.beam.sdk.transforms.windowing
- 
A  Trigger that executes according to its main trigger until its "finally" trigger fires. 
- org.apache.beam.io.debezium - package org.apache.beam.io.debezium
- 
Transforms for reading from DebeziumIO. 
- org.apache.beam.io.requestresponse - package org.apache.beam.io.requestresponse
- 
Package provides Beam I/O transform support for safely reading from and writing to Web APIs. 
- org.apache.beam.runners.dataflow - package org.apache.beam.runners.dataflow
- 
Provides a Beam runner that executes pipelines on the Google Cloud Dataflow service. 
- org.apache.beam.runners.dataflow.options - package org.apache.beam.runners.dataflow.options
- 
- org.apache.beam.runners.dataflow.util - package org.apache.beam.runners.dataflow.util
- 
Provides miscellaneous internal utilities used by the Google Cloud Dataflow runner. 
- org.apache.beam.runners.direct - package org.apache.beam.runners.direct
- 
- org.apache.beam.runners.flink - package org.apache.beam.runners.flink
- 
Internal implementation of the Beam runner for Apache Flink. 
- org.apache.beam.runners.flink.adapter - package org.apache.beam.runners.flink.adapter
- 
Adaptors for using Beam transforms in Apache Flink pipelines. 
- org.apache.beam.runners.flink.metrics - package org.apache.beam.runners.flink.metrics
- 
Internal metrics implementation of the Beam runner for Apache Flink. 
- org.apache.beam.runners.fnexecution.artifact - package org.apache.beam.runners.fnexecution.artifact
- 
Pipeline execution-time artifact-management services, including abstract implementations of the
 Artifact Retrieval Service. 
- org.apache.beam.runners.fnexecution.control - package org.apache.beam.runners.fnexecution.control
- 
Utilities for a Beam runner to interact with the Fn API Control Servicevia java abstractions.
 
- org.apache.beam.runners.fnexecution.data - package org.apache.beam.runners.fnexecution.data
- 
Utilities for a Beam runner to interact with the Fn API Data Servicevia java abstractions.
 
- org.apache.beam.runners.fnexecution.environment - package org.apache.beam.runners.fnexecution.environment
- 
Classes used to instantiate and manage SDK harness environments. 
- org.apache.beam.runners.fnexecution.environment.testing - package org.apache.beam.runners.fnexecution.environment.testing
- 
Test utilities for the environment management package. 
- org.apache.beam.runners.fnexecution.logging - package org.apache.beam.runners.fnexecution.logging
- 
Classes used to log informational messages over the Beam Fn Logging Service.
 
- org.apache.beam.runners.fnexecution.provisioning - package org.apache.beam.runners.fnexecution.provisioning
- 
Provision api services. 
- org.apache.beam.runners.fnexecution.state - package org.apache.beam.runners.fnexecution.state
- 
State API services. 
- org.apache.beam.runners.fnexecution.status - package org.apache.beam.runners.fnexecution.status
- 
Worker Status API services. 
- org.apache.beam.runners.fnexecution.translation - package org.apache.beam.runners.fnexecution.translation
- 
Shared utilities for a Beam runner to translate portable pipelines. 
- org.apache.beam.runners.fnexecution.wire - package org.apache.beam.runners.fnexecution.wire
- 
Wire coders for communications between runner and SDK harness. 
- org.apache.beam.runners.jet - package org.apache.beam.runners.jet
- 
Implementation of the Beam runner for Hazelcast Jet. 
- org.apache.beam.runners.jet.metrics - package org.apache.beam.runners.jet.metrics
- 
Helper classes for implementing metrics in the Hazelcast Jet based runner. 
- org.apache.beam.runners.jet.processors - package org.apache.beam.runners.jet.processors
- 
Individual DAG node processors used by the Beam runner for Hazelcast Jet. 
- org.apache.beam.runners.jobsubmission - package org.apache.beam.runners.jobsubmission
- 
Job management services for use in beam runners. 
- org.apache.beam.runners.local - package org.apache.beam.runners.local
- 
Utilities useful when executing a pipeline on a single machine. 
- org.apache.beam.runners.portability - package org.apache.beam.runners.portability
- 
Support for executing a pipeline locally over the Beam fn API. 
- org.apache.beam.runners.portability.testing - package org.apache.beam.runners.portability.testing
- 
Testing utilities for the reference runner. 
- org.apache.beam.runners.prism - package org.apache.beam.runners.prism
- 
Support for executing a pipeline on Prism. 
- org.apache.beam.runners.spark - package org.apache.beam.runners.spark
- 
Internal implementation of the Beam runner for Apache Spark. 
- org.apache.beam.runners.spark.coders - package org.apache.beam.runners.spark.coders
- 
Beam coders and coder-related utilities for running on Apache Spark. 
- org.apache.beam.runners.spark.io - package org.apache.beam.runners.spark.io
- 
Spark-specific transforms for I/O. 
- org.apache.beam.runners.spark.metrics - package org.apache.beam.runners.spark.metrics
- 
Provides internal utilities for implementing Beam metrics using Spark accumulators. 
- org.apache.beam.runners.spark.metrics.sink - package org.apache.beam.runners.spark.metrics.sink
- 
Spark sinks that supports beam metrics and aggregators. 
- org.apache.beam.runners.spark.stateful - package org.apache.beam.runners.spark.stateful
- 
Spark-specific stateful operators. 
- org.apache.beam.runners.spark.structuredstreaming - package org.apache.beam.runners.spark.structuredstreaming
- 
Internal implementation of the Beam runner for Apache Spark. 
- org.apache.beam.runners.spark.structuredstreaming.examples - package org.apache.beam.runners.spark.structuredstreaming.examples
-  
- org.apache.beam.runners.spark.structuredstreaming.io - package org.apache.beam.runners.spark.structuredstreaming.io
- 
Spark-specific transforms for I/O. 
- org.apache.beam.runners.spark.structuredstreaming.metrics - package org.apache.beam.runners.spark.structuredstreaming.metrics
- 
Provides internal utilities for implementing Beam metrics using Spark accumulators. 
- org.apache.beam.runners.spark.structuredstreaming.metrics.sink - package org.apache.beam.runners.spark.structuredstreaming.metrics.sink
- 
Spark sinks that supports beam metrics and aggregators. 
- org.apache.beam.runners.spark.structuredstreaming.translation - package org.apache.beam.runners.spark.structuredstreaming.translation
- 
Internal translators for running Beam pipelines on Spark. 
- org.apache.beam.runners.spark.structuredstreaming.translation.batch - package org.apache.beam.runners.spark.structuredstreaming.translation.batch
- 
Internal utilities to translate Beam pipelines to Spark batching. 
- org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions - package org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions
- 
Internal implementation of the Beam runner for Apache Spark. 
- org.apache.beam.runners.spark.structuredstreaming.translation.helpers - package org.apache.beam.runners.spark.structuredstreaming.translation.helpers
- 
Internal helpers to translate Beam pipelines to Spark streaming. 
- org.apache.beam.runners.spark.structuredstreaming.translation.utils - package org.apache.beam.runners.spark.structuredstreaming.translation.utils
- 
Internal utils to translate Beam pipelines to Spark streaming. 
- org.apache.beam.runners.spark.util - package org.apache.beam.runners.spark.util
- 
Internal utilities to translate Beam pipelines to Spark. 
- org.apache.beam.runners.twister2 - package org.apache.beam.runners.twister2
- 
Internal implementation of the Beam runner for Twister2. 
- org.apache.beam.runners.twister2.translation.wrappers - package org.apache.beam.runners.twister2.translation.wrappers
- 
Internal implementation of the Beam runner for Twister2. 
- org.apache.beam.runners.twister2.translators - package org.apache.beam.runners.twister2.translators
- 
Internal implementation of the Beam runner for Twister2. 
- org.apache.beam.runners.twister2.translators.batch - package org.apache.beam.runners.twister2.translators.batch
- 
Internal implementation of the Beam runner for Twister2. 
- org.apache.beam.runners.twister2.translators.functions - package org.apache.beam.runners.twister2.translators.functions
- 
Internal implementation of the Beam runner for Twister2. 
- org.apache.beam.runners.twister2.translators.functions.internal - package org.apache.beam.runners.twister2.translators.functions.internal
- 
Internal implementation of the Beam runner for Twister2. 
- org.apache.beam.runners.twister2.translators.streaming - package org.apache.beam.runners.twister2.translators.streaming
- 
Internal implementation of the Beam runner for Twister2. 
- org.apache.beam.runners.twister2.utils - package org.apache.beam.runners.twister2.utils
- 
Internal implementation of the Beam runner for Twister2. 
- org.apache.beam.sdk - package org.apache.beam.sdk
- 
Provides a simple, powerful model for building both batch and streaming parallel data processing
  Pipelines. 
- org.apache.beam.sdk.annotations - package org.apache.beam.sdk.annotations
- 
Defines annotations used across the SDK. 
- org.apache.beam.sdk.coders - package org.apache.beam.sdk.coders
- 
Defines  Coders to specify how data is encoded to and
 decoded from byte strings. 
- org.apache.beam.sdk.expansion - package org.apache.beam.sdk.expansion
- 
Contains classes needed to expose transforms to other SDKs. 
- org.apache.beam.sdk.expansion.service - package org.apache.beam.sdk.expansion.service
- 
Classes used to expand cross-language transforms. 
- org.apache.beam.sdk.extensions.arrow - package org.apache.beam.sdk.extensions.arrow
- 
Extensions for using Apache Arrow with Beam. 
- org.apache.beam.sdk.extensions.avro - package org.apache.beam.sdk.extensions.avro
- 
Defines  Schema and other classes for representing schema'd
 data in a  Pipeline using Apache Avro. 
- org.apache.beam.sdk.extensions.avro.coders - package org.apache.beam.sdk.extensions.avro.coders
- 
Defines  Coders to specify how data is encoded to and
 decoded from byte strings using Apache Avro. 
- org.apache.beam.sdk.extensions.avro.io - package org.apache.beam.sdk.extensions.avro.io
- 
Defines transforms for reading and writing Avro storage format. 
- org.apache.beam.sdk.extensions.avro.schemas - package org.apache.beam.sdk.extensions.avro.schemas
- 
Defines  Schema and other classes for representing schema'd
 data in a  Pipeline using Apache Avro. 
- org.apache.beam.sdk.extensions.avro.schemas.io.payloads - package org.apache.beam.sdk.extensions.avro.schemas.io.payloads
- 
Provides abstractions for schema-aware AvroIO. 
- org.apache.beam.sdk.extensions.avro.schemas.utils - package org.apache.beam.sdk.extensions.avro.schemas.utils
- 
Defines utilities for deailing with schemas using Apache Avro. 
- org.apache.beam.sdk.extensions.gcp.auth - package org.apache.beam.sdk.extensions.gcp.auth
- 
Defines classes related to interacting with Credentialsfor pipeline
 creation and execution containing Google Cloud Platform components.
 
- org.apache.beam.sdk.extensions.gcp.options - package org.apache.beam.sdk.extensions.gcp.options
- 
Defines  PipelineOptions for configuring pipeline execution
 for Google Cloud Platform components. 
- org.apache.beam.sdk.extensions.gcp.storage - package org.apache.beam.sdk.extensions.gcp.storage
- 
Defines IO connectors for Google Cloud Storage. 
- org.apache.beam.sdk.extensions.gcp.util - package org.apache.beam.sdk.extensions.gcp.util
- 
Defines Google Cloud Platform component utilities that can be used by Beam runners. 
- org.apache.beam.sdk.extensions.gcp.util.channels - package org.apache.beam.sdk.extensions.gcp.util.channels
- 
Package contains java channel wrappers used with GCS. 
- org.apache.beam.sdk.extensions.gcp.util.gcsfs - package org.apache.beam.sdk.extensions.gcp.util.gcsfs
- 
Defines utilities used to interact with Google Cloud Storage. 
- org.apache.beam.sdk.extensions.jackson - package org.apache.beam.sdk.extensions.jackson
- 
Utilities for parsing and creating JSON serialized objects. 
- org.apache.beam.sdk.extensions.joinlibrary - package org.apache.beam.sdk.extensions.joinlibrary
- 
Utilities for performing SQL-style joins of keyed  PCollections. 
- org.apache.beam.sdk.extensions.ml - package org.apache.beam.sdk.extensions.ml
- 
Provides DoFns for integration with Google Cloud AI Video Intelligence service. 
- org.apache.beam.sdk.extensions.ordered - package org.apache.beam.sdk.extensions.ordered
- 
Provides a transform for ordered processing. 
- org.apache.beam.sdk.extensions.ordered.combiner - package org.apache.beam.sdk.extensions.ordered.combiner
- 
Default implementation of the global sequence combiner used by  OrderedEventProcessor when processing events using global
 sequences. 
- org.apache.beam.sdk.extensions.protobuf - package org.apache.beam.sdk.extensions.protobuf
- 
Defines a  Coder for Protocol Buffers messages,  ProtoCoder. 
- org.apache.beam.sdk.extensions.python - package org.apache.beam.sdk.extensions.python
- 
Extensions for invoking Python transforms from the Beam Java SDK. 
- org.apache.beam.sdk.extensions.python.transforms - package org.apache.beam.sdk.extensions.python.transforms
- 
Extensions for invoking Python transforms from the Beam Java SDK. 
- org.apache.beam.sdk.extensions.sbe - package org.apache.beam.sdk.extensions.sbe
- 
Extension for working with SBE messages in Beam. 
- org.apache.beam.sdk.extensions.schemaio.expansion - package org.apache.beam.sdk.extensions.schemaio.expansion
- 
External Transform Registration for SchemaIOs. 
- org.apache.beam.sdk.extensions.sketching - package org.apache.beam.sdk.extensions.sketching
- 
Utilities for computing statistical indicators using probabilistic sketches. 
- org.apache.beam.sdk.extensions.sorter - package org.apache.beam.sdk.extensions.sorter
- 
Utility for performing local sort of potentially large sets of values. 
- org.apache.beam.sdk.extensions.sql - package org.apache.beam.sdk.extensions.sql
- 
BeamSQL provides a new interface to run a SQL statement with Beam. 
- org.apache.beam.sdk.extensions.sql.example - package org.apache.beam.sdk.extensions.sql.example
- 
Example how to use Data Catalog table provider. 
- org.apache.beam.sdk.extensions.sql.example.model - package org.apache.beam.sdk.extensions.sql.example.model
- 
Java classes used to for modeling the examples. 
- org.apache.beam.sdk.extensions.sql.expansion - package org.apache.beam.sdk.extensions.sql.expansion
- 
External Transform Registration for Beam SQL. 
- org.apache.beam.sdk.extensions.sql.impl - package org.apache.beam.sdk.extensions.sql.impl
- 
Implementation classes of BeamSql. 
- org.apache.beam.sdk.extensions.sql.impl.cep - package org.apache.beam.sdk.extensions.sql.impl.cep
- 
Utilities for Complex Event Processing (CEP). 
- org.apache.beam.sdk.extensions.sql.impl.nfa - package org.apache.beam.sdk.extensions.sql.impl.nfa
- 
Package of Non-deterministic Finite Automata (NFA) for MATCH_RECOGNIZE.
 
- org.apache.beam.sdk.extensions.sql.impl.parser - package org.apache.beam.sdk.extensions.sql.impl.parser
- 
Beam SQL parsing additions to Calcite SQL. 
- org.apache.beam.sdk.extensions.sql.impl.planner - package org.apache.beam.sdk.extensions.sql.impl.planner
- 
BeamQueryPlanneris the main interface.
 
- org.apache.beam.sdk.extensions.sql.impl.rel - package org.apache.beam.sdk.extensions.sql.impl.rel
- 
BeamSQL specified nodes, to replace RelNode.
 
- org.apache.beam.sdk.extensions.sql.impl.rule - package org.apache.beam.sdk.extensions.sql.impl.rule
- 
- org.apache.beam.sdk.extensions.sql.impl.schema - package org.apache.beam.sdk.extensions.sql.impl.schema
- 
define table schema, to map with Beam IO components. 
- org.apache.beam.sdk.extensions.sql.impl.transform - package org.apache.beam.sdk.extensions.sql.impl.transform
- 
- org.apache.beam.sdk.extensions.sql.impl.transform.agg - package org.apache.beam.sdk.extensions.sql.impl.transform.agg
- 
Implementation of standard SQL aggregation functions, e.g. 
- org.apache.beam.sdk.extensions.sql.impl.udaf - package org.apache.beam.sdk.extensions.sql.impl.udaf
- 
UDAF classes. 
- org.apache.beam.sdk.extensions.sql.impl.udf - package org.apache.beam.sdk.extensions.sql.impl.udf
- 
UDF classes. 
- org.apache.beam.sdk.extensions.sql.impl.utils - package org.apache.beam.sdk.extensions.sql.impl.utils
- 
Utility classes. 
- org.apache.beam.sdk.extensions.sql.meta - package org.apache.beam.sdk.extensions.sql.meta
- 
Metadata related classes. 
- org.apache.beam.sdk.extensions.sql.meta.provider - package org.apache.beam.sdk.extensions.sql.meta.provider
- 
Table providers. 
- org.apache.beam.sdk.extensions.sql.meta.provider.avro - package org.apache.beam.sdk.extensions.sql.meta.provider.avro
- 
Table schema for AvroIO. 
- org.apache.beam.sdk.extensions.sql.meta.provider.bigquery - package org.apache.beam.sdk.extensions.sql.meta.provider.bigquery
- 
Table schema for BigQuery. 
- org.apache.beam.sdk.extensions.sql.meta.provider.bigtable - package org.apache.beam.sdk.extensions.sql.meta.provider.bigtable
- 
Table schema for BigTable. 
- org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog - package org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog
- 
Table schema for Google Cloud Data Catalog. 
- org.apache.beam.sdk.extensions.sql.meta.provider.datastore - package org.apache.beam.sdk.extensions.sql.meta.provider.datastore
- 
Table schema for DataStore. 
- org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog - package org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog
- 
Table schema for HCatalog. 
- org.apache.beam.sdk.extensions.sql.meta.provider.kafka - package org.apache.beam.sdk.extensions.sql.meta.provider.kafka
- 
Table schema for KafkaIO. 
- org.apache.beam.sdk.extensions.sql.meta.provider.mongodb - package org.apache.beam.sdk.extensions.sql.meta.provider.mongodb
- 
Table schema for MongoDb. 
- org.apache.beam.sdk.extensions.sql.meta.provider.parquet - package org.apache.beam.sdk.extensions.sql.meta.provider.parquet
- 
Table schema for ParquetIO. 
- org.apache.beam.sdk.extensions.sql.meta.provider.pubsub - package org.apache.beam.sdk.extensions.sql.meta.provider.pubsub
- 
- org.apache.beam.sdk.extensions.sql.meta.provider.pubsublite - package org.apache.beam.sdk.extensions.sql.meta.provider.pubsublite
- 
Provides abstractions for schema-aware IOs. 
- org.apache.beam.sdk.extensions.sql.meta.provider.seqgen - package org.apache.beam.sdk.extensions.sql.meta.provider.seqgen
- 
Table schema for streaming sequence generator. 
- org.apache.beam.sdk.extensions.sql.meta.provider.test - package org.apache.beam.sdk.extensions.sql.meta.provider.test
- 
Table schema for in-memory test data. 
- org.apache.beam.sdk.extensions.sql.meta.provider.text - package org.apache.beam.sdk.extensions.sql.meta.provider.text
- 
Table schema for text files. 
- org.apache.beam.sdk.extensions.sql.meta.store - package org.apache.beam.sdk.extensions.sql.meta.store
- 
Meta stores. 
- org.apache.beam.sdk.extensions.sql.provider - package org.apache.beam.sdk.extensions.sql.provider
- 
Package containing UDF providers for testing. 
- org.apache.beam.sdk.extensions.sql.udf - package org.apache.beam.sdk.extensions.sql.udf
- 
Provides interfaces for defining user-defined functions in Beam SQL. 
- org.apache.beam.sdk.extensions.sql.zetasql - package org.apache.beam.sdk.extensions.sql.zetasql
- 
ZetaSQL Dialect package. 
- org.apache.beam.sdk.extensions.sql.zetasql.translation - package org.apache.beam.sdk.extensions.sql.zetasql.translation
- 
Conversion logic between ZetaSQL resolved query nodes and Calcite rel nodes. 
- org.apache.beam.sdk.extensions.sql.zetasql.translation.impl - package org.apache.beam.sdk.extensions.sql.zetasql.translation.impl
- 
Java implementation of ZetaSQL functions. 
- org.apache.beam.sdk.extensions.sql.zetasql.unnest - package org.apache.beam.sdk.extensions.sql.zetasql.unnest
- 
Temporary solution to support ZetaSQL UNNEST. 
- org.apache.beam.sdk.extensions.timeseries - package org.apache.beam.sdk.extensions.timeseries
- 
Utilities for operating on timeseries data. 
- org.apache.beam.sdk.extensions.zetasketch - package org.apache.beam.sdk.extensions.zetasketch
- 
PTransforms to compute statistical sketches on data streams based on the  ZetaSketch implementation. 
- org.apache.beam.sdk.fn - package org.apache.beam.sdk.fn
- 
The top level package for the Fn Execution Java libraries. 
- org.apache.beam.sdk.fn.channel - package org.apache.beam.sdk.fn.channel
- 
gRPC channel management. 
- org.apache.beam.sdk.fn.data - package org.apache.beam.sdk.fn.data
- 
Classes to interact with the portability framework data plane. 
- org.apache.beam.sdk.fn.server - package org.apache.beam.sdk.fn.server
- 
gPRC server factory. 
- org.apache.beam.sdk.fn.splittabledofn - package org.apache.beam.sdk.fn.splittabledofn
- 
- org.apache.beam.sdk.fn.stream - package org.apache.beam.sdk.fn.stream
- 
gRPC stream management. 
- org.apache.beam.sdk.fn.test - package org.apache.beam.sdk.fn.test
- 
Utilities for testing use of this package. 
- org.apache.beam.sdk.fn.windowing - package org.apache.beam.sdk.fn.windowing
- 
Common utilities related to windowing during execution of a pipeline. 
- org.apache.beam.sdk.function - package org.apache.beam.sdk.function
- 
Java 8 functional interface extensions. 
- org.apache.beam.sdk.harness - package org.apache.beam.sdk.harness
- 
Utilities for configuring worker environment. 
- org.apache.beam.sdk.io - package org.apache.beam.sdk.io
- 
Defines transforms for reading and writing common storage formats, including  org.apache.beam.sdk.io.AvroIO, and  TextIO. 
- org.apache.beam.sdk.io.amqp - package org.apache.beam.sdk.io.amqp
- 
Transforms for reading and writing using AMQP 1.0 protocol. 
- org.apache.beam.sdk.io.aws2.common - package org.apache.beam.sdk.io.aws2.common
- 
Common code for AWS sources and sinks such as retry configuration. 
- org.apache.beam.sdk.io.aws2.dynamodb - package org.apache.beam.sdk.io.aws2.dynamodb
- 
Defines IO connectors for Amazon Web Services DynamoDB. 
- org.apache.beam.sdk.io.aws2.kinesis - package org.apache.beam.sdk.io.aws2.kinesis
- 
Transforms for reading from Amazon Kinesis. 
- org.apache.beam.sdk.io.aws2.options - package org.apache.beam.sdk.io.aws2.options
- 
Defines  PipelineOptions for configuring pipeline execution
 for Amazon Web Services components. 
- org.apache.beam.sdk.io.aws2.s3 - package org.apache.beam.sdk.io.aws2.s3
- 
Defines IO connectors for Amazon Web Services S3. 
- org.apache.beam.sdk.io.aws2.schemas - package org.apache.beam.sdk.io.aws2.schemas
- 
Schemas for AWS model classes. 
- org.apache.beam.sdk.io.aws2.sns - package org.apache.beam.sdk.io.aws2.sns
- 
Defines IO connectors for Amazon Web Services SNS. 
- org.apache.beam.sdk.io.aws2.sqs - package org.apache.beam.sdk.io.aws2.sqs
- 
Defines IO connectors for Amazon Web Services SQS. 
- org.apache.beam.sdk.io.azure.blobstore - package org.apache.beam.sdk.io.azure.blobstore
- 
Defines IO connectors for Azure Blob Storage. 
- org.apache.beam.sdk.io.azure.cosmos - package org.apache.beam.sdk.io.azure.cosmos
- 
Defines IO connectors for Azure Cosmos DB. 
- org.apache.beam.sdk.io.azure.options - package org.apache.beam.sdk.io.azure.options
- 
Defines IO connectors for Microsoft Azure Blobstore. 
- org.apache.beam.sdk.io.cassandra - package org.apache.beam.sdk.io.cassandra
- 
Transforms for reading and writing from/to Apache Cassandra. 
- org.apache.beam.sdk.io.cdap - package org.apache.beam.sdk.io.cdap
- 
Transforms for reading and writing from CDAP. 
- org.apache.beam.sdk.io.cdap.context - package org.apache.beam.sdk.io.cdap.context
- 
Context for CDAP classes. 
- org.apache.beam.sdk.io.clickhouse - package org.apache.beam.sdk.io.clickhouse
- 
Transform for writing to ClickHouse. 
- org.apache.beam.sdk.io.contextualtextio - package org.apache.beam.sdk.io.contextualtextio
- 
Transforms for reading from Files with contextual Information. 
- org.apache.beam.sdk.io.csv - package org.apache.beam.sdk.io.csv
- 
Transforms for reading and writing CSV files. 
- org.apache.beam.sdk.io.csv.providers - package org.apache.beam.sdk.io.csv.providers
- 
Transforms for reading and writing CSV files. 
- org.apache.beam.sdk.io.elasticsearch - package org.apache.beam.sdk.io.elasticsearch
- 
Common test utilities for Elasticsearch. 
- org.apache.beam.sdk.io.fileschematransform - package org.apache.beam.sdk.io.fileschematransform
- 
Defines transforms for File reading and writing support with Schema Transform. 
- org.apache.beam.sdk.io.fs - package org.apache.beam.sdk.io.fs
- 
Apache Beam FileSystem interfaces and their default implementations. 
- org.apache.beam.sdk.io.gcp.bigquery - package org.apache.beam.sdk.io.gcp.bigquery
- 
Defines transforms for reading and writing from Google BigQuery. 
- org.apache.beam.sdk.io.gcp.bigquery.providers - package org.apache.beam.sdk.io.gcp.bigquery.providers
- 
Defines SchemaTransformProviders for reading and writing from Google BigQuery. 
- org.apache.beam.sdk.io.gcp.bigtable - package org.apache.beam.sdk.io.gcp.bigtable
- 
Defines transforms for reading and writing from Google Cloud Bigtable. 
- org.apache.beam.sdk.io.gcp.bigtable.changestreams - package org.apache.beam.sdk.io.gcp.bigtable.changestreams
- 
Change stream for Google Cloud Bigtable. 
- org.apache.beam.sdk.io.gcp.bigtable.changestreams.action - package org.apache.beam.sdk.io.gcp.bigtable.changestreams.action
- 
Business logic to process change stream for Google Cloud Bigtable. 
- org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao - package org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao
- 
Data access object for change stream for Google Cloud Bigtable. 
- org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn - package org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn
- 
DoFn and SDF definitions to process Google Cloud Bigtable Change Streams. 
- org.apache.beam.sdk.io.gcp.bigtable.changestreams.encoder - package org.apache.beam.sdk.io.gcp.bigtable.changestreams.encoder
- 
Encoders for writing and reading from Metadata Table for Google Cloud Bigtable Change Streams. 
- org.apache.beam.sdk.io.gcp.bigtable.changestreams.estimator - package org.apache.beam.sdk.io.gcp.bigtable.changestreams.estimator
- 
Classes related to estimating the throughput of the change streams SDFs. 
- org.apache.beam.sdk.io.gcp.bigtable.changestreams.model - package org.apache.beam.sdk.io.gcp.bigtable.changestreams.model
- 
User models for the Google Cloud Bigtable change stream API. 
- org.apache.beam.sdk.io.gcp.bigtable.changestreams.reconciler - package org.apache.beam.sdk.io.gcp.bigtable.changestreams.reconciler
- 
Partition reconciler for Google Cloud Bigtable Change Streams. 
- org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction - package org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction
- 
Custom RestrictionTracker for Google Cloud Bigtable Change Streams. 
- org.apache.beam.sdk.io.gcp.common - package org.apache.beam.sdk.io.gcp.common
- 
Defines common Google Cloud Platform IO support classes. 
- org.apache.beam.sdk.io.gcp.datastore - package org.apache.beam.sdk.io.gcp.datastore
- 
Provides an API for reading from and writing to  Google Cloud Datastore over different
 versions of the Cloud Datastore Client libraries. 
- org.apache.beam.sdk.io.gcp.firestore - package org.apache.beam.sdk.io.gcp.firestore
- 
- org.apache.beam.sdk.io.gcp.healthcare - package org.apache.beam.sdk.io.gcp.healthcare
- 
Provides an API for reading from and writing to  Google Cloud Datastore over different
 versions of the Cloud Datastore Client libraries. 
- org.apache.beam.sdk.io.gcp.pubsub - package org.apache.beam.sdk.io.gcp.pubsub
- 
- org.apache.beam.sdk.io.gcp.pubsublite - package org.apache.beam.sdk.io.gcp.pubsublite
- 
Defines transforms for reading and writing from Google Cloud Pub/Sub Lite. 
- org.apache.beam.sdk.io.gcp.pubsublite.internal - package org.apache.beam.sdk.io.gcp.pubsublite.internal
- 
Defines transforms for reading and writing from Google Cloud Pub/Sub Lite. 
- org.apache.beam.sdk.io.gcp.spanner - package org.apache.beam.sdk.io.gcp.spanner
- 
- org.apache.beam.sdk.io.gcp.spanner.changestreams - package org.apache.beam.sdk.io.gcp.spanner.changestreams
- 
- org.apache.beam.sdk.io.gcp.spanner.changestreams.action - package org.apache.beam.sdk.io.gcp.spanner.changestreams.action
- 
Action processors for each of the types of Change Stream records received. 
- org.apache.beam.sdk.io.gcp.spanner.changestreams.cache - package org.apache.beam.sdk.io.gcp.spanner.changestreams.cache
- 
Caching strategy for watermark. 
- org.apache.beam.sdk.io.gcp.spanner.changestreams.dao - package org.apache.beam.sdk.io.gcp.spanner.changestreams.dao
- 
Database Access Objects for querying change streams and modifying the Connector's metadata
 tables. 
- org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn - package org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn
- 
DoFn and SDF definitions to process Google Cloud Spanner Change Streams. 
- org.apache.beam.sdk.io.gcp.spanner.changestreams.encoder - package org.apache.beam.sdk.io.gcp.spanner.changestreams.encoder
- 
User model for the Spanner change stream API. 
- org.apache.beam.sdk.io.gcp.spanner.changestreams.estimator - package org.apache.beam.sdk.io.gcp.spanner.changestreams.estimator
- 
Classes related to estimating the throughput of the change streams SDFs. 
- org.apache.beam.sdk.io.gcp.spanner.changestreams.mapper - package org.apache.beam.sdk.io.gcp.spanner.changestreams.mapper
- 
Mapping related functionality, such as from ResultSets to Change
 Stream models.
 
- org.apache.beam.sdk.io.gcp.spanner.changestreams.model - package org.apache.beam.sdk.io.gcp.spanner.changestreams.model
- 
User models for the Spanner change stream API. 
- org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction - package org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction
- 
Custom restriction tracker related classes. 
- org.apache.beam.sdk.io.gcp.testing - package org.apache.beam.sdk.io.gcp.testing
- 
Defines utilities for unit testing Google Cloud Platform components of Apache Beam pipelines. 
- org.apache.beam.sdk.io.googleads - package org.apache.beam.sdk.io.googleads
- 
Defines transforms for reading from Google Ads. 
- org.apache.beam.sdk.io.hadoop - package org.apache.beam.sdk.io.hadoop
- 
Classes shared by Hadoop based IOs. 
- org.apache.beam.sdk.io.hadoop.format - package org.apache.beam.sdk.io.hadoop.format
- 
Defines transforms for writing to Data sinks that implement  HadoopFormatIO . 
- org.apache.beam.sdk.io.hbase - package org.apache.beam.sdk.io.hbase
- 
Transforms for reading and writing from/to Apache HBase. 
- org.apache.beam.sdk.io.hcatalog - package org.apache.beam.sdk.io.hcatalog
- 
Transforms for reading and writing using HCatalog. 
- org.apache.beam.sdk.io.hdfs - package org.apache.beam.sdk.io.hdfs
- 
FileSystem implementation for any Hadoop  FileSystem. 
- org.apache.beam.sdk.io.iceberg - package org.apache.beam.sdk.io.iceberg
- 
Iceberg connectors. 
- org.apache.beam.sdk.io.influxdb - package org.apache.beam.sdk.io.influxdb
- 
Transforms for reading and writing from/to InfluxDB. 
- org.apache.beam.sdk.io.jdbc - package org.apache.beam.sdk.io.jdbc
- 
Transforms for reading and writing from JDBC. 
- org.apache.beam.sdk.io.jdbc.providers - package org.apache.beam.sdk.io.jdbc.providers
- 
Transforms for reading and writing from JDBC. 
- org.apache.beam.sdk.io.jms - package org.apache.beam.sdk.io.jms
- 
Transforms for reading and writing from JMS (Java Messaging Service). 
- org.apache.beam.sdk.io.json - package org.apache.beam.sdk.io.json
- 
Transforms for reading and writing JSON files. 
- org.apache.beam.sdk.io.json.providers - package org.apache.beam.sdk.io.json.providers
- 
Transforms for reading and writing JSON files. 
- org.apache.beam.sdk.io.kafka - package org.apache.beam.sdk.io.kafka
- 
Transforms for reading and writing from Apache Kafka. 
- org.apache.beam.sdk.io.kafka.serialization - package org.apache.beam.sdk.io.kafka.serialization
- 
Kafka serializers and deserializers. 
- org.apache.beam.sdk.io.kafka.upgrade - package org.apache.beam.sdk.io.kafka.upgrade
- 
A library to support upgrading Kafka transforms without upgrading the pipeline. 
- org.apache.beam.sdk.io.kudu - package org.apache.beam.sdk.io.kudu
- 
Transforms for reading and writing from/to Apache Kudu. 
- org.apache.beam.sdk.io.mongodb - package org.apache.beam.sdk.io.mongodb
- 
Transforms for reading and writing from MongoDB. 
- org.apache.beam.sdk.io.mqtt - package org.apache.beam.sdk.io.mqtt
- 
Transforms for reading and writing from MQTT. 
- org.apache.beam.sdk.io.neo4j - package org.apache.beam.sdk.io.neo4j
- 
Transforms for reading from and writing to from Neo4j. 
- org.apache.beam.sdk.io.parquet - package org.apache.beam.sdk.io.parquet
- 
Transforms for reading and writing from Parquet. 
- org.apache.beam.sdk.io.pulsar - package org.apache.beam.sdk.io.pulsar
- 
Transforms for reading and writing from Apache Pulsar. 
- org.apache.beam.sdk.io.rabbitmq - package org.apache.beam.sdk.io.rabbitmq
- 
Transforms for reading and writing from RabbitMQ. 
- org.apache.beam.sdk.io.range - package org.apache.beam.sdk.io.range
- 
Provides thread-safe helpers for implementing dynamic work rebalancing in position-based bounded
 sources. 
- org.apache.beam.sdk.io.redis - package org.apache.beam.sdk.io.redis
- 
Transforms for reading and writing from Redis. 
- org.apache.beam.sdk.io.singlestore - package org.apache.beam.sdk.io.singlestore
- 
Transforms for reading and writing from SingleStoreDB. 
- org.apache.beam.sdk.io.singlestore.schematransform - package org.apache.beam.sdk.io.singlestore.schematransform
- 
SingleStoreIO SchemaTransforms. 
- org.apache.beam.sdk.io.snowflake - package org.apache.beam.sdk.io.snowflake
- 
Snowflake IO transforms. 
- org.apache.beam.sdk.io.snowflake.crosslanguage - package org.apache.beam.sdk.io.snowflake.crosslanguage
- 
Cross-language for SnowflakeIO. 
- org.apache.beam.sdk.io.snowflake.data - package org.apache.beam.sdk.io.snowflake.data
- 
Snowflake IO data types. 
- org.apache.beam.sdk.io.snowflake.data.datetime - package org.apache.beam.sdk.io.snowflake.data.datetime
- 
Snowflake IO date/time types. 
- org.apache.beam.sdk.io.snowflake.data.geospatial - package org.apache.beam.sdk.io.snowflake.data.geospatial
- 
Snowflake IO geospatial types. 
- org.apache.beam.sdk.io.snowflake.data.logical - package org.apache.beam.sdk.io.snowflake.data.logical
- 
Snowflake IO logical types. 
- org.apache.beam.sdk.io.snowflake.data.numeric - package org.apache.beam.sdk.io.snowflake.data.numeric
- 
Snowflake IO numeric types. 
- org.apache.beam.sdk.io.snowflake.data.structured - package org.apache.beam.sdk.io.snowflake.data.structured
- 
Snowflake IO structured types. 
- org.apache.beam.sdk.io.snowflake.data.text - package org.apache.beam.sdk.io.snowflake.data.text
- 
Snowflake IO text types. 
- org.apache.beam.sdk.io.snowflake.enums - package org.apache.beam.sdk.io.snowflake.enums
- 
Snowflake IO data types. 
- org.apache.beam.sdk.io.snowflake.services - package org.apache.beam.sdk.io.snowflake.services
- 
Snowflake IO services and POJOs. 
- org.apache.beam.sdk.io.solace - package org.apache.beam.sdk.io.solace
- 
Solace IO connector. 
- org.apache.beam.sdk.io.solace.broker - package org.apache.beam.sdk.io.solace.broker
- 
Solace IO broker-related classes. 
- org.apache.beam.sdk.io.solace.data - package org.apache.beam.sdk.io.solace.data
- 
Solace IO connector - data-related classes. 
- org.apache.beam.sdk.io.solace.read - package org.apache.beam.sdk.io.solace.read
- 
Solace IO connector - read connector classes. 
- org.apache.beam.sdk.io.solace.write - package org.apache.beam.sdk.io.solace.write
- 
SolaceIO Write connector. 
- org.apache.beam.sdk.io.solr - package org.apache.beam.sdk.io.solr
- 
Transforms for reading and writing from/to Solr. 
- org.apache.beam.sdk.io.sparkreceiver - package org.apache.beam.sdk.io.sparkreceiver
- 
Transforms for reading and writing from streaming CDAP plugins. 
- org.apache.beam.sdk.io.splunk - package org.apache.beam.sdk.io.splunk
- 
Transforms for writing events to Splunk's Http Event Collector (HEC). 
- org.apache.beam.sdk.io.thrift - package org.apache.beam.sdk.io.thrift
- 
Transforms for reading and writing to Thrift files. 
- org.apache.beam.sdk.io.tika - package org.apache.beam.sdk.io.tika
- 
Transform for reading and parsing files with Apache Tika. 
- org.apache.beam.sdk.io.xml - package org.apache.beam.sdk.io.xml
- 
Transforms for reading and writing Xml files. 
- org.apache.beam.sdk.jmh.io - package org.apache.beam.sdk.jmh.io
- 
Benchmarks for IO. 
- org.apache.beam.sdk.jmh.schemas - package org.apache.beam.sdk.jmh.schemas
- 
Benchmarks for schemas. 
- org.apache.beam.sdk.jmh.util - package org.apache.beam.sdk.jmh.util
- 
Benchmarks for core SDK utility classes. 
- org.apache.beam.sdk.managed - package org.apache.beam.sdk.managed
- 
Managed reads and writes. 
- org.apache.beam.sdk.managed.testing - package org.apache.beam.sdk.managed.testing
- 
Test transform for Managed API. 
- org.apache.beam.sdk.metrics - package org.apache.beam.sdk.metrics
- 
Metrics allow exporting information about the execution of a pipeline. 
- org.apache.beam.sdk.options - package org.apache.beam.sdk.options
- 
- org.apache.beam.sdk.providers - package org.apache.beam.sdk.providers
- 
- org.apache.beam.sdk.schemas - package org.apache.beam.sdk.schemas
- 
Defines  Schema and other classes for representing schema'd
 data in a  Pipeline. 
- org.apache.beam.sdk.schemas.annotations - package org.apache.beam.sdk.schemas.annotations
- 
Defines  Schema and other classes for representing schema'd
 data in a  Pipeline. 
- org.apache.beam.sdk.schemas.io - package org.apache.beam.sdk.schemas.io
- 
Provides abstractions for schema-aware IOs. 
- org.apache.beam.sdk.schemas.io.payloads - package org.apache.beam.sdk.schemas.io.payloads
- 
Provides abstractions for schema-aware IOs. 
- org.apache.beam.sdk.schemas.logicaltypes - package org.apache.beam.sdk.schemas.logicaltypes
- 
A set of common LogicalTypes for use with schemas. 
- org.apache.beam.sdk.schemas.parser - package org.apache.beam.sdk.schemas.parser
- 
Defines utilities for deailing with schemas. 
- org.apache.beam.sdk.schemas.parser.generated - package org.apache.beam.sdk.schemas.parser.generated
- 
Defines utilities for deailing with schemas. 
- org.apache.beam.sdk.schemas.transforms - package org.apache.beam.sdk.schemas.transforms
- 
Defines transforms that work on PCollections with schemas.. 
- org.apache.beam.sdk.schemas.transforms.providers - package org.apache.beam.sdk.schemas.transforms.providers
- 
Defines transforms that work on PCollections with schemas.. 
- org.apache.beam.sdk.schemas.utils - package org.apache.beam.sdk.schemas.utils
- 
Defines utilities for deailing with schemas. 
- org.apache.beam.sdk.state - package org.apache.beam.sdk.state
- 
Classes and interfaces for interacting with state. 
- org.apache.beam.sdk.testing - package org.apache.beam.sdk.testing
- 
Defines utilities for unit testing Apache Beam pipelines. 
- org.apache.beam.sdk.transforms - package org.apache.beam.sdk.transforms
- 
Defines  PTransforms for transforming data in a pipeline. 
- org.apache.beam.sdk.transforms.display - package org.apache.beam.sdk.transforms.display
- 
- org.apache.beam.sdk.transforms.errorhandling - package org.apache.beam.sdk.transforms.errorhandling
- 
Provides utilities for handling errors in Pipelines. 
- org.apache.beam.sdk.transforms.join - package org.apache.beam.sdk.transforms.join
- 
Defines the  CoGroupByKey transform for joining
 multiple PCollections. 
- org.apache.beam.sdk.transforms.resourcehints - package org.apache.beam.sdk.transforms.resourcehints
- 
- org.apache.beam.sdk.transforms.splittabledofn - package org.apache.beam.sdk.transforms.splittabledofn
- 
- org.apache.beam.sdk.transforms.windowing - package org.apache.beam.sdk.transforms.windowing
- 
Defines the  Window transform for dividing the
 elements in a PCollection into windows, and the  Trigger for controlling when those elements are output. 
- org.apache.beam.sdk.transformservice.launcher - package org.apache.beam.sdk.transformservice.launcher
- 
A library that can be used to start up a Docker-composed based Beam transform service. 
- org.apache.beam.sdk.values - package org.apache.beam.sdk.values
- 
- ORPHANED_NEW_PARTITION_CLEANED_COUNT - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
Counter for the total number of orphaned new partitions cleaned up. 
- OrphanedMetadataCleaner - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.reconciler
-  
- OrphanedMetadataCleaner() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.reconciler.OrphanedMetadataCleaner
-  
- out() - Static method in class org.apache.beam.runners.spark.io.ConsoleIO.Write
- 
- out(int) - Static method in class org.apache.beam.runners.spark.io.ConsoleIO.Write
- 
- OUT - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Read
- 
The tag for the main output of FHIR resources. 
- OUT - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Search
- 
The tag for the main output of FHIR Resources from a search. 
- OUT - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.FhirIOPatientEverything
- 
The tag for the main output of FHIR Resources from a GetPatientEverything request. 
- OUT - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.HL7v2Read
- 
The tag for the main output of HL7v2 read responses. 
- OUT - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.Read
- 
The tag for the main output of HL7v2 Messages. 
- outbound(DataStreams.OutputChunkConsumer<ByteString>) - Static method in class org.apache.beam.sdk.fn.stream.DataStreams
- 
Converts a single element delimited OutputStreaminto multipleByteStrings.
 
- outbound(DataStreams.OutputChunkConsumer<ByteString>, int) - Static method in class org.apache.beam.sdk.fn.stream.DataStreams
- 
Converts a single element delimited OutputStreaminto multipleByteStringsusing the specified maximum chunk size.
 
- OutboundObserverFactory - Class in org.apache.beam.sdk.fn.stream
- 
Creates factories which determine an underlying StreamObserverimplementation to use in
 to interact with fn execution APIs.
 
- OutboundObserverFactory() - Constructor for class org.apache.beam.sdk.fn.stream.OutboundObserverFactory
-  
- OutboundObserverFactory.BasicFactory<ReqT,RespT> - Interface in org.apache.beam.sdk.fn.stream
- 
Creates an outbound observer for the given inbound observer. 
- outboundObserverFor(StreamObserver<ReqT>) - Method in interface org.apache.beam.sdk.fn.stream.OutboundObserverFactory.BasicFactory
-  
- outboundObserverFor(OutboundObserverFactory.BasicFactory<ReqT, RespT>, StreamObserver<ReqT>) - Method in class org.apache.beam.sdk.fn.stream.OutboundObserverFactory
- 
Creates an outbound observer for the given inbound observer by potentially inserting hooks into
 the inbound and outbound observers. 
- OUTER - Static variable in class org.apache.beam.sdk.coders.Coder.Context
- 
Deprecated. The outer context: the value being encoded or decoded takes up the remainder of the
 record/stream contents. 
- OutgoingMessage() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.OutgoingMessage
-  
- OUTPUT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- output() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedEventProcessorResult
-  
- output(T) - Method in interface org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.ParserCallback
- 
Output the object. 
- output(T, Instant) - Method in interface org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.ParserCallback
- 
Output the object using the specified timestamp. 
- output(OutputT, Instant, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFn.FinishBundleContext
- 
Adds the given element to the main output PCollectionat the given timestamp in the
 given window.
 
- output(TupleTag<T>, T, Instant, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFn.FinishBundleContext
- 
Adds the given element to the output PCollectionwith the given tag at the given
 timestamp in the given window.
 
- output(T) - Method in interface org.apache.beam.sdk.transforms.DoFn.OutputReceiver
-  
- output(OutputT) - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
- 
Adds the given element to the main output PCollection.
 
- output(TupleTag<T>, T) - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
- 
Adds the given element to the output PCollectionwith the given tag.
 
- output() - Method in class org.apache.beam.sdk.transforms.WithFailures.Result
-  
- OUTPUT_DIR - Static variable in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
- 
MRJobConfig.MAPREDUCE_JOB_DIR.
 
- OUTPUT_FORMAT_CLASS_ATTR - Static variable in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
- 
MRJobConfig.OUTPUT_FORMAT_CLASS_ATTR.
 
- OUTPUT_INFO - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- OUTPUT_KEY_CLASS - Static variable in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
- 
MRJobConfig.OUTPUT_KEY_CLASS.
 
- OUTPUT_NAME - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- OUTPUT_ROWS_TAG - Static variable in class org.apache.beam.sdk.expansion.service.WindowIntoTransformProvider
-  
- OUTPUT_ROWS_TAG - Static variable in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider
-  
- OUTPUT_ROWS_TAG - Static variable in class org.apache.beam.sdk.schemas.transforms.providers.FlattenTransformProvider
-  
- OUTPUT_ROWS_TAG - Static variable in class org.apache.beam.sdk.schemas.transforms.providers.JavaExplodeTransformProvider
-  
- OUTPUT_ROWS_TAG - Static variable in class org.apache.beam.sdk.schemas.transforms.providers.JavaFilterTransformProvider
-  
- OUTPUT_ROWS_TAG - Static variable in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider
-  
- OUTPUT_ROWS_TAG - Static variable in class org.apache.beam.sdk.schemas.transforms.providers.LoggingTransformProvider
-  
- OUTPUT_SCHEMA - Static variable in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformProvider
-  
- OUTPUT_SCHEMA - Static variable in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider
-  
- OUTPUT_TAG - Static variable in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider
-  
- OUTPUT_TAG - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformProvider
-  
- OUTPUT_TAG - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformProvider
-  
- OUTPUT_TAG - Static variable in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider
-  
- OUTPUT_TAG - Static variable in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider
-  
- OUTPUT_TAG - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider
-  
- OUTPUT_TAG - Static variable in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformProvider
-  
- OUTPUT_TAG - Static variable in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider
-  
- OUTPUT_VALUE_CLASS - Static variable in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
- 
MRJobConfig.OUTPUT_VALUE_CLASS.
 
- outputCollectionNames() - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.expansion.service.WindowIntoTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.extensions.sql.expansion.SqlTransformSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.csv.providers.CsvWriteTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformProvider
- 
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryExportReadSchemaTransformProvider
- 
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryFileLoadsSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.iceberg.IcebergReadSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.json.providers.JsonWriteTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadProvider
- 
- outputCollectionNames() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteProvider
- 
- outputCollectionNames() - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.schemas.transforms.providers.FlattenTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaExplodeTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaFilterTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider
-  
- outputCollectionNames() - Method in class org.apache.beam.sdk.schemas.transforms.providers.LoggingTransformProvider
-  
- outputCollectionNames() - Method in interface org.apache.beam.sdk.schemas.transforms.SchemaTransformProvider
- 
Returns the output collection names of this transform. 
- outputColumnMap - Variable in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
-  
- outputFormatProvider - Variable in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
- 
- outputOf(ProcessFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
- 
Returns a type descriptor for the output of the given  ProcessFunction, subject to Java
 type erasure: may contain unresolved type variables if the type was erased. 
- outputOf(SerializableFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
- 
- outputOf(Contextful.Fn<InputT, OutputT>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
- 
- OutputRangeTracker(OffsetRange) - Constructor for class org.apache.beam.sdk.transforms.PeriodicSequence.OutputRangeTracker
-  
- OutputReceiverFactory - Interface in org.apache.beam.runners.fnexecution.control
- 
A factory that can create output receivers during an executable stage. 
- OutputReference - Class in org.apache.beam.runners.dataflow.util
- 
A representation used by  Steps to reference the
 output of other  Steps. 
- OutputReference(String, String) - Constructor for class org.apache.beam.runners.dataflow.util.OutputReference
-  
- outputRuntimeOptions() - Method in interface org.apache.beam.sdk.options.PipelineOptions
- 
- outputSchema() - Method in class org.apache.beam.sdk.schemas.transforms.Cast
-  
- outputSchemaCoder - Variable in class org.apache.beam.sdk.schemas.utils.ConvertHelpers.ConvertedSchemaInformation
-  
- OutputTagFilter<OutputT,InputT> - Class in org.apache.beam.runners.twister2.translators.functions
- 
Output tag filter. 
- OutputTagFilter() - Constructor for class org.apache.beam.runners.twister2.translators.functions.OutputTagFilter
-  
- OutputTagFilter(int) - Constructor for class org.apache.beam.runners.twister2.translators.functions.OutputTagFilter
-  
- outputWindowedValue(T, Instant, Collection<? extends BoundedWindow>, PaneInfo) - Method in interface org.apache.beam.sdk.transforms.DoFn.OutputReceiver
-  
- outputWindowedValue(OutputT, Instant, Collection<? extends BoundedWindow>, PaneInfo) - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
- 
Adds the given element to the main output PCollection, with the given windowing
 metadata.
 
- outputWindowedValue(TupleTag<T>, T, Instant, Collection<? extends BoundedWindow>, PaneInfo) - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
- 
Adds the given element to the main output PCollection, with the given windowing
 metadata.
 
- outputWithTimestamp(T, Instant) - Method in interface org.apache.beam.sdk.transforms.DoFn.OutputReceiver
-  
- outputWithTimestamp(OutputT, Instant) - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
- 
Adds the given element to the main output PCollection, with the given timestamp.
 
- outputWithTimestamp(TupleTag<T>, T, Instant) - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
- 
Adds the given element to the specified output PCollection, with the given timestamp.
 
- overlaps(ByteKeyRange) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
- 
Returns  true if the specified  ByteKeyRange overlaps this range. 
- overrideEncodingPositions(UUID, Map<String, Integer>) - Static method in class org.apache.beam.sdk.coders.RowCoder
- 
Override encoding positions for the given schema. 
- overrideEncodingPositions(UUID, Map<String, Integer>) - Static method in class org.apache.beam.sdk.coders.RowCoderGenerator
-  
- overrideEncodingPositions(UUID, Map<String, Integer>) - Static method in class org.apache.beam.sdk.schemas.SchemaCoder
- 
Override encoding positions for the given schema. 
- PackageUtil - Class in org.apache.beam.runners.dataflow.util
- 
Helper routines for packages. 
- PackageUtil.StagedFile - Class in org.apache.beam.runners.dataflow.util
-  
- pane() - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContext
- 
Returns information about the pane within this window into which the input element has been
 assigned. 
- PaneInfo - Class in org.apache.beam.sdk.transforms.windowing
- 
Provides information about the pane an element belongs to. 
- PaneInfo.PaneInfoCoder - Class in org.apache.beam.sdk.transforms.windowing
- 
A Coder for encoding PaneInfo instances. 
- PaneInfo.Timing - Enum in org.apache.beam.sdk.transforms.windowing
- 
Enumerates the possibilities for the timing of this pane firing related to the input and output
 watermarks for its computation. 
- paneInfoFromBytes(byte[]) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.Utils
-  
- paneInfoToBytes(PaneInfo) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.Utils
-  
- PARALLEL_INPUT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- ParameterListBuilder() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase.ParameterListBuilder
-  
- parameters - Variable in class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase
- 
Types of parameter for the function call. 
- Params() - Constructor for class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
- 
Construct a default Params object. 
- ParamsCoder() - Constructor for class org.apache.beam.sdk.io.DefaultFilenamePolicy.ParamsCoder
-  
- ParDo - Class in org.apache.beam.sdk.transforms
- 
ParDo is the core element-wise transform in Apache Beam, invoking a user-specified
 function on each of the elements of the input  PCollection to produce zero or more output
 elements, all of which are collected into the output  PCollection. 
- ParDo() - Constructor for class org.apache.beam.sdk.transforms.ParDo
-  
- ParDo.MultiOutput<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
- 
A  PTransform that, when applied to a  PCollection<InputT>, invokes a
 user-specified  DoFn<InputT, OutputT> on all its elements, which can emit elements to
 any of the  PTransform's output  PCollections, which are bundled into a result
  PCollectionTuple. 
- ParDo.SingleOutput<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
- 
A  PTransform that, when applied to a  PCollection<InputT>, invokes a
 user-specified  DoFn<InputT, OutputT> on all its elements, with all its outputs
 collected into an output  PCollection<OutputT>. 
- ParDoMultiOutputTranslatorBatch<InputT,OutputT> - Class in org.apache.beam.runners.twister2.translators.batch
- 
ParDo translator. 
- ParDoMultiOutputTranslatorBatch() - Constructor for class org.apache.beam.runners.twister2.translators.batch.ParDoMultiOutputTranslatorBatch
-  
- ParDoMultiOverrideFactory<InputT,OutputT> - Class in org.apache.beam.runners.direct
- 
A  PTransformOverrideFactory that provides overrides for applications of a  ParDo
 in the direct runner. 
- ParDoMultiOverrideFactory() - Constructor for class org.apache.beam.runners.direct.ParDoMultiOverrideFactory
-  
- ParDoP<InputT,OutputT> - Class in org.apache.beam.runners.jet.processors
- 
Jet Processorimplementation for Beam's ParDo primitive (when no
 user-state is being used).
 
- ParDoP.Supplier<InputT,OutputT> - Class in org.apache.beam.runners.jet.processors
- 
Jet  Processor supplier that will provide instances of  ParDoP. 
- parent() - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold.Context
- 
Type of parent node in a tree. 
- PARENT_TOKENS - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.InitialPartition
- 
The empty set representing the initial partition parent tokens. 
- ParquetConfiguration() - Constructor for class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.ParquetConfiguration
-  
- parquetConfigurationBuilder() - Static method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration
-  
- ParquetIO - Class in org.apache.beam.sdk.io.parquet
- 
IO to read and write Parquet files. 
- ParquetIO.Parse<T> - Class in org.apache.beam.sdk.io.parquet
- 
- ParquetIO.ParseFiles<T> - Class in org.apache.beam.sdk.io.parquet
- 
- ParquetIO.Read - Class in org.apache.beam.sdk.io.parquet
- 
- ParquetIO.ReadFiles - Class in org.apache.beam.sdk.io.parquet
- 
- ParquetIO.ReadFiles.BlockTracker - Class in org.apache.beam.sdk.io.parquet
-  
- ParquetIO.Sink - Class in org.apache.beam.sdk.io.parquet
- 
- ParquetReadSchemaTransformFormatProvider - Class in org.apache.beam.sdk.io.fileschematransform
-  
- ParquetReadSchemaTransformFormatProvider() - Constructor for class org.apache.beam.sdk.io.fileschematransform.ParquetReadSchemaTransformFormatProvider
-  
- ParquetTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.parquet
- 
- ParquetTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.parquet.ParquetTableProvider
-  
- ParquetWriteSchemaTransformFormatProvider - Class in org.apache.beam.sdk.io.fileschematransform
- 
- ParquetWriteSchemaTransformFormatProvider() - Constructor for class org.apache.beam.sdk.io.fileschematransform.ParquetWriteSchemaTransformFormatProvider
-  
- Parse() - Constructor for class org.apache.beam.sdk.extensions.avro.io.AvroIO.Parse
-  
- parse(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner
- 
Parse input SQL query, and return a SqlNodeas grammar tree.
 
- parse(String) - Method in interface org.apache.beam.sdk.extensions.sql.impl.QueryPlanner
- 
Parse input SQL query, and return a SqlNodeas grammar tree.
 
- parse(String) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
-  
- parse(String) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
- 
- parse(String) - Static method in enum org.apache.beam.sdk.io.clickhouse.TableSchema.DefaultType
-  
- parse(Class<T>, CSVFormat) - Static method in class org.apache.beam.sdk.io.csv.CsvIO
- 
Instantiates a  CsvIOParse for parsing CSV string records into custom  Schema-mapped  Class<T>es from the records' assumed  CsvFormat. 
- parse(GridFSDBFile, MongoDbGridFSIO.ParserCallback<T>) - Method in interface org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Parser
-  
- Parse() - Constructor for class org.apache.beam.sdk.io.parquet.ParquetIO.Parse
-  
- parse() - Static method in class org.apache.beam.sdk.io.tika.TikaIO
- 
Parses files matching a given filepattern. 
- Parse() - Constructor for class org.apache.beam.sdk.io.tika.TikaIO.Parse
-  
- parse(String) - Static method in class org.apache.beam.sdk.schemas.parser.FieldAccessDescriptorParser
-  
- parse(T) - Method in class org.apache.beam.sdk.testing.JsonMatcher
-  
- ParseAll() - Constructor for class org.apache.beam.sdk.extensions.avro.io.AvroIO.ParseAll
- 
Deprecated. 
- parseAllGenericRecords(SerializableFunction<GenericRecord, T>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- parseArgs(String[]) - Static method in class org.apache.beam.runners.flink.FlinkJobServerDriver
-  
- parseDate(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
-  
- parseDateToValue(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
-  
- parseDefaultExpression(TableSchema.ColumnType, String) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
- 
Get default value of a column based on expression. 
- parseDicomWebpath(String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.WebPathParser
-  
- ParsedMetricName() - Constructor for class org.apache.beam.sdk.metrics.LabeledMetricNameUtils.ParsedMetricName
-  
- ParseException - Exception in org.apache.beam.sdk.extensions.sql.impl
- 
Exception thrown when Beam SQL is unable to parse the statement. 
- ParseException(Throwable) - Constructor for exception org.apache.beam.sdk.extensions.sql.impl.ParseException
-  
- ParseException(String, Throwable) - Constructor for exception org.apache.beam.sdk.extensions.sql.impl.ParseException
-  
- ParseFiles() - Constructor for class org.apache.beam.sdk.extensions.avro.io.AvroIO.ParseFiles
-  
- ParseFiles() - Constructor for class org.apache.beam.sdk.io.parquet.ParquetIO.ParseFiles
-  
- parseFiles() - Static method in class org.apache.beam.sdk.io.tika.TikaIO
- 
- ParseFiles() - Constructor for class org.apache.beam.sdk.io.tika.TikaIO.ParseFiles
-  
- parseFilesGenericRecords(SerializableFunction<GenericRecord, T>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- parseFilesGenericRecords(SerializableFunction<GenericRecord, T>) - Static method in class org.apache.beam.sdk.io.parquet.ParquetIO
- 
Reads  GenericRecord from Parquet files and converts to user defined type using provided
  parseFn. 
- parseGenericRecords(SerializableFunction<GenericRecord, T>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
Reads Avro file(s) containing records of an unspecified schema and converting each record to a
 custom type. 
- parseGenericRecords(SerializableFunction<GenericRecord, T>) - Static method in class org.apache.beam.sdk.io.parquet.ParquetIO
- 
Reads  GenericRecord from a Parquet file (or multiple Parquet files matching the
 pattern) and converts to user defined type using provided parseFn. 
- parseInitialContinuationTokens(Row) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.encoder.MetadataTableEncoder
- 
Return a list of initial token from a row. 
- ParseJsons<OutputT> - Class in org.apache.beam.sdk.extensions.jackson
- 
- ParseJsons.ParseJsonsWithFailures<FailureT> - Class in org.apache.beam.sdk.extensions.jackson
- 
A  PTransform that adds exception handling to  ParseJsons. 
- parseLockUuid(Row) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.encoder.MetadataTableEncoder
- 
Returns the uuid from a row. 
- parseMetricName(String) - Static method in class org.apache.beam.sdk.metrics.LabeledMetricNameUtils
- 
Parse a 'metric name' String that was created with 'MetricNameBuilder'. 
- ParsePayloadAsPubsubMessageProto() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessages.ParsePayloadAsPubsubMessageProto
-  
- parseProperties(String) - Static method in class org.apache.beam.sdk.extensions.sql.TableUtils
-  
- ParsePubsubMessageProtoAsPayload() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessages.ParsePubsubMessageProtoAsPayload
-  
- ParsePubsubMessageProtoAsPayloadFromWindowedValue() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite.ParsePubsubMessageProtoAsPayloadFromWindowedValue
-  
- parseQuery(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
-  
- parseQuery(String, QueryPlanner.QueryParameters) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
-  
- ParseResult - Class in org.apache.beam.sdk.io.tika
- 
The result of parsing a single file with Tika: contains the file's location, metadata, extracted
 text, and optionally an error. 
- ParseResult() - Constructor for class org.apache.beam.sdk.transforms.JsonToRow.ParseResult
-  
- parseRows(Schema, CSVFormat) - Static method in class org.apache.beam.sdk.io.csv.CsvIO
- 
- parseTableSpec(String) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
- 
Parse a table specification in the form "[project_id]:[dataset_id].[table_id]"or"[project_id].[dataset_id].[table_id]"or"[dataset_id].[table_id]".
 
- parseTableUrn(String) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
-  
- parseTime(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
-  
- parseTimestampAsMsSinceEpoch(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
- 
Return timestamp as ms-since-unix-epoch corresponding to timestamp.
 
- parseTimestampWithLocalTimeZone(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
-  
- parseTimestampWithoutTimeZone(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
-  
- parseTimestampWithTimeZone(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
-  
- parseTimestampWithTZToValue(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
-  
- parseTimestampWithUTCTimeZone(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
-  
- parseTimeToValue(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
-  
- parseTokenFromRow(Row) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.encoder.MetadataTableEncoder
- 
Read the continuation token cell of a row from ReadRows. 
- parseWatermarkFromRow(Row) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.encoder.MetadataTableEncoder
- 
Read the watermark cell of a row from ReadRows. 
- parseWatermarkLastUpdatedFromRow(Row) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.encoder.MetadataTableEncoder
- 
Return the timestamp (the time it was updated) of the watermark cell. 
- ParseWithError() - Constructor for class org.apache.beam.sdk.transforms.JsonToRow.JsonToRowWithErrFn.ParseWithError
-  
- Partition<T> - Class in org.apache.beam.sdk.transforms
- 
Partitiontakes aPCollection<T>and aPartitionFn, uses thePartitionFnto split the elements of the inputPCollectionintoNpartitions,
 and returns aPCollectionList<T>that bundlesNPCollection<T>s
 containing the split elements.
 
- Partition.PartitionFn<T> - Interface in org.apache.beam.sdk.transforms
- 
A function object that chooses an output partition for an element. 
- Partition.PartitionWithSideInputsFn<T> - Interface in org.apache.beam.sdk.transforms
- 
A function object that chooses an output partition for an element. 
- PARTITION_CREATED_TO_SCHEDULED_MS - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
- PARTITION_MERGE_COUNT - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
Counter for the total number of partition merges identified during the execution of the
 Connector. 
- PARTITION_RECONCILED_WITH_TOKEN_COUNT - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
Counter for the total number of partitions reconciled with continuation tokens. 
- PARTITION_RECONCILED_WITHOUT_TOKEN_COUNT - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
Counter for the total number of partitions reconciled without continuation tokens. 
- PARTITION_RECORD_COUNT - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
Counter for the total number of partitions identified during the execution of the Connector. 
- PARTITION_RECORD_MERGE_COUNT - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
Counter for the total number of partition merges identified during the execution of the
 Connector. 
- PARTITION_RECORD_SPLIT_COUNT - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
Counter for the total number of partition splits / moves identified during the execution of the
 Connector. 
- PARTITION_SCHEDULED_TO_RUNNING_MS - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamMetrics
- 
- PARTITION_SPLIT_COUNT - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
Counter for the total number of partition splits / moves identified during the execution of the
 Connector. 
- PARTITION_STREAM_COUNT - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
Counter for the total number of active partitions being streamed. 
- PARTITION_TOKEN - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.InitialPartition
- 
The token of the initial partition. 
- PartitionContext() - Constructor for class org.apache.beam.sdk.io.kafka.TimestampPolicy.PartitionContext
-  
- partitioner() - Method in class org.apache.beam.runners.spark.io.SourceRDD.Unbounded
-  
- PARTITIONER_CLASS_ATTR - Static variable in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
- 
MRJobConfig.PARTITIONER_CLASS_ATTR.
 
- partitionFor(T, int) - Method in interface org.apache.beam.sdk.transforms.Partition.PartitionFn
- 
Chooses the partition into which to put the given element. 
- partitionFor(T, int, Contextful.Fn.Context) - Method in interface org.apache.beam.sdk.transforms.Partition.PartitionWithSideInputsFn
- 
Chooses the partition into which to put the given element. 
- PartitioningWindowFn<T,W extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
- 
A  WindowFn that places each value into exactly one window based on its timestamp and
 never merges windows. 
- PartitioningWindowFn() - Constructor for class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
-  
- PartitionMark(String, int, long, long) - Constructor for class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
-  
- PartitionMetadata - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.model
- 
Model for the partition metadata database table used in the Connector. 
- PartitionMetadata(String, HashSet<String>, Timestamp, Timestamp, long, PartitionMetadata.State, Timestamp, Timestamp, Timestamp, Timestamp, Timestamp) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata
-  
- PartitionMetadata.Builder - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.model
- 
Partition metadata builder for better user experience. 
- PartitionMetadata.State - Enum in org.apache.beam.sdk.io.gcp.spanner.changestreams.model
- 
The state at which a partition can be in the system:
 
   CREATED: the partition has been created, but no query has been done against it yet. 
- PartitionMetadataAdminDao - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.dao
- 
Data access object for creating and dropping the partition metadata table. 
- PartitionMetadataDao - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.dao
- 
Data access object for the Connector metadata tables. 
- PartitionMetadataDao.InTransactionContext - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.dao
- 
Represents the execution of a read / write transaction in Cloud Spanner. 
- PartitionMetadataDao.TransactionResult<T> - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.dao
- 
Represents a result from executing a Cloud Spanner read / write transaction. 
- partitionMetadataMapper() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.mapper.MapperFactory
- 
Creates and returns a single instance of a mapper class capable of transforming a  Struct into a  PartitionMetadata class. 
- PartitionMetadataMapper - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.mapper
- 
- PartitionMetadataTableNames - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.dao
- 
Configuration for a partition metadata table. 
- PartitionMetadataTableNames(String, String, String) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataTableNames
-  
- partitionQuery() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.Read
- 
Factory method to create a new type safe builder for PartitionQueryRequestoperations.
 
- PartitionReconciler - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.reconciler
- 
There can be a race when many splits and merges happen to a single partition in quick succession. 
- PartitionReconciler(MetadataTableDao, ChangeStreamMetrics) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.reconciler.PartitionReconciler
-  
- PartitionRecord - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.model
- 
- PartitionRecord(Range.ByteStringRange, List<ChangeStreamContinuationToken>, Instant, List<NewPartition>) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.PartitionRecord
-  
- PartitionRecord(Range.ByteStringRange, Instant, Instant, List<NewPartition>) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.PartitionRecord
-  
- PartitionRecord(Range.ByteStringRange, Instant, String, Instant, List<NewPartition>, Instant) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.PartitionRecord
-  
- PartitionRecord(Range.ByteStringRange, List<ChangeStreamContinuationToken>, String, Instant, List<NewPartition>, Instant) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.PartitionRecord
-  
- partitionsToString(List<Range.ByteStringRange>) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ByteStringRangeHelper
- 
Convert partitions to a string for debugging. 
- PAssert - Class in org.apache.beam.sdk.testing
- 
An assertion on the contents of a  PCollection incorporated into the pipeline. 
- PAssert.DefaultConcludeTransform - Class in org.apache.beam.sdk.testing
- 
Default transform to check that a PAssert was successful. 
- PAssert.GroupThenAssert<T> - Class in org.apache.beam.sdk.testing
- 
A transform that applies an assertion-checking function over iterables of ActualTto
 the entirety of the contents of its input.
 
- PAssert.GroupThenAssertForSingleton<T> - Class in org.apache.beam.sdk.testing
- 
A transform that applies an assertion-checking function to the sole element of a  PCollection. 
- PAssert.IterableAssert<T> - Interface in org.apache.beam.sdk.testing
- 
Builder interface for assertions applicable to iterables and PCollection contents. 
- PAssert.MatcherCheckerFn<T> - Class in org.apache.beam.sdk.testing
- 
Check that the passed-in matchers match the existing data. 
- PAssert.OneSideInputAssert<ActualT> - Class in org.apache.beam.sdk.testing
- 
An assertion checker that takes a single  PCollectionView<ActualT>
 and an assertion over  ActualT, and checks it within a Beam pipeline. 
- PAssert.PAssertionSite - Class in org.apache.beam.sdk.testing
- 
Track the place where an assertion is defined. 
- PAssert.PCollectionContentsAssert<T> - Class in org.apache.beam.sdk.testing
- 
- PAssert.PCollectionListContentsAssert<T> - Class in org.apache.beam.sdk.testing
- 
- PAssert.SingletonAssert<T> - Interface in org.apache.beam.sdk.testing
- 
Builder interface for assertions applicable to a single value. 
- PassThroughLogicalType<T> - Class in org.apache.beam.sdk.schemas.logicaltypes
- 
A base class for LogicalTypes that use the same Java type as the underlying base type. 
- PassThroughLogicalType(String, Schema.FieldType, Object, Schema.FieldType) - Constructor for class org.apache.beam.sdk.schemas.logicaltypes.PassThroughLogicalType
-  
- password(String) - Method in class org.apache.beam.sdk.io.solace.broker.BasicAuthJcsmpSessionServiceFactory.Builder
- 
Set Solace password. 
- password() - Method in class org.apache.beam.sdk.io.solace.broker.BasicAuthJcsmpSessionServiceFactory
- 
The password to use for authentication. 
- password(String) - Method in class org.apache.beam.sdk.io.solace.broker.BasicAuthSempClientFactory.Builder
- 
Set Solace password. 
- passwordSecretName(String) - Method in class org.apache.beam.sdk.io.solace.broker.GCPSecretSessionServiceFactory.Builder
- 
The Secret Manager secret name where the password is stored. 
- passwordSecretName() - Method in class org.apache.beam.sdk.io.solace.broker.GCPSecretSessionServiceFactory
-  
- passwordSecretVersion(String) - Method in class org.apache.beam.sdk.io.solace.broker.GCPSecretSessionServiceFactory.Builder
- 
Optional. 
- passwordSecretVersion() - Method in class org.apache.beam.sdk.io.solace.broker.GCPSecretSessionServiceFactory
-  
- pastEndOfWindow() - Static method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark
- 
Creates a trigger that fires when the watermark passes the end of the window. 
- pastFirstElementInPane() - Static method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
- 
Creates a trigger that fires when the current processing time passes the processing time at
 which this trigger saw the first element in a pane. 
- patchTableDescription(TableReference, String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
- 
Patch BigQuery  Table description. 
- patchTableDescription(TableReference, String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.DatasetServiceImpl
-  
- patchTableDescription(TableReference, String) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- path - Variable in class org.apache.beam.sdk.jmh.io.TextSourceBenchmark.Data
-  
- path() - Method in class org.apache.beam.sdk.schemas.transforms.Cast.CompatibilityError
-  
- path() - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold.Context
- 
Field path from a root of a schema. 
- pathString - Variable in class org.apache.beam.sdk.jmh.io.TextSourceBenchmark.Data
-  
- PathValidator - Interface in org.apache.beam.sdk.extensions.gcp.storage
- 
For internal use only; no backwards compatibility guarantees. 
- PathValidatorFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcsOptions.PathValidatorFactory
-  
- PatientEverythingParameter() - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.FhirIOPatientEverything.PatientEverythingParameter
-  
- PatternCondition - Class in org.apache.beam.sdk.extensions.sql.impl.cep
- 
PatternConditionstores the function to decide whether a row is a match of a single
 pattern.
 
- payload() - Method in class org.apache.beam.sdk.io.splunk.SplunkWriteError
-  
- PAYLOAD_TOO_LARGE - Static variable in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySinkMetrics
-  
- PayloadSerializer - Interface in org.apache.beam.sdk.schemas.io.payloads
-  
- PayloadSerializerKafkaTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.kafka
-  
- PayloadSerializerProvider - Interface in org.apache.beam.sdk.schemas.io.payloads
-  
- PayloadSerializers - Class in org.apache.beam.sdk.schemas.io.payloads
-  
- payloadToConfig(ExternalTransforms.ExternalConfigurationPayload, Class<ConfigT>) - Static method in class org.apache.beam.sdk.expansion.service.ExpansionService
- 
Attempt to create an instance of ConfigTfrom anExternalTransforms.ExternalConfigurationPayload.
 
- PBegin - Class in org.apache.beam.sdk.values
- 
- PBegin(Pipeline) - Constructor for class org.apache.beam.sdk.values.PBegin
- 
- PCollection<T> - Class in org.apache.beam.sdk.values
- 
- PCollection.IsBounded - Enum in org.apache.beam.sdk.values
- 
The enumeration of cases for whether a  PCollection is bounded. 
- PCOLLECTION_NAME - Static variable in class org.apache.beam.sdk.extensions.sql.SqlTransform
-  
- PCollectionContentsAssert(PCollection<T>, PAssert.PAssertionSite) - Constructor for class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-  
- PCollectionContentsAssert(PCollection<T>, PAssert.AssertionWindows, SimpleFunction<Iterable<ValueInSingleWindow<T>>, Iterable<T>>, PAssert.PAssertionSite) - Constructor for class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-  
- PCollectionList<T> - Class in org.apache.beam.sdk.values
- 
- PCollectionListContentsAssert(PCollectionList<T>) - Constructor for class org.apache.beam.sdk.testing.PAssert.PCollectionListContentsAssert
-  
- PCollectionRowTuple - Class in org.apache.beam.sdk.values
- 
- pCollections() - Static method in class org.apache.beam.sdk.transforms.Flatten
- 
- PCollectionTuple - Class in org.apache.beam.sdk.values
- 
- PCollectionView<T> - Interface in org.apache.beam.sdk.values
- 
- PCollectionViews - Class in org.apache.beam.sdk.values
- 
For internal use only; no backwards compatibility guarantees. 
- PCollectionViews() - Constructor for class org.apache.beam.sdk.values.PCollectionViews
-  
- PCollectionViews.HasDefaultValue<T> - Interface in org.apache.beam.sdk.values
-  
- PCollectionViews.InMemoryListFromMultimapViewFn<T> - Class in org.apache.beam.sdk.values
- 
Implementation which is able to adapt a multimap materialization to an in-memory List<T>.
 
- PCollectionViews.InMemoryListViewFn<T> - Class in org.apache.beam.sdk.values
- 
Implementation which is able to adapt an iterable materialization to an in-memory List<T>.
 
- PCollectionViews.InMemoryMapFromVoidKeyViewFn<K,V> - Class in org.apache.beam.sdk.values
- 
Implementation which is able to adapt a multimap materialization to an in-memory Map<K,
 V>.
 
- PCollectionViews.InMemoryMapViewFn<K,V> - Class in org.apache.beam.sdk.values
- 
Implementation which is able to adapt an iterable materialization to an in-memory Map<K,
 V>.
 
- PCollectionViews.InMemoryMultimapFromVoidKeyViewFn<K,V> - Class in org.apache.beam.sdk.values
- 
Implementation which is able to adapt a multimap materialization to an in-memory Map<K,
 Iterable<V>>.
 
- PCollectionViews.InMemoryMultimapViewFn<K,V> - Class in org.apache.beam.sdk.values
- 
Implementation which is able to adapt an iterable materialization to an in-memory Map<K,
 Iterable<V>>.
 
- PCollectionViews.IsSingletonView<T> - Interface in org.apache.beam.sdk.values
-  
- PCollectionViews.IterableBackedListViewFn<T> - Class in org.apache.beam.sdk.values
- 
Implementation which is able to adapt an iterable materialization to a List<T>.
 
- PCollectionViews.IterableViewFn<T> - Class in org.apache.beam.sdk.values
- 
- PCollectionViews.IterableViewFn2<T> - Class in org.apache.beam.sdk.values
- 
Implementation which is able to adapt an iterable materialization to a Iterable<T>.
 
- PCollectionViews.ListViewFn<T> - Class in org.apache.beam.sdk.values
- 
- PCollectionViews.ListViewFn2<T> - Class in org.apache.beam.sdk.values
- 
Implementation which is able to adapt a multimap materialization to a List<T>.
 
- PCollectionViews.MapViewFn<K,V> - Class in org.apache.beam.sdk.values
- 
- PCollectionViews.MapViewFn2<K,V> - Class in org.apache.beam.sdk.values
- 
Implementation which is able to adapt a multimap materialization to a Map<K, V>.
 
- PCollectionViews.MultimapViewFn<K,V> - Class in org.apache.beam.sdk.values
- 
- PCollectionViews.MultimapViewFn2<K,V> - Class in org.apache.beam.sdk.values
- 
Implementation which is able to adapt a multimap materialization to a Map<K,
 Iterable<V>>.
 
- PCollectionViews.SimplePCollectionView<ElemT,PrimitiveViewT,ViewT,W extends BoundedWindow> - Class in org.apache.beam.sdk.values
- 
A class for  PCollectionView implementations, with additional type parameters that are
 not visible at pipeline assembly time when the view is used as a side input. 
- PCollectionViews.SingletonViewFn<T> - Class in org.apache.beam.sdk.values
- 
- PCollectionViews.SingletonViewFn2<T> - Class in org.apache.beam.sdk.values
- 
Implementation which is able to adapt an iterable materialization to a T.
 
- PCollectionViews.TypeDescriptorSupplier<T> - Interface in org.apache.beam.sdk.values
-  
- PCollectionViews.ValueOrMetadata<T,MetaT> - Class in org.apache.beam.sdk.values
- 
Stores values or metadata about values. 
- PCollectionViews.ValueOrMetadataCoder<T,MetaT> - Class in org.apache.beam.sdk.values
- 
- PCollectionViewTranslatorBatch<ElemT,ViewT> - Class in org.apache.beam.runners.twister2.translators.batch
- 
PCollectionView translator. 
- PCollectionViewTranslatorBatch() - Constructor for class org.apache.beam.runners.twister2.translators.batch.PCollectionViewTranslatorBatch
-  
- PDone - Class in org.apache.beam.sdk.values
- 
- peekOutputElements() - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- peekOutputElements(TupleTag<T>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- peekOutputElementsInWindow(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- peekOutputElementsInWindow(TupleTag<OutputT>, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- peekOutputElementsWithTimestamp() - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- perElement() - Static method in class org.apache.beam.sdk.transforms.Count
- 
- PeriodicImpulse - Class in org.apache.beam.sdk.transforms
- 
A  PTransform which produces a sequence of elements at fixed runtime intervals. 
- PeriodicSequence - Class in org.apache.beam.sdk.transforms
- 
A  PTransform which generates a sequence of timestamped elements at given runtime
 intervals. 
- PeriodicSequence.OutputRangeTracker - Class in org.apache.beam.sdk.transforms
-  
- PeriodicSequence.SequenceDefinition - Class in org.apache.beam.sdk.transforms
-  
- PeriodicStatusPageDirectoryFactory() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions.PeriodicStatusPageDirectoryFactory
-  
- perKey() - Static method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct
- 
Like  ApproximateDistinct.globally() but per key, i.e computes the approximate number of distinct values per
 key in a  PCollection<KV<K, V>> and returns  PCollection<KV<K, Long>>. 
- perKey() - Static method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies
- 
Like  SketchFrequencies.globally() but per key, i.e a Count-min sketch per key in  PCollection<KV<K, V>> and returns a  PCollection<KV<K, {@link CountMinSketch}>>. 
- perKey() - Static method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles
- 
- perKey() - Static method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct
-  
- PerKey() - Constructor for class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.PerKey
-  
- perKey() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Extract
- 
Returns a PTransformthat takes an inputPCollection<KV<K, byte[]>>of (key,
 HLL++ sketch) pairs and returns aPCollection<KV<K, Long>>of (key, estimated count
 of distinct elements extracted from each sketch).
 
- perKey() - Method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Init.Builder
- 
Returns a Combine.PerKeyPTransformthat takes an inputPCollection<KV<K, InputT>>and returns aPCollection<KV<K, byte[]>>which consists
 of the per-key HLL++ sketch computed from the values matching each key in the inputPCollection.
 
- perKey() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.MergePartial
- 
Returns a Combine.PerKeyPTransformthat takes an inputPCollection<KV<K, byte[]>>of (key, HLL++ sketch) pairs and returns aPCollection<KV<K, byte[]>>of (key, new sketch merged from the input sketches under the
 key).
 
- perKey(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles
- 
Returns a PTransformthat takes aPCollection<KV<K, V>>and returns aPCollection<KV<K, List<V>>>that contains an output element mapping each distinct key in the
 inputPCollectionto aListof the approximateN-tiles of the values
 associated with that key in the inputPCollection.
 
- perKey(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles
- 
- perKey(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateUnique
- 
Deprecated. Returns a PTransformthat takes aPCollection<KV<K, V>>and returns aPCollection<KV<K, Long>>that contains an output element mapping each distinct key in the
 inputPCollectionto an estimate of the number of distinct values associated with that
 key in the inputPCollection.
 
- perKey(double) - Static method in class org.apache.beam.sdk.transforms.ApproximateUnique
- 
Deprecated. 
- PerKey(int) - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique.PerKey
- 
Deprecated. 
- PerKey(double) - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique.PerKey
- 
Deprecated. 
- perKey(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
- 
Returns a  Combine.PerKeyPTransform that first groups its input  PCollection of  KVs by keys and windows, then invokes the given function on each of the
 values lists to produce a combined value, and then returns a  PCollection of  KVs
 mapping each distinct key to its combined value for each window. 
- perKey(SerializableBiFunction<V, V, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
- 
Returns a  Combine.PerKeyPTransform that first groups its input  PCollection of  KVs by keys and windows, then invokes the given function on each of the
 values lists to produce a combined value, and then returns a  PCollection of  KVs
 mapping each distinct key to its combined value for each window. 
- perKey(CombineFnBase.GlobalCombineFn<? super InputT, ?, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Combine
- 
Returns a  Combine.PerKeyPTransform that first groups its input  PCollection of  KVs by keys and windows, then invokes the given function on each of the
 values lists to produce a combined value, and then returns a  PCollection of  KVs
 mapping each distinct key to its combined value for each window. 
- perKey() - Static method in class org.apache.beam.sdk.transforms.Count
- 
Returns a  PTransform that counts the number of elements associated with each key of its
 input  PCollection. 
- perKey() - Static method in class org.apache.beam.sdk.transforms.Latest
- 
Returns a  PTransform that takes as input a  PCollection<KV<K, V>> and returns a
  PCollection<KV<K, V>> whose contents is the latest element per-key according to its
 event time. 
- perKey() - Static method in class org.apache.beam.sdk.transforms.Max
- 
Returns a PTransformthat takes an inputPCollection<KV<K, T>>and returns aPCollection<KV<K, T>>that contains an output element mapping each distinct key in the
 inputPCollectionto the maximum according to the natural ordering ofTof the
 values associated with that key in the inputPCollection.
 
- perKey(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Max
- 
Returns a PTransformthat takes an inputPCollection<KV<K, T>>and returns aPCollection<KV<K, T>>that contains one output element per key mapping each to the
 maximum of the values associated with that key in the inputPCollection.
 
- perKey() - Static method in class org.apache.beam.sdk.transforms.Mean
- 
Returns a PTransformthat takes an inputPCollection<KV<K, N>>and returns aPCollection<KV<K, Double>>that contains an output element mapping each distinct key in
 the inputPCollectionto the mean of the values associated with that key in the inputPCollection.
 
- perKey() - Static method in class org.apache.beam.sdk.transforms.Min
- 
Returns a PTransformthat takes an inputPCollection<KV<K, T>>and returns aPCollection<KV<K, T>>that contains an output element mapping each distinct key in the
 inputPCollectionto the minimum according to the natural ordering ofTof the
 values associated with that key in the inputPCollection.
 
- perKey(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Min
- 
Returns a PTransformthat takes an inputPCollection<KV<K, T>>and returns aPCollection<KV<K, T>>that contains one output element per key mapping each to the
 minimum of the values associated with that key in the inputPCollection.
 
- perKey(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Top
- 
Returns a PTransformthat takes an inputPCollection<KV<K, V>>and returns aPCollection<KV<K, List<V>>>that contains an output element mapping each distinct key
 in the inputPCollectionto the largestcountvalues associated with that key
 in the inputPCollection<KV<K, V>>, in decreasing order, sorted using the givenComparator<V>.
 
- PerKeyDigest() - Constructor for class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.PerKeyDigest
-  
- PerKeyDistinct() - Constructor for class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.PerKeyDistinct
-  
- PerKeySketch() - Constructor for class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.PerKeySketch
-  
- pin() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StreamAppendClient
- 
Pin this object. 
- PInput - Interface in org.apache.beam.sdk.values
- 
The interface for things that might be input to a  PTransform. 
- pipeline() - Method in class org.apache.beam.runners.jobsubmission.JobPreparation
-  
- Pipeline - Class in org.apache.beam.sdk
- 
- Pipeline(PipelineOptions) - Constructor for class org.apache.beam.sdk.Pipeline
-  
- Pipeline.PipelineExecutionException - Exception in org.apache.beam.sdk
- 
Thrown during execution of a  Pipeline, whenever user code within that  Pipeline
 throws an exception. 
- Pipeline.PipelineVisitor - Interface in org.apache.beam.sdk
- 
For internal use only; no backwards-compatibility guarantees. 
- Pipeline.PipelineVisitor.CompositeBehavior - Enum in org.apache.beam.sdk
- 
Control enum for indicating whether or not a traversal should process the contents of a
 composite transform or not. 
- Pipeline.PipelineVisitor.Defaults - Class in org.apache.beam.sdk
- 
- PIPELINE_PROTO_CODER_ID - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- PIPELINED - Static variable in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- PipelineExecutionException(Throwable) - Constructor for exception org.apache.beam.sdk.Pipeline.PipelineExecutionException
- 
- PipelineMessageReceiver - Interface in org.apache.beam.runners.local
- 
Handles failures in the form of exceptions. 
- pipelineOptions() - Method in class org.apache.beam.runners.fnexecution.provisioning.JobInfo
-  
- PipelineOptions - Interface in org.apache.beam.sdk.options
- 
PipelineOptions are used to configure Pipelines. 
- PipelineOptions.AtomicLongFactory - Class in org.apache.beam.sdk.options
- 
DefaultValueFactory which supplies an ID that is guaranteed to be unique within the
 given process. 
- PipelineOptions.CheckEnabled - Enum in org.apache.beam.sdk.options
- 
Enumeration of the possible states for a given check. 
- PipelineOptions.DirectRunner - Class in org.apache.beam.sdk.options
- 
A  DefaultValueFactory that obtains the class of the  DirectRunner if it exists
 on the classpath, and throws an exception otherwise. 
- PipelineOptions.JobNameFactory - Class in org.apache.beam.sdk.options
- 
- PipelineOptions.UserAgentFactory - Class in org.apache.beam.sdk.options
- 
Returns a user agent string constructed from ReleaseInfo.getName()andReleaseInfo.getVersion(), in the format[name]/[version].
 
- PipelineOptionsFactory - Class in org.apache.beam.sdk.options
- 
- PipelineOptionsFactory() - Constructor for class org.apache.beam.sdk.options.PipelineOptionsFactory
-  
- PipelineOptionsFactory.Builder - Class in org.apache.beam.sdk.options
- 
- PipelineOptionsRegistrar - Interface in org.apache.beam.sdk.options
- 
PipelineOptions creators have the ability to automatically have their  PipelineOptions registered with this SDK by creating a  ServiceLoader entry and a
 concrete implementation of this interface. 
- PipelineOptionsValidator - Class in org.apache.beam.sdk.options
- 
- PipelineOptionsValidator() - Constructor for class org.apache.beam.sdk.options.PipelineOptionsValidator
-  
- PipelineResult - Interface in org.apache.beam.sdk
- 
- PipelineResult.State - Enum in org.apache.beam.sdk
- 
Possible job states, for both completed and ongoing jobs. 
- PipelineRunner<ResultT extends PipelineResult> - Class in org.apache.beam.sdk
- 
- PipelineRunner() - Constructor for class org.apache.beam.sdk.PipelineRunner
-  
- PipelineTranslator - Class in org.apache.beam.runners.spark.structuredstreaming.translation
- 
The pipeline translator translates a Beam  Pipeline into a Spark correspondence, that can
 then be evaluated. 
- PipelineTranslator() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
-  
- PipelineTranslator.TranslationState - Interface in org.apache.beam.runners.spark.structuredstreaming.translation
- 
Shared, mutable state during the translation of a pipeline and omitted afterwards. 
- PipelineTranslator.UnresolvedTranslation<InT,T> - Interface in org.apache.beam.runners.spark.structuredstreaming.translation
- 
Unresolved translation, allowing to optimize the generated Spark DAG. 
- PipelineTranslatorBatch - Class in org.apache.beam.runners.spark.structuredstreaming.translation.batch
- 
- PipelineTranslatorBatch() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.batch.PipelineTranslatorBatch
-  
- PipelineTranslatorUtils - Class in org.apache.beam.runners.fnexecution.translation
- 
Utilities for pipeline translation. 
- placementId() - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIPredict
-  
- Plugin<K,V> - Class in org.apache.beam.sdk.io.cdap
- 
Class wrapper for a CDAP plugin. 
- Plugin() - Constructor for class org.apache.beam.sdk.io.cdap.Plugin
-  
- Plugin.Builder<K,V> - Class in org.apache.beam.sdk.io.cdap
- 
- pluginConfig - Variable in class org.apache.beam.sdk.io.cdap.Plugin
-  
- PluginConfigInstantiationUtils - Class in org.apache.beam.sdk.io.cdap
- 
Class for getting any filled  PluginConfig configuration object. 
- PluginConfigInstantiationUtils() - Constructor for class org.apache.beam.sdk.io.cdap.PluginConfigInstantiationUtils
-  
- PluginConstants - Class in org.apache.beam.sdk.io.cdap
- 
Class for CDAP plugin constants. 
- PluginConstants() - Constructor for class org.apache.beam.sdk.io.cdap.PluginConstants
-  
- PluginConstants.Format - Enum in org.apache.beam.sdk.io.cdap
- 
Format types. 
- PluginConstants.FormatProvider - Enum in org.apache.beam.sdk.io.cdap
- 
Format provider types. 
- PluginConstants.Hadoop - Enum in org.apache.beam.sdk.io.cdap
- 
Hadoop types. 
- PluginConstants.PluginType - Enum in org.apache.beam.sdk.io.cdap
- 
Plugin types. 
- PLUS - Static variable in class org.apache.beam.sdk.extensions.sql.impl.cep.Quantifier
-  
- plus(RelOptCost) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
-  
- plus(NodeStats) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
-  
- PLUS_RELUCTANT - Static variable in class org.apache.beam.sdk.extensions.sql.impl.cep.Quantifier
-  
- plusDelayOf(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
- 
Adds some delay to the original target time. 
- poisonInstructionId(String) - Method in class org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer
- 
Poisons an instruction id. 
- POJOUtils - Class in org.apache.beam.sdk.schemas.utils
- 
A set of utilities to generate getter and setter classes for POJOs. 
- POJOUtils() - Constructor for class org.apache.beam.sdk.schemas.utils.POJOUtils
-  
- PollFn() - Constructor for class org.apache.beam.sdk.transforms.Watch.Growth.PollFn
-  
- pollFor(Duration) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery.PollingAssertion
-  
- pollJob(JobReference, int) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.JobService
- 
Waits for the job is Done, and returns the job. 
- pollJob(JobReference, int) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
-  
- pollOperation(Operation, Long) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Poll operation. 
- pollOperation(Operation, Long) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- pooledClientFactory(BuilderT) - Static method in class org.apache.beam.sdk.io.aws2.common.ObjectPool
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Parse
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ParseAll
- 
Deprecated. 
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ParseFiles
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadAll
- 
Deprecated. 
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadFiles
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.ApproximateDistinctFn
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.CountMinSketchFn
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.TDigestQuantilesFn
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.sql.TypedCombineFnDelegate
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.Globally
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.PerKey
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.amqp.AmqpIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.BoundedReadFromUnboundedSource
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.CompressedSource
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.ReadFiles
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.csv.CsvIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
- 
Populates the display data. 
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileBasedSink.FilenamePolicy
- 
Populates the display data. 
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileBasedSink
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileBasedSource
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileIO.Match
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileIO.MatchAll
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileIO.ReadMatches
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageTableSource
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.WriteWithResults
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.datastore.RampupThrottlingFn
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write.PubsubBoundedWriter
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.WriteGrouped
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.GenerateSequence
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.WriteRowMutations
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceProviderFromDataSourceConfiguration
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.PoolableDataSourceProvider
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadWithPartitions
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.TypedWithoutMetadata
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverProviderFromDriverConfiguration
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.ReadAll
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Parse
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.ParseFiles
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.ReadFiles
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.Read.Bounded
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.Read.Unbounded
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
- 
Populate the display data with connectionConfiguration details. 
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
-  
- populateDisplayData(SingleStoreIO.DataSourceConfiguration, DisplayData.Builder) - Static method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.DataSourceConfiguration
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.ReadWithPartitions
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceProviderFromDataSourceConfiguration
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.Source
- 
Register display data for the given transform or component. 
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TextIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
- 
Deprecated. 
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TextIO.ReadFiles
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TextIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.ReadFiles
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Parse
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.tika.TikaIO.ParseFiles
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.WriteFiles
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.Globally
- 
Deprecated. 
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.PerKey
- 
Deprecated. 
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.Globally
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.GloballyAsSingletonView
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.PerKeyWithHotKeyFanout
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
-  
- populateDisplayData(DisplayData.Builder) - Method in interface org.apache.beam.sdk.transforms.display.HasDisplayData
- 
Register display data for the given transform or component. 
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.DoFn
- 
Register display data for the given transform or component. 
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Filter
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.FlatMapElements.FlatMapWithFailures
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.FlatMapElements
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.GroupByKey
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.InferableFunction
- 
Register display data for the given transform or component. 
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.MapElements.MapWithFailures
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.MapElements
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
- 
Register display data for the given transform or component. 
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Partition
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.PTransform
- 
Register display data for the given transform or component. 
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.Window.Assign
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.Window
-  
- populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
- 
Register display data for the given transform or component. 
- PortableBigQueryDestinations - Class in org.apache.beam.sdk.io.gcp.bigquery.providers
-  
- PortableBigQueryDestinations(Schema, BigQueryWriteConfiguration) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.providers.PortableBigQueryDestinations
-  
- portableMetrics() - Method in class org.apache.beam.runners.flink.FlinkPortableRunnerResult
-  
- portableMetrics() - Method in interface org.apache.beam.runners.jobsubmission.PortablePipelineResult
- 
Returns the object to access monitoring infos from the pipeline. 
- PortableMetrics - Class in org.apache.beam.runners.portability
-  
- PortablePipelineJarCreator - Class in org.apache.beam.runners.jobsubmission
- 
- PortablePipelineJarCreator(Class) - Constructor for class org.apache.beam.runners.jobsubmission.PortablePipelineJarCreator
-  
- PortablePipelineJarUtils - Class in org.apache.beam.runners.jobsubmission
- 
Contains common code for writing and reading portable pipeline jars. 
- PortablePipelineJarUtils() - Constructor for class org.apache.beam.runners.jobsubmission.PortablePipelineJarUtils
-  
- PortablePipelineOptions - Interface in org.apache.beam.sdk.options
- 
Pipeline options common to all portable runners. 
- PortablePipelineResult - Interface in org.apache.beam.runners.jobsubmission
- 
- PortablePipelineRunner - Interface in org.apache.beam.runners.jobsubmission
- 
Runs a portable Beam pipeline on some execution engine. 
- PortableRunner - Class in org.apache.beam.runners.portability
- 
- PortableRunnerRegistrar - Class in org.apache.beam.runners.portability
- 
Registrar for the portable runner. 
- PortableRunnerRegistrar() - Constructor for class org.apache.beam.runners.portability.PortableRunnerRegistrar
-  
- position() - Method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingSeekableByteChannel
-  
- position(long) - Method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingSeekableByteChannel
-  
- positional() - Method in class org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters
-  
- PositionAwareCombineFn() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamBuiltinAnalyticFunctions.PositionAwareCombineFn
-  
- POSTGRES - Static variable in class org.apache.beam.sdk.io.jdbc.JdbcUtil
-  
- PostProcessingMetricsDoFn - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn
- 
- PostProcessingMetricsDoFn(ChangeStreamMetrics) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.PostProcessingMetricsDoFn
-  
- POutput - Interface in org.apache.beam.sdk.values
- 
The interface for things that might be output from a  PTransform. 
- PRE_DEFINED_WINDOW_FUNCTIONS - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCatalog
-  
- precisionForRelativeError(double) - Static method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct
- 
Computes the precision based on the desired relative error. 
- predictAll() - Static method in class org.apache.beam.sdk.extensions.ml.RecommendationAIIO
-  
- prefetch() - Method in class org.apache.beam.sdk.fn.stream.DataStreams.DataStreamDecoder
-  
- prefetch() - Method in interface org.apache.beam.sdk.fn.stream.PrefetchableIterable
- 
Ensures that the next iterator returned has been prefetched. 
- prefetch() - Method in class org.apache.beam.sdk.fn.stream.PrefetchableIterables.Default
-  
- prefetch() - Method in interface org.apache.beam.sdk.fn.stream.PrefetchableIterator
- 
- PrefetchableIterable<T> - Interface in org.apache.beam.sdk.fn.stream
- 
- PrefetchableIterables - Class in org.apache.beam.sdk.fn.stream
- 
This class contains static utility functions that operate on or return objects of type  PrefetchableIterable. 
- PrefetchableIterables() - Constructor for class org.apache.beam.sdk.fn.stream.PrefetchableIterables
-  
- PrefetchableIterables.Default<T> - Class in org.apache.beam.sdk.fn.stream
- 
- PrefetchableIterator<T> - Interface in org.apache.beam.sdk.fn.stream
- 
Iteratorthat supports prefetching the next set of records.
 
- PrefetchableIterators - Class in org.apache.beam.sdk.fn.stream
-  
- PrefetchableIterators() - Constructor for class org.apache.beam.sdk.fn.stream.PrefetchableIterators
-  
- prefetchOnMerge(MergingStateAccessor<K, W>) - Method in class org.apache.beam.runners.twister2.translators.functions.internal.SystemReduceFnBuffering
-  
- prefetchOnTrigger(StateAccessor<K>) - Method in class org.apache.beam.runners.twister2.translators.functions.internal.SystemReduceFnBuffering
-  
- prepare(JobApi.PrepareJobRequest, StreamObserver<JobApi.PrepareJobResponse>) - Method in class org.apache.beam.runners.jobsubmission.InMemoryJobService
-  
- prepare(JobApi.PrepareJobRequest, StreamObserver<JobApi.PrepareJobResponse>) - Method in class org.apache.beam.runners.portability.testing.TestJobService
-  
- prepare(TSetContext) - Method in class org.apache.beam.runners.twister2.translation.wrappers.Twister2BoundedSource
-  
- prepare(TSetContext) - Method in class org.apache.beam.runners.twister2.translators.functions.AssignWindowsFunction
-  
- prepare(TSetContext) - Method in class org.apache.beam.runners.twister2.translators.functions.ByteToElemFunction
-  
- prepare(TSetContext) - Method in class org.apache.beam.runners.twister2.translators.functions.ByteToWindowFunction
-  
- prepare(TSetContext) - Method in class org.apache.beam.runners.twister2.translators.functions.ByteToWindowFunctionPrimitive
-  
- prepare(TSetContext) - Method in class org.apache.beam.runners.twister2.translators.functions.DoFnFunction
-  
- prepare(TSetContext) - Method in class org.apache.beam.runners.twister2.translators.functions.ElemToBytesFunction
-  
- prepare(TSetContext) - Method in class org.apache.beam.runners.twister2.translators.functions.GroupByWindowFunction
-  
- prepare(TSetContext) - Method in class org.apache.beam.runners.twister2.translators.functions.MapToTupleFunction
-  
- prepare(TSetContext) - Method in class org.apache.beam.runners.twister2.translators.functions.OutputTagFilter
-  
- prepare(TSetContext) - Method in class org.apache.beam.runners.twister2.translators.functions.Twister2SinkFunction
-  
- prepareCall(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- prepareCall(String, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- prepareCall(String, int, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- prepareFilesToStage(SparkCommonPipelineOptions) - Static method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
- 
Classpath contains non jar files (eg. 
- prepareForProcessing() - Method in class org.apache.beam.sdk.transforms.DoFn
- 
- prepareForTranslation(RunnerApi.Pipeline) - Method in interface org.apache.beam.runners.flink.FlinkPortablePipelineTranslator
-  
- preparePrivateKey(String, String) - Static method in class org.apache.beam.sdk.io.snowflake.KeyPairUtils
-  
- PreparePubsubWriteDoFn<InputT> - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- prepareRun() - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
- prepareStatement(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- prepareStatement(String, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- prepareStatement(String, int, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- prepareStatement(String, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- prepareStatement(String, int[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- prepareStatement(String, String[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- prepareWrite(WritableByteChannel) - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
- 
Called with the channel that a subclass will write its header, footer, and values to. 
- PrepareWrite<InputT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
- PrepareWrite(DynamicDestinations<InputT, DestinationT>, SerializableFunction<InputT, OutputT>) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.PrepareWrite
-  
- PRESERVES_KEYS - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- PRESET_HELPERS - Static variable in class org.apache.beam.sdk.io.jdbc.JdbcUtil
-  
- previous(Timestamp) - Static method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampUtils
-  
- primary() - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
-  
- PrimitiveParDoSingleFactory<InputT,OutputT> - Class in org.apache.beam.runners.dataflow
- 
- PrimitiveParDoSingleFactory() - Constructor for class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory
-  
- PrimitiveParDoSingleFactory.ParDoSingle<InputT,OutputT> - Class in org.apache.beam.runners.dataflow
- 
A single-output primitive  ParDo. 
- PrimitiveParDoSingleFactory.PayloadTranslator - Class in org.apache.beam.runners.dataflow
- 
- PrimitiveParDoSingleFactory.Registrar - Class in org.apache.beam.runners.dataflow
- 
- printHelp(PrintStream) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
- 
Outputs the set of registered options with the PipelineOptionsFactory with a description for
 each one if available to the output stream. 
- printHelp(PrintStream, Class<? extends PipelineOptions>) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
- 
Outputs the set of options available to be set for the passed in  PipelineOptions
 interface. 
- PrismPipelineOptions - Interface in org.apache.beam.runners.prism
- 
- PrismRegistrar - Class in org.apache.beam.runners.prism
- 
- PrismRegistrar.Options - Class in org.apache.beam.runners.prism
- 
- PrismRegistrar.Runner - Class in org.apache.beam.runners.prism
- 
- PrismRunner - Class in org.apache.beam.runners.prism
- 
- PrismRunner(PrismPipelineOptions) - Constructor for class org.apache.beam.runners.prism.PrismRunner
-  
- process(Map<String, String>, RestrictionTracker<KafkaSourceConsumerFn.OffsetHolder, Map<String, Object>>, DoFn.OutputReceiver<T>) - Method in class org.apache.beam.io.debezium.KafkaSourceConsumerFn
- 
Process the retrieved element and format it for output. 
- process(List<JobMessage>) - Method in interface org.apache.beam.runners.dataflow.util.MonitoringUtil.JobMessagesHandler
- 
Process the rows. 
- process(List<JobMessage>) - Method in class org.apache.beam.runners.dataflow.util.MonitoringUtil.LoggingHandler
-  
- process(KV<Row, ValueT>, Instant, TimerMap, TimerMap, ValueState<SortedMap<Instant, TimestampedValue<ValueT>>>, DoFn.OutputReceiver<ValueT>) - Method in class org.apache.beam.sdk.extensions.timeseries.FillGaps.FillGapsDoFn
-  
- process(PipelineOptions, KV<String, StorageApiFlushAndFinalizeDoFn.Operation>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiFlushAndFinalizeDoFn
-  
- process(InputT, Instant, BoundedWindow, PaneInfo, DoFn.MultiOutputReceiver) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PreparePubsubWriteDoFn
-  
- process(SequencedMessage, DoFn.MultiOutputReceiver) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.ErrorFn
-  
- process(Row, DoFn.MultiOutputReceiver) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.ErrorCounterFn
-  
- process(DataChangeRecord, DoFn.MultiOutputReceiver) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.DataChangeRecordToRow
-  
- process(byte[], DoFn.MultiOutputReceiver) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformProvider.ErrorFn
-  
- process(DoFn<KV<FileIO.ReadableFile, OffsetRange>, T>.ProcessContext) - Method in class org.apache.beam.sdk.io.ReadAllViaFileBasedSourceTransform.AbstractReadFileRangesFn
-  
- process(DoFn<FileIO.ReadableFile, KV<FileIO.ReadableFile, OffsetRange>>.ProcessContext) - Method in class org.apache.beam.sdk.io.ReadAllViaFileBasedSourceTransform.SplitIntoRangesFn
-  
- process(DoFn<InputT, KV<InputT, List<TimestampedValue<OutputT>>>>.ProcessContext, RestrictionTracker<Watch.GrowthState, KV<Watch.Growth.PollResult<OutputT>, TerminationStateT>>, ManualWatermarkEstimator<Instant>) - Method in class org.apache.beam.sdk.transforms.Watch.WatchGrowthFn
-  
- processArrayOfNestedStringField(RowBundles.ArrayOfNestedStringBundle, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.GetterBasedSchemaProviderBenchmark
-  
- processArrayOfStringField(RowBundles.ArrayOfStringBundle, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.GetterBasedSchemaProviderBenchmark
-  
- processBundle(Iterable<? extends InputT>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- processBundle(InputT...) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- ProcessBundleDescriptors - Class in org.apache.beam.runners.fnexecution.control
- 
Utility methods for creating BeamFnApi.ProcessBundleDescriptorinstances.
 
- ProcessBundleDescriptors() - Constructor for class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors
-  
- ProcessBundleDescriptors.BagUserStateSpec<K,V,W extends BoundedWindow> - Class in org.apache.beam.runners.fnexecution.control
- 
A container type storing references to the key, value, and window  Coder used when
 handling bag user state requests. 
- ProcessBundleDescriptors.ExecutableProcessBundleDescriptor - Class in org.apache.beam.runners.fnexecution.control
-  
- ProcessBundleDescriptors.SideInputSpec<T,W extends BoundedWindow> - Class in org.apache.beam.runners.fnexecution.control
- 
A container type storing references to the value, and window  Coder used when handling
 side input state requests. 
- ProcessBundleDescriptors.TimerSpec<K,V,W extends BoundedWindow> - Class in org.apache.beam.runners.fnexecution.control
- 
A container type storing references to the key, timer and payload coders and the remote input
 destination used when handling timer requests. 
- processByteBufferField(RowBundles.ByteBufferBundle, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.GetterBasedSchemaProviderBenchmark
-  
- processBytesField(RowBundles.BytesBundle, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.GetterBasedSchemaProviderBenchmark
-  
- ProcessContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.ProcessContext
-  
- ProcessContinuation() - Constructor for class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
-  
- processDateTimeField(RowBundles.DateTimeBundle, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.GetterBasedSchemaProviderBenchmark
-  
- processElement(DoFn<KV<K, Iterable<KV<Instant, WindowedValue<KV<K, V>>>>>, OutputT>.ProcessContext, BoundedWindow) - Method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides.BatchStatefulDoFn
-  
- processElement(DoFn<Iterable<T>, T>.ProcessContext, BoundedWindow) - Method in class org.apache.beam.runners.dataflow.DataflowRunner.StreamingPCollectionViewWriterFn
-  
- processElement(WindowedValue<InputT>) - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
-  
- processElement(Row, DoFn.OutputReceiver<Row>) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSetOperatorsTransforms.SetOperatorFilteringDoFn
-  
- processElement(DoFn<Row, Void>.ProcessContext) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSqlOutputToConsoleFn
-  
- processElement(DoFn<KV<DestinationT, ElementT>, KV<DestinationT, StorageApiWritePayload>>.ProcessContext, PipelineOptions, KV<DestinationT, ElementT>, Instant, DoFn.MultiOutputReceiver) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiConvertMessages.ConvertMessagesDoFn
-  
- processElement(Iterable<KV<DestinationT, WriteTables.Result>>, DoFn<Iterable<KV<DestinationT, WriteTables.Result>>, Iterable<KV<TableDestination, WriteTables.Result>>>.ProcessContext, BoundedWindow) - Method in class org.apache.beam.sdk.io.gcp.bigquery.UpdateSchemaDestination
-  
- processElement(InitialPipelineState, RestrictionTracker<OffsetRange, Long>, DoFn.OutputReceiver<PartitionRecord>, ManualWatermarkEstimator<Instant>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.DetectNewPartitionsDoFn
-  
- processElement(KV<ByteString, ChangeStreamRecord>, DoFn.OutputReceiver<KV<ByteString, ChangeStreamMutation>>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.FilterForMutationDoFn
-  
- processElement(DoFn.OutputReceiver<InitialPipelineState>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.InitializeDoFn
-  
- processElement(PartitionRecord, RestrictionTracker<StreamProgress, StreamProgress>, DoFn.OutputReceiver<KV<ByteString, ChangeStreamRecord>>, ManualWatermarkEstimator<Instant>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.ReadChangeStreamPartitionDoFn
-  
- processElement(DoFn<T, T>.ProcessContext) - Method in class org.apache.beam.sdk.io.gcp.datastore.RampupThrottlingFn
- 
Emit only as many elements as the exponentially increasing budget allows. 
- processElement(DoFn<HL7v2ReadParameter, HL7v2ReadResponse>.ProcessContext) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.HL7v2Read.FetchHL7v2Message.HL7v2MessageGetFn
- 
Process element. 
- processElement(DoFn<String, HL7v2Message>.ProcessContext) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.Read.FetchHL7v2Message.HL7v2MessageGetFn
- 
Process element. 
- processElement(PubsubMessage, Instant) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write.PubsubBoundedWriter
-  
- processElement(Row, DoFn.MultiOutputReceiver) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformProvider.ErrorFn
-  
- processElement(PubSubMessage) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.PubsubLiteSink
-  
- processElement(PubSubMessage, DoFn.OutputReceiver<PubSubMessage>) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.SetUuidFromPubSubMessage.SetUuidFn
-  
- processElement(DoFn.OutputReceiver<Void>) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.CleanUpReadChangeStreamDoFn
-  
- processElement(RestrictionTracker<TimestampRange, Timestamp>, DoFn.OutputReceiver<PartitionMetadata>, ManualWatermarkEstimator<Instant>) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.DetectNewPartitionsDoFn
- 
- processElement(DoFn.OutputReceiver<PartitionMetadata>) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.InitializeDoFn
-  
- processElement(DataChangeRecord, DoFn.OutputReceiver<DataChangeRecord>) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.PostProcessingMetricsDoFn
- 
Stage to measure a data records latencies and metrics. 
- processElement(PartitionMetadata, RestrictionTracker<TimestampRange, Timestamp>, DoFn.OutputReceiver<DataChangeRecord>, ManualWatermarkEstimator<Instant>, DoFn.BundleFinalizer) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.ReadChangeStreamPartitionDoFn
- 
Performs a change stream query for a given partition. 
- processElement(DoFn<Void, SpannerSchema>.ProcessContext) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadSpannerSchema
-  
- processElement(PulsarSourceDescriptor, RestrictionTracker<OffsetRange, Long>, WatermarkEstimator, DoFn.OutputReceiver<PulsarMessage>) - Method in class org.apache.beam.sdk.io.pulsar.ReadFromPulsarDoFn
-  
- processElement(byte[]) - Method in class org.apache.beam.sdk.io.pulsar.WriteToPulsarDoFn
-  
- processElement(DoFn<Object, Object>.ProcessContext) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read.CleanTmpFilesFromGcsFn
-  
- processElement(DoFn<String, String[]>.ProcessContext) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read.MapCsvToStringArrayFn
-  
- processElement(Solace.Record, DoFn.OutputReceiver<KV<Integer, Solace.Record>>) - Method in class org.apache.beam.sdk.io.solace.write.AddShardKeyDoFn
-  
- processElement(Solace.Record, DoFn.OutputReceiver<Solace.PublishResult>) - Method in class org.apache.beam.sdk.io.solace.write.RecordToPublishResultDoFn
-  
- processElement(KV<Integer, Solace.Record>, Timer, Instant) - Method in class org.apache.beam.sdk.io.solace.write.UnboundedBatchedSolaceWriter
-  
- processElement(KV<Integer, Solace.Record>, Instant, ValueState<Integer>) - Method in class org.apache.beam.sdk.io.solace.write.UnboundedStreamingSolaceWriter
-  
- processElement(InputT) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- processElement(ErrorT, DoFn.OutputReceiver<ErrorT>) - Method in class org.apache.beam.sdk.transforms.errorhandling.ErrorHandler.PTransformErrorHandler.WriteErrorMetrics.CountErrors
-  
- processElement(String, DoFn.MultiOutputReceiver) - Method in class org.apache.beam.sdk.transforms.JsonToRow.JsonToRowWithErrFn.ParseWithError
-  
- processElement(T, DoFn.OutputReceiver<KV<Integer, T>>) - Method in class org.apache.beam.sdk.transforms.Reshuffle.AssignShardFn
-  
- processElement(DoFn<T, KV<Long, PCollectionViews.ValueOrMetadata<T, OffsetRange>>>.ProcessContext, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.View.ToListViewDoFn
-  
- processElement(DoFn<ValueWithRecordId<T>, T>.ProcessContext) - Method in class org.apache.beam.sdk.values.ValueWithRecordId.StripIdsDoFn
-  
- processElementWithRunner(DoFnRunner<KV<?, ?>, OutputT>, WindowedValue<KV<?, ?>>) - Method in class org.apache.beam.runners.jet.processors.StatefulParDoP
-  
- ProcessEnvironment - Class in org.apache.beam.runners.fnexecution.environment
- 
Environment for process-based execution. 
- ProcessEnvironmentFactory - Class in org.apache.beam.runners.fnexecution.environment
- 
- ProcessEnvironmentFactory.Provider - Class in org.apache.beam.runners.fnexecution.environment
- 
Provider of ProcessEnvironmentFactory. 
- ProcessFunction<InputT,OutputT> - Interface in org.apache.beam.sdk.transforms
- 
A function that computes an output value of type OutputTfrom an input value of typeInputTand isSerializable.
 
- PROCESSING_DELAY_FROM_COMMIT_TIMESTAMP - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.ChangeStreamMetrics
- 
Distribution for measuring processing delay from commit timestamp. 
- processingStatuses() - Method in class org.apache.beam.sdk.extensions.ordered.OrderedEventProcessorResult
-  
- ProcessingTimeEvent() - Constructor for class org.apache.beam.sdk.testing.TestStream.ProcessingTimeEvent
-  
- ProcessingTimePolicy() - Constructor for class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.ProcessingTimePolicy
-  
- ProcessingTimeWatermarkPolicy() - Constructor for class org.apache.beam.sdk.io.aws2.kinesis.WatermarkPolicyFactory.ProcessingTimeWatermarkPolicy
-  
- processIntField(RowBundles.IntBundle, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.GetterBasedSchemaProviderBenchmark
-  
- ProcessManager - Class in org.apache.beam.runners.fnexecution.environment
- 
A simple process manager which forks processes and kills them if necessary. 
- ProcessManager.RunningProcess - Class in org.apache.beam.runners.fnexecution.environment
-  
- processMapOfIntField(RowBundles.MapOfIntBundle, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.GetterBasedSchemaProviderBenchmark
-  
- processMapOfNestedIntField(RowBundles.MapOfNestedIntBundle, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.GetterBasedSchemaProviderBenchmark
-  
- processNestedBytesField(RowBundles.NestedBytesBundle, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.GetterBasedSchemaProviderBenchmark
-  
- processNestedIntField(RowBundles.NestedIntBundle, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.GetterBasedSchemaProviderBenchmark
-  
- processNewPartition(NewPartition, DoFn.OutputReceiver<PartitionRecord>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.ProcessNewPartitionsAction
- 
Process a single new partition. 
- processNewPartitionsAction(ChangeStreamMetrics, MetadataTableDao, Instant) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.ActionFactory
-  
- ProcessNewPartitionsAction - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.action
-  
- ProcessNewPartitionsAction(ChangeStreamMetrics, MetadataTableDao, Instant) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.ProcessNewPartitionsAction
-  
- processNewRow(Row) - Method in class org.apache.beam.sdk.extensions.sql.impl.nfa.NFA
-  
- processRows(Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.RowBundle
- 
Runs benchmark iteration on a bundle of rows. 
- processStringBuilderField(RowBundles.StringBuilderBundle, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.GetterBasedSchemaProviderBenchmark
-  
- processStringField(RowBundles.StringBundle, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.GetterBasedSchemaProviderBenchmark
-  
- processTimestampedElement(TimestampedValue<InputT>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- processValue(ReduceFn<K, T, Iterable<T>, W>.ProcessValueContext) - Method in class org.apache.beam.runners.twister2.translators.functions.internal.SystemReduceFnBuffering
-  
- processWindowedElement(InputT, Instant, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- produceResult() - Method in interface org.apache.beam.sdk.extensions.ordered.MutableState
- 
This method is called after each state mutation. 
- ProducerRecordCoder<K,V> - Class in org.apache.beam.sdk.io.kafka
- 
- ProducerRecordCoder(Coder<K>, Coder<V>) - Constructor for class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
-  
- program - Variable in class org.apache.beam.sdk.extensions.sql.impl.rel.CalcRelSplitter
-  
- Progress() - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker.Progress
-  
- project - Variable in class org.apache.beam.sdk.io.gcp.healthcare.WebPathParser.DicomWebPath
-  
- PROJECT_ID_REGEXP - Static variable in class org.apache.beam.runners.dataflow.DataflowRunner
- 
Project IDs must contain lowercase letters, digits, or dashes. 
- projectId() - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAICreateCatalogItem
-  
- projectId() - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportCatalogItems
-  
- projectId() - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportUserEvents
-  
- projectId() - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIPredict
-  
- projectId() - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIWriteUserEvent
-  
- projectId - Variable in class org.apache.beam.sdk.io.gcp.datastore.DataStoreV1SchemaIOProvider.DataStoreV1SchemaIO
-  
- ProjectionConsumer - Interface in org.apache.beam.sdk.schemas
- 
- ProjectionProducer<T> - Interface in org.apache.beam.sdk.schemas
- 
- projectPathFromId(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
-  
- projectPathFromPath(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
-  
- ProjectSupport - Enum in org.apache.beam.sdk.extensions.sql.meta
-  
- properties(ObjectNode) - Method in class org.apache.beam.sdk.extensions.sql.meta.Table.Builder
-  
- properties() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
-  
- PROPERTY_BEAM_TEST_PIPELINE_OPTIONS - Static variable in class org.apache.beam.sdk.testing.TestPipeline
- 
- propertyName() - Method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
- 
Returns the property name associated with this provider. 
- propertyName() - Method in class org.apache.beam.sdk.options.ValueProvider.RuntimeValueProvider
- 
Returns the property name that corresponds to this provider. 
- PropertyNames - Class in org.apache.beam.runners.dataflow.util
- 
Constant property names used by the SDK in CloudWorkflow specifications. 
- PropertyNames() - Constructor for class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- ProtobufByteStringOutputStream() - Constructor for class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark.ProtobufByteStringOutputStream
-  
- ProtobufCoderProviderRegistrar - Class in org.apache.beam.sdk.extensions.protobuf
- 
- ProtobufCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.extensions.protobuf.ProtobufCoderProviderRegistrar
-  
- ProtoByteUtils - Class in org.apache.beam.sdk.extensions.protobuf
- 
Utility class for working with Protocol Buffer (Proto) data. 
- ProtoByteUtils() - Constructor for class org.apache.beam.sdk.extensions.protobuf.ProtoByteUtils
-  
- ProtoCoder<T extends com.google.protobuf.Message> - Class in org.apache.beam.sdk.extensions.protobuf
- 
A  Coder using Google Protocol Buffers binary format. 
- ProtoCoder(Class<T>, Set<Class<?>>) - Constructor for class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
- 
Private constructor. 
- ProtoDomain - Class in org.apache.beam.sdk.extensions.protobuf
- 
ProtoDomain is a container class for Protobuf descriptors. 
- ProtoDynamicMessageSchema<T> - Class in org.apache.beam.sdk.extensions.protobuf
-  
- ProtoFromBytes<T extends com.google.protobuf.Message> - Class in org.apache.beam.sdk.io.gcp.pubsublite.internal
-  
- ProtoMessageSchema - Class in org.apache.beam.sdk.extensions.protobuf
-  
- ProtoMessageSchema() - Constructor for class org.apache.beam.sdk.extensions.protobuf.ProtoMessageSchema
-  
- protoModeToJsonMode(TableFieldSchema.Mode) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProto
-  
- ProtoPayloadSerializerProvider - Class in org.apache.beam.sdk.extensions.protobuf
-  
- ProtoPayloadSerializerProvider() - Constructor for class org.apache.beam.sdk.extensions.protobuf.ProtoPayloadSerializerProvider
-  
- ProtoSchemaLogicalTypes - Class in org.apache.beam.sdk.extensions.protobuf
- 
- ProtoSchemaLogicalTypes() - Constructor for class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes
-  
- ProtoSchemaLogicalTypes.DurationConvert - Class in org.apache.beam.sdk.extensions.protobuf
-  
- ProtoSchemaLogicalTypes.Fixed32 - Class in org.apache.beam.sdk.extensions.protobuf
- 
A Fixed32 type. 
- ProtoSchemaLogicalTypes.Fixed64 - Class in org.apache.beam.sdk.extensions.protobuf
- 
A Fixed64 type. 
- ProtoSchemaLogicalTypes.SFixed32 - Class in org.apache.beam.sdk.extensions.protobuf
- 
A SFixed32 type. 
- ProtoSchemaLogicalTypes.SFixed64 - Class in org.apache.beam.sdk.extensions.protobuf
- 
An SFixed64 type. 
- ProtoSchemaLogicalTypes.SInt32 - Class in org.apache.beam.sdk.extensions.protobuf
- 
A SInt32 type. 
- ProtoSchemaLogicalTypes.SInt64 - Class in org.apache.beam.sdk.extensions.protobuf
- 
A SIn64 type. 
- ProtoSchemaLogicalTypes.TimestampConvert - Class in org.apache.beam.sdk.extensions.protobuf
-  
- ProtoSchemaLogicalTypes.UInt32 - Class in org.apache.beam.sdk.extensions.protobuf
- 
A UInt32 type. 
- ProtoSchemaLogicalTypes.UInt64 - Class in org.apache.beam.sdk.extensions.protobuf
- 
A UIn64 type. 
- protoSchemaToTableSchema(TableSchema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProto
-  
- protoTableFieldToTableField(TableFieldSchema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProto
-  
- protoTableSchemaFromAvroSchema(Schema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.AvroGenericRecordToStorageApiProto
- 
Given an Avro Schema, returns a protocol-buffer TableSchema that can be used to write data
 through BigQuery Storage API. 
- ProtoToBytes<T extends com.google.protobuf.Message> - Class in org.apache.beam.sdk.io.gcp.pubsublite.internal
-  
- ProtoToBytes() - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.internal.ProtoToBytes
-  
- protoTypeToJsonType(TableFieldSchema.Type) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProto
-  
- provide(String, Map<String, String>) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- Provider(PipelineOptions) - Constructor for class org.apache.beam.runners.fnexecution.environment.DockerEnvironmentFactory.Provider
-  
- Provider(PipelineOptions) - Constructor for class org.apache.beam.runners.fnexecution.environment.EmbeddedEnvironmentFactory.Provider
-  
- Provider() - Constructor for class org.apache.beam.runners.fnexecution.environment.ExternalEnvironmentFactory.Provider
-  
- Provider(PipelineOptions) - Constructor for class org.apache.beam.runners.fnexecution.environment.ProcessEnvironmentFactory.Provider
-  
- Provider(InstructionRequestHandler) - Constructor for class org.apache.beam.runners.fnexecution.environment.StaticRemoteEnvironmentFactory.Provider
-  
- provider() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQuerySchemaTransformTranslation.BigQueryStorageReadSchemaTransformTranslator
-  
- provider() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQuerySchemaTransformTranslation.BigQueryWriteSchemaTransformTranslator
-  
- provider() - Method in class org.apache.beam.sdk.io.thrift.ThriftSchema.Customizer
-  
- provider() - Static method in class org.apache.beam.sdk.io.thrift.ThriftSchema
- 
Schema provider that maps any thrift type to a Beam schema, assuming that any typedefs that
 might have been used in the thrift definitions will preserve all required metadata to infer the
 beam type (which is the case for any primitive typedefs and alike). 
- provider() - Method in class org.apache.beam.sdk.schemas.transforms.SchemaTransformTranslation.SchemaTransformPayloadTranslator
-  
- Providers - Class in org.apache.beam.sdk.schemas.io
- 
Helpers for implementing the "Provider" pattern. 
- Providers.Identifyable - Interface in org.apache.beam.sdk.schemas.io
-  
- PTransform<InputT extends PInput,OutputT extends POutput> - Class in org.apache.beam.sdk.transforms
- 
A  PTransform<InputT, OutputT> is an operation that takes an  InputT (some subtype
 of  PInput) and produces an  OutputT (some subtype of  POutput). 
- PTransform() - Constructor for class org.apache.beam.sdk.transforms.PTransform
-  
- PTransform(String) - Constructor for class org.apache.beam.sdk.transforms.PTransform
-  
- PTransformErrorHandler(PTransform<PCollection<ErrorT>, OutputT>, Pipeline, Coder<ErrorT>) - Constructor for class org.apache.beam.sdk.transforms.errorhandling.ErrorHandler.PTransformErrorHandler
- 
Constructs a new ErrorHandler, but should not be called directly. 
- publish(PubsubClient.TopicPath, List<PubsubClient.OutgoingMessage>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
- 
Publish outgoingMessagesto Pubsubtopic.
 
- publish(PubsubClient.TopicPath, List<PubsubClient.OutgoingMessage>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
-  
- publish(PubsubClient.TopicPath, List<PubsubClient.OutgoingMessage>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
-  
- publish(PubsubClient.TopicPath, List<PubsubClient.OutgoingMessage>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- publish(List<PubsubMessage>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
- 
- publishBatch(List<Solace.Record>, boolean, SerializableFunction<Solace.Record, Destination>, DeliveryMode) - Method in interface org.apache.beam.sdk.io.solace.broker.MessageProducer
- 
Publishes a batch of messages to the broker. 
- publishBatch(List<Solace.Record>, boolean, SerializableFunction<Solace.Record, Destination>, DeliveryMode) - Method in class org.apache.beam.sdk.io.solace.broker.SolaceMessageProducer
-  
- PublisherOptions - Class in org.apache.beam.sdk.io.gcp.pubsublite
- 
Options needed for a Pub/Sub Lite Publisher. 
- PublisherOptions() - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.PublisherOptions
-  
- PublisherOptions.Builder - Class in org.apache.beam.sdk.io.gcp.pubsublite
-  
- publishLatencyMetrics() - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Write
- 
Publish latency metrics using Beam metrics. 
- PublishResult() - Constructor for class org.apache.beam.sdk.io.solace.data.Solace.PublishResult
-  
- PublishResultHandler - Class in org.apache.beam.sdk.io.solace.broker
- 
This class is required to handle callbacks from Solace, to find out if messages were actually
 published or there were any kind of error. 
- PublishResultHandler(Queue<Solace.PublishResult>) - Constructor for class org.apache.beam.sdk.io.solace.broker.PublishResultHandler
-  
- publishResults(UnboundedSolaceWriter.BeamContextWrapper) - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- publishSingleMessage(Solace.Record, Destination, boolean, DeliveryMode) - Method in interface org.apache.beam.sdk.io.solace.broker.MessageProducer
- 
Publishes a message to the broker. 
- publishSingleMessage(Solace.Record, Destination, boolean, DeliveryMode) - Method in class org.apache.beam.sdk.io.solace.broker.SolaceMessageProducer
-  
- PUBSUB_DYNAMIC_DESTINATIONS - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- PUBSUB_ID_ATTRIBUTE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- PUBSUB_SERIALIZED_ATTRIBUTES_FN - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- PUBSUB_SUBSCRIPTION - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- PUBSUB_SUBSCRIPTION_OVERRIDE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- PUBSUB_TIMESTAMP_ATTRIBUTE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- PUBSUB_TOPIC - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- PUBSUB_TOPIC_OVERRIDE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- PubsubClient - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
An (abstract) helper class for talking to Pubsub via an underlying transport. 
- PubsubClient() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
-  
- PubsubClient.IncomingMessage - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
A message received from Pubsub. 
- PubsubClient.OutgoingMessage - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
A message to be sent to Pubsub. 
- PubsubClient.ProjectPath - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Path representing a cloud project id. 
- PubsubClient.PubsubClientFactory - Interface in org.apache.beam.sdk.io.gcp.pubsub
- 
Factory for creating clients. 
- PubsubClient.SchemaPath - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Path representing a Pubsub schema. 
- PubsubClient.SubscriptionPath - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Path representing a Pubsub subscription. 
- PubsubClient.TopicPath - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Path representing a Pubsub topic. 
- PubsubCoderProviderRegistrar - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
- PubsubCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubCoderProviderRegistrar
-  
- PubsubDlqProvider - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- PubsubDlqProvider() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubDlqProvider
-  
- PubsubGrpcClient - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
A helper class for talking to Pubsub via grpc. 
- PubsubIO - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Read and Write  PTransforms for Cloud Pub/Sub streams. 
- PubsubIO.PubsubSubscription - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Class representing a Cloud Pub/Sub Subscription. 
- PubsubIO.PubsubTopic - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Class representing a Cloud Pub/Sub Topic. 
- PubsubIO.Read<T> - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Implementation of read methods. 
- PubsubIO.Write<T> - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Implementation of write methods. 
- PubsubIO.Write.PubsubBoundedWriter - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Writer to Pubsub which batches messages from bounded collections. 
- PubsubJsonClient - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
A Pubsub client using JSON transport. 
- PubsubLiteIO - Class in org.apache.beam.sdk.io.gcp.pubsublite
- 
I/O transforms for reading from Google Pub/Sub Lite. 
- PubsubLiteReadSchemaTransformConfiguration() - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration
-  
- PubsubLiteReadSchemaTransformProvider - Class in org.apache.beam.sdk.io.gcp.pubsublite
-  
- PubsubLiteReadSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider
-  
- PubsubLiteReadSchemaTransformProvider.ErrorFn - Class in org.apache.beam.sdk.io.gcp.pubsublite
-  
- PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration - Class in org.apache.beam.sdk.io.gcp.pubsublite
-  
- PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration.Builder - Class in org.apache.beam.sdk.io.gcp.pubsublite
-  
- PubsubLiteSink - Class in org.apache.beam.sdk.io.gcp.pubsublite.internal
- 
A sink which publishes messages to Pub/Sub Lite. 
- PubsubLiteSink(PublisherOptions) - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.internal.PubsubLiteSink
-  
- PubsubLiteTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.pubsublite
- 
Pub/Sub Lite table provider. 
- PubsubLiteTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.pubsublite.PubsubLiteTableProvider
-  
- PubsubLiteWriteSchemaTransformConfiguration() - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration
-  
- PubsubLiteWriteSchemaTransformProvider - Class in org.apache.beam.sdk.io.gcp.pubsublite
-  
- PubsubLiteWriteSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider
-  
- PubsubLiteWriteSchemaTransformProvider.ErrorCounterFn - Class in org.apache.beam.sdk.io.gcp.pubsublite
-  
- PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration - Class in org.apache.beam.sdk.io.gcp.pubsublite
-  
- PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration.Builder - Class in org.apache.beam.sdk.io.gcp.pubsublite
-  
- PubsubLiteWriteSchemaTransformProvider.SetUuidFromPubSubMessage - Class in org.apache.beam.sdk.io.gcp.pubsublite
-  
- PubsubLiteWriteSchemaTransformProvider.SetUuidFromPubSubMessage.SetUuidFn - Class in org.apache.beam.sdk.io.gcp.pubsublite
-  
- PubsubMessage - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Class representing a Pub/Sub message. 
- PubsubMessage(byte[], Map<String, String>) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
-  
- PubsubMessage(byte[], Map<String, String>, String) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
-  
- PubsubMessage(byte[], Map<String, String>, String, String) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
-  
- PubsubMessagePayloadOnlyCoder - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
A coder for PubsubMessage treating the raw bytes being decoded as the message's payload. 
- PubsubMessagePayloadOnlyCoder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessagePayloadOnlyCoder
-  
- PubsubMessages - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Common util functions for converting between PubsubMessage proto and  PubsubMessage. 
- PubsubMessages.DeserializeBytesIntoPubsubMessagePayloadOnly - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- PubsubMessages.ParsePayloadAsPubsubMessageProto - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- PubsubMessages.ParsePubsubMessageProtoAsPayload - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- PubsubMessageWithAttributesAndMessageIdAndOrderingKeyCoder - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
A coder for PubsubMessage including all fields of a PubSub message from server. 
- PubsubMessageWithAttributesAndMessageIdAndOrderingKeyCoder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesAndMessageIdAndOrderingKeyCoder
-  
- PubsubMessageWithAttributesAndMessageIdCoder - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
A coder for PubsubMessage including attributes and the message id from the PubSub server. 
- PubsubMessageWithAttributesAndMessageIdCoder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesAndMessageIdCoder
-  
- PubsubMessageWithAttributesCoder - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
A coder for PubsubMessage including attributes. 
- PubsubMessageWithAttributesCoder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
-  
- PubsubMessageWithMessageIdCoder - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
A coder for PubsubMessage treating the raw bytes being decoded as the message's payload, with the
 message id from the PubSub server. 
- PubsubMessageWithMessageIdCoder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithMessageIdCoder
-  
- PubsubMessageWithTopicCoder - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
A coder for PubsubMessage including the topic from the PubSub server. 
- PubsubMessageWithTopicCoder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithTopicCoder
-  
- PubsubOptions - Interface in org.apache.beam.sdk.io.gcp.pubsub
- 
Properties that can be set when using Google Cloud Pub/Sub with the Apache Beam SDK. 
- PubSubPayloadTranslation - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- PubSubPayloadTranslation() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubSubPayloadTranslation
-  
- PubSubPayloadTranslation.ReadRegistrar - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- PubSubPayloadTranslation.WriteRegistrar - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- PubsubReadSchemaTransformConfiguration - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Configuration for reading from Pub/Sub. 
- PubsubReadSchemaTransformConfiguration() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration
-  
- PubsubReadSchemaTransformConfiguration.Builder - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- PubsubReadSchemaTransformConfiguration.ErrorHandling - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- PubsubReadSchemaTransformConfiguration.ErrorHandling.Builder - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- PubsubReadSchemaTransformProvider - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
- PubsubReadSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformProvider
-  
- PubsubSchemaIOProvider - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
- PubsubSchemaIOProvider() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubSchemaIOProvider
-  
- PubsubTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.pubsub
- 
- PubsubTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubTableProvider
-  
- PubsubTestClient - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
A (partial) implementation of  PubsubClient for use by unit tests. 
- PubsubTestClient() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- PubsubTestClient.PubsubTestClientFactory - Interface in org.apache.beam.sdk.io.gcp.pubsub
- 
Closing the factory will validate all expected messages were processed. 
- PubsubUnboundedSink - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
A PTransform which streams messages to Pubsub. 
- PubsubUnboundedSink(PubsubClient.PubsubClientFactory, ValueProvider<PubsubClient.TopicPath>, String, String, int) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
-  
- PubsubUnboundedSink(PubsubClient.PubsubClientFactory, ValueProvider<PubsubClient.TopicPath>, String, String, int, String) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
-  
- PubsubUnboundedSink(PubsubClient.PubsubClientFactory, ValueProvider<PubsubClient.TopicPath>, String, String, int, int, int) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
-  
- PubsubUnboundedSink(PubsubClient.PubsubClientFactory, ValueProvider<PubsubClient.TopicPath>, String, String, int, int, int, String) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
-  
- PubsubUnboundedSource - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Users should use PubsubIO#readinstead.
 
- PubsubUnboundedSource(PubsubClient.PubsubClientFactory, ValueProvider<PubsubClient.ProjectPath>, ValueProvider<PubsubClient.TopicPath>, ValueProvider<PubsubClient.SubscriptionPath>, String, String, boolean) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
- 
Construct an unbounded source to consume from the Pubsub subscription.
 
- PubsubUnboundedSource(Clock, PubsubClient.PubsubClientFactory, ValueProvider<PubsubClient.ProjectPath>, ValueProvider<PubsubClient.TopicPath>, ValueProvider<PubsubClient.SubscriptionPath>, String, String, boolean) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
- 
Construct an unbounded source to consume from the Pubsub subscription.
 
- PubsubUnboundedSource(PubsubClient.PubsubClientFactory, ValueProvider<PubsubClient.ProjectPath>, ValueProvider<PubsubClient.TopicPath>, ValueProvider<PubsubClient.SubscriptionPath>, String, String, boolean, boolean) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
- 
Construct an unbounded source to consume from the Pubsub subscription.
 
- PubsubWriteSchemaTransformConfiguration - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Configuration for writing to Pub/Sub. 
- PubsubWriteSchemaTransformConfiguration() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration
-  
- PubsubWriteSchemaTransformConfiguration.Builder - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- PubsubWriteSchemaTransformConfiguration.ErrorHandling - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- PubsubWriteSchemaTransformConfiguration.ErrorHandling.Builder - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- PubsubWriteSchemaTransformProvider - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
- PubsubWriteSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformProvider
-  
- PubsubWriteSchemaTransformProvider.ErrorFn - Class in org.apache.beam.sdk.io.gcp.pubsub
-  
- pull(long, PubsubClient.SubscriptionPath, int, boolean) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
- 
Request the next batch of up to batchSizemessages fromsubscription.
 
- pull(long, PubsubClient.SubscriptionPath, int, boolean) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
-  
- pull(long, PubsubClient.SubscriptionPath, int, boolean) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
-  
- pull(long, PubsubClient.SubscriptionPath, int, boolean) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
-  
- PulsarIO - Class in org.apache.beam.sdk.io.pulsar
- 
Class for reading and writing from Apache Pulsar. 
- PulsarIO.Read - Class in org.apache.beam.sdk.io.pulsar
-  
- PulsarIO.Write - Class in org.apache.beam.sdk.io.pulsar
-  
- PulsarMessage - Class in org.apache.beam.sdk.io.pulsar
- 
Class representing a Pulsar Message record. 
- PulsarMessage(String, Long, Object) - Constructor for class org.apache.beam.sdk.io.pulsar.PulsarMessage
-  
- PulsarMessage(String, Long) - Constructor for class org.apache.beam.sdk.io.pulsar.PulsarMessage
-  
- PulsarMessageCoder - Class in org.apache.beam.sdk.io.pulsar
-  
- PulsarMessageCoder() - Constructor for class org.apache.beam.sdk.io.pulsar.PulsarMessageCoder
-  
- PulsarSourceDescriptor - Class in org.apache.beam.sdk.io.pulsar
-  
- PulsarSourceDescriptor() - Constructor for class org.apache.beam.sdk.io.pulsar.PulsarSourceDescriptor
-  
- PUSH_DOWN_OPTION - Static variable in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
-  
- pushArrayBuffer(ArrayBuffer<?>, Option<Object>, Option<StreamBlockId>) - Method in class org.apache.beam.sdk.io.sparkreceiver.WrappedSupervisor
-  
- pushBytes(ByteBuffer, Option<Object>, Option<StreamBlockId>) - Method in class org.apache.beam.sdk.io.sparkreceiver.WrappedSupervisor
-  
- pushIterator(Iterator<?>, Option<Object>, Option<StreamBlockId>) - Method in class org.apache.beam.sdk.io.sparkreceiver.WrappedSupervisor
-  
- pushSingle(Object) - Method in class org.apache.beam.sdk.io.sparkreceiver.WrappedSupervisor
-  
- put(String, InstructionRequestHandler) - Method in interface org.apache.beam.runners.fnexecution.control.ControlClientPool.Sink
- 
- put(T) - Method in class org.apache.beam.sdk.fn.CancellableQueue
- 
Adds an element to this queue. 
- put(K2, V2) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
-  
- put(K, V) - Method in interface org.apache.beam.sdk.state.MapState
- 
Associates the specified value with the specified key in this state. 
- put(K, V) - Method in interface org.apache.beam.sdk.state.MultimapState
- 
Associates the specified value with the specified key in this multimap. 
- putAll(Map<? extends K2, ? extends V2>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
-  
- putDataset(PCollection<T>, Dataset<WindowedValue<T>>, boolean) - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator.TranslationState
-  
- putDataset(PCollection<T>, Dataset<WindowedValue<T>>) - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator.TranslationState
-  
- putDataset(PCollection<T>, Dataset<WindowedValue<T>>, boolean) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- putIfAbsent(K, V) - Method in interface org.apache.beam.sdk.state.MapState
- 
A deferred read-followed-by-write. 
- putSchemaIfAbsent(TableReference, TableSchema) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableSchemaCache
- 
Registers schema for a table if one is not already present. 
- putUnresolved(PCollection<OutT>, PipelineTranslator.UnresolvedTranslation<InT, OutT>) - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator.TranslationState
-  
- putUnresolved(PCollection<T>, PipelineTranslator.UnresolvedTranslation<InputT, T>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- PValue - Interface in org.apache.beam.sdk.values
- 
For internal use. 
- PValueBase - Class in org.apache.beam.sdk.values
- 
For internal use. 
- PValueBase(Pipeline) - Constructor for class org.apache.beam.sdk.values.PValueBase
-  
- PValueBase() - Constructor for class org.apache.beam.sdk.values.PValueBase
- 
No-arg constructor to allow subclasses to implement Serializable.
 
- PValues - Class in org.apache.beam.sdk.values
- 
For internal use. 
- PythonCallable - Class in org.apache.beam.sdk.schemas.logicaltypes
- 
A logical type for PythonCallableSource objects. 
- PythonCallable() - Constructor for class org.apache.beam.sdk.schemas.logicaltypes.PythonCallable
-  
- PythonExternalTransform<InputT extends PInput,OutputT extends POutput> - Class in org.apache.beam.sdk.extensions.python
- 
Wrapper for invoking external Python transforms. 
- PythonExternalTransformOptions - Interface in org.apache.beam.sdk.extensions.python
- 
- PythonExternalTransformOptionsRegistrar - Class in org.apache.beam.sdk.extensions.python
- 
- PythonExternalTransformOptionsRegistrar() - Constructor for class org.apache.beam.sdk.extensions.python.PythonExternalTransformOptionsRegistrar
-  
- PythonMap<InputT,OutputT> - Class in org.apache.beam.sdk.extensions.python.transforms
- 
Wrapper for invoking external Python Maptransforms..
 
- PythonService - Class in org.apache.beam.sdk.extensions.python
- 
Utility to bootstrap and start a Beam Python service. 
- PythonService(String, List<String>, List<String>) - Constructor for class org.apache.beam.sdk.extensions.python.PythonService
-  
- PythonService(String, List<String>) - Constructor for class org.apache.beam.sdk.extensions.python.PythonService
-  
- PythonService(String, String...) - Constructor for class org.apache.beam.sdk.extensions.python.PythonService
-  
- RabbitMqIO - Class in org.apache.beam.sdk.io.rabbitmq
- 
A IO to publish or consume messages with a RabbitMQ broker. 
- RabbitMqIO.Read - Class in org.apache.beam.sdk.io.rabbitmq
- 
A  PTransform to consume messages from RabbitMQ server. 
- RabbitMqIO.Write - Class in org.apache.beam.sdk.io.rabbitmq
- 
A  PTransform to publish messages to a RabbitMQ server. 
- RabbitMqMessage - Class in org.apache.beam.sdk.io.rabbitmq
- 
It contains the message payload, and additional metadata like routing key or attributes. 
- RabbitMqMessage(byte[]) - Constructor for class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- RabbitMqMessage(String, GetResponse) - Constructor for class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- RabbitMqMessage(String, byte[], String, String, Map<String, Object>, Integer, Integer, String, String, String, String, Date, String, String, String, String) - Constructor for class org.apache.beam.sdk.io.rabbitmq.RabbitMqMessage
-  
- RampupThrottlingFn<T> - Class in org.apache.beam.sdk.io.gcp.datastore
- 
An implementation of a client-side throttler that enforces a gradual ramp-up, broadly in line
 with Datastore best practices. 
- RampupThrottlingFn(ValueProvider<Integer>, PCollectionView<Instant>) - Constructor for class org.apache.beam.sdk.io.gcp.datastore.RampupThrottlingFn
-  
- RampupThrottlingFn(int, PCollectionView<Instant>) - Constructor for class org.apache.beam.sdk.io.gcp.datastore.RampupThrottlingFn
-  
- random() - Static method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.Uuid
-  
- RandomAccessData - Class in org.apache.beam.runners.dataflow.util
- 
An elastic-sized byte array which allows you to manipulate it as a stream, or access it directly. 
- RandomAccessData() - Constructor for class org.apache.beam.runners.dataflow.util.RandomAccessData
- 
Constructs a RandomAccessData with a default buffer size. 
- RandomAccessData(byte[]) - Constructor for class org.apache.beam.runners.dataflow.util.RandomAccessData
- 
Constructs a RandomAccessData with the initial buffer. 
- RandomAccessData(int) - Constructor for class org.apache.beam.runners.dataflow.util.RandomAccessData
- 
Constructs a RandomAccessData with the given buffer size. 
- RandomAccessData.RandomAccessDataCoder - Class in org.apache.beam.runners.dataflow.util
- 
A  Coder which encodes the valid parts of this stream. 
- RandomAccessData.UnsignedLexicographicalComparator - Class in org.apache.beam.runners.dataflow.util
- 
A Comparatorthat compares two byte arrays lexicographically.
 
- RandomAccessDataCoder() - Constructor for class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
-  
- range - Variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRangeTracker
-  
- range - Variable in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
-  
- RANGE_OFFSET - Static variable in class org.apache.beam.sdk.io.contextualtextio.RecordWithMetadata
-  
- RangeTracker<PositionT> - Interface in org.apache.beam.sdk.io.range
- 
A  RangeTracker is a thread-safe helper object for implementing dynamic work rebalancing
 in position-based  BoundedSource.BoundedReader subclasses. 
- Rate() - Constructor for class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration.Rate
-  
- RateLimitPolicy - Interface in org.apache.beam.sdk.io.aws2.kinesis
-  
- RateLimitPolicyFactory - Interface in org.apache.beam.sdk.io.aws2.kinesis
- 
Implement this interface to create a RateLimitPolicy.
 
- RateLimitPolicyFactory.DefaultRateLimiter - Class in org.apache.beam.sdk.io.aws2.kinesis
- 
Default rate limiter that throttles reading from a shard using an exponential backoff if the
 response is empty or if the consumer is throttled by AWS. 
- RateLimitPolicyFactory.DelayIntervalRateLimiter - Class in org.apache.beam.sdk.io.aws2.kinesis
-  
- RawUnionValue - Class in org.apache.beam.sdk.transforms.join
- 
This corresponds to an integer union tag and value. 
- RawUnionValue(int, Object) - Constructor for class org.apache.beam.sdk.transforms.join.RawUnionValue
- 
Constructs a partial union from the given union tag and value. 
- read() - Static method in class org.apache.beam.io.debezium.DebeziumIO
- 
Read data from a Debezium source. 
- Read() - Constructor for class org.apache.beam.io.debezium.DebeziumIO.Read
-  
- read(DataInputView) - Method in class org.apache.beam.runners.flink.adapter.FlinkKey
-  
- read(JavaStreamingContext, SerializablePipelineOptions, UnboundedSource<T, CheckpointMarkT>, String) - Static method in class org.apache.beam.runners.spark.io.SparkUnboundedSource
-  
- read(Kryo, Input) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.SideInputValues.BaseSideInputValues
-  
- read(Class<T>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
Reads records of the given type from an Avro file (or multiple Avro files matching a pattern). 
- Read() - Constructor for class org.apache.beam.sdk.extensions.avro.io.AvroIO.Read
-  
- read(ByteBuffer) - Method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingReadableByteChannel
-  
- read(ByteBuffer) - Method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingSeekableByteChannel
-  
- read(T) - Method in interface org.apache.beam.sdk.fn.stream.DataStreams.OutputChunkConsumer
-  
- read() - Static method in class org.apache.beam.sdk.io.amqp.AmqpIO
-  
- Read() - Constructor for class org.apache.beam.sdk.io.amqp.AmqpIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO
-  
- Read() - Constructor for class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO
- 
- Read() - Constructor for class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO
-  
- Read() - Constructor for class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Read
-  
- read(Class<T>) - Static method in class org.apache.beam.sdk.io.azure.cosmos.CosmosIO
- 
- Read() - Constructor for class org.apache.beam.sdk.io.azure.cosmos.CosmosIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.cassandra.CassandraIO
- 
- Read() - Constructor for class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.cdap.CdapIO
-  
- Read() - Constructor for class org.apache.beam.sdk.io.cdap.CdapIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO
- 
A  PTransform that reads from one or more text files and returns a bounded  PCollection containing one  element for each line in the input files. 
- Read() - Constructor for class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO
-  
- Read() - Constructor for class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
- 
- read(SerializableFunction<SchemaAndRecord, T>) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
- 
Reads from a BigQuery table or query and returns a  PCollection with one element per
 each row of the table or query result, parsed from the BigQuery AVRO format using the specified
 function. 
- read() - Static method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO
- 
- Read() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-  
- read() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1
- 
- Read() - Constructor for class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-  
- read() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1
- 
The class returned by this method provides the ability to create  PTransforms
 for read operations available in the Firestore V1 API provided by  FirestoreStub. 
- Read() - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Read
- 
Instantiates a new Read. 
- read(String) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO
- 
Read all HL7v2 Messages from a single store. 
- read(ValueProvider<String>) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO
- 
Read all HL7v2 Messages from a single store. 
- Read() - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.Read
-  
- Read() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
-  
- read(SubscriberOptions) - Static method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteIO
- 
Read messages from Pub/Sub Lite. 
- read(Object, Decoder) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.encoder.TimestampEncoding
- 
Deserializes a  Timestamp from the given  Decoder. 
- read() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO
- 
- Read() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- read() - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17
-  
- Read() - Constructor for class org.apache.beam.sdk.io.googleads.GoogleAdsV17.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
- 
- Read() - Constructor for class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.hbase.HBaseIO
- 
- read() - Static method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO
- 
Read data from Hive. 
- Read() - Constructor for class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO
-  
- Read() - Constructor for class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
- 
Read data from a JDBC datasource. 
- Read() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.jms.JmsIO
-  
- Read() - Constructor for class org.apache.beam.sdk.io.jms.JmsIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.kafka.KafkaIO
- 
- Read() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
-  
- read() - Static method in class org.apache.beam.sdk.io.kudu.KuduIO
-  
- Read() - Constructor for class org.apache.beam.sdk.io.kudu.KuduIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO
- 
Read data from GridFS. 
- Read() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbIO
- 
Read data from MongoDB. 
- Read() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.mqtt.MqttIO
-  
- Read() - Constructor for class org.apache.beam.sdk.io.mqtt.MqttIO.Read
-  
- read(Schema) - Static method in class org.apache.beam.sdk.io.parquet.ParquetIO
- 
Reads  GenericRecord from a Parquet file (or multiple Parquet files matching the
 pattern). 
- Read() - Constructor for class org.apache.beam.sdk.io.parquet.ParquetIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.pulsar.PulsarIO
- 
Read from Apache Pulsar. 
- Read() - Constructor for class org.apache.beam.sdk.io.pulsar.PulsarIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO
-  
- Read() - Constructor for class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Read
-  
- Read - Class in org.apache.beam.sdk.io
- 
- Read() - Constructor for class org.apache.beam.sdk.io.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.redis.RedisIO
- 
Read data from a Redis server. 
- Read() - Constructor for class org.apache.beam.sdk.io.redis.RedisIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO
- 
Read data from a SingleStoreDB datasource. 
- Read() - Constructor for class org.apache.beam.sdk.io.singlestore.SingleStoreIO.Read
-  
- read(SnowflakeBatchServiceConfig) - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceImpl
- 
Reading data from Snowflake tables in batch processing. 
- read(SnowflakeBatchServiceConfig) - Method in interface org.apache.beam.sdk.io.snowflake.services.SnowflakeServices.BatchService
-  
- read(SnowflakeStreamingServiceConfig) - Method in interface org.apache.beam.sdk.io.snowflake.services.SnowflakeServices.StreamingService
-  
- read(SnowflakeStreamingServiceConfig) - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeStreamingServiceImpl
- 
Reading data from Snowflake in streaming mode is not supported. 
- read(SnowflakeServices) - Static method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO
- 
- read() - Static method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO
- 
- Read() - Constructor for class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.solace.SolaceIO
- 
- read(TypeDescriptor<T>, SerializableFunction<BytesXMLMessage, T>, SerializableFunction<T, Instant>) - Static method in class org.apache.beam.sdk.io.solace.SolaceIO
- 
- read() - Static method in class org.apache.beam.sdk.io.solr.SolrIO
-  
- Read() - Constructor for class org.apache.beam.sdk.io.solr.SolrIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.sparkreceiver.SparkReceiverIO
-  
- Read() - Constructor for class org.apache.beam.sdk.io.sparkreceiver.SparkReceiverIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.TextIO
- 
A  PTransform that reads from one or more text files and returns a bounded  PCollection containing one element for each line of the input files. 
- Read() - Constructor for class org.apache.beam.sdk.io.TextIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.TFRecordIO
- 
A  PTransform that reads from a TFRecord file (or multiple TFRecord files matching a
 pattern) and returns a  PCollection containing the decoding of each of the records of
 the TFRecord file(s) as a byte array. 
- Read() - Constructor for class org.apache.beam.sdk.io.TFRecordIO.Read
-  
- read() - Static method in class org.apache.beam.sdk.io.xml.XmlIO
- 
Reads XML files as a  PCollection of a given type mapped via JAXB. 
- Read() - Constructor for class org.apache.beam.sdk.io.xml.XmlIO.Read
-  
- read(String) - Static method in class org.apache.beam.sdk.managed.Managed
- 
- read() - Method in interface org.apache.beam.sdk.state.BagState
-  
- read() - Method in interface org.apache.beam.sdk.state.CombiningState
-  
- read() - Method in interface org.apache.beam.sdk.state.ReadableState
- 
Read the current value, blocking until it is available. 
- read() - Method in interface org.apache.beam.sdk.state.ValueState
- 
Read the current value, blocking until it is available. 
- Read.Bounded<T> - Class in org.apache.beam.sdk.io
- 
- Read.Builder - Class in org.apache.beam.sdk.io
- 
Helper class for building  Read transforms. 
- Read.Unbounded<T> - Class in org.apache.beam.sdk.io
- 
- READ_DATA_URN - Static variable in class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar
-  
- READ_JSON_URN - Static variable in class org.apache.beam.io.debezium.DebeziumTransformRegistrar
-  
- READ_TRANSFORMS - Static variable in class org.apache.beam.sdk.managed.Managed
-  
- READ_URN - Static variable in class org.apache.beam.sdk.io.gcp.pubsublite.internal.ExternalTransformRegistrarImpl
-  
- READ_URN - Static variable in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar
-  
- READ_URN - Static variable in class org.apache.beam.sdk.io.snowflake.crosslanguage.SnowflakeTransformRegistrar
-  
- ReadableFileCoder - Class in org.apache.beam.sdk.io
- 
- ReadableState<T> - Interface in org.apache.beam.sdk.state
- 
- ReadableStates - Class in org.apache.beam.sdk.state
- 
For internal use only; no backwards-compatibility guarantees. 
- ReadableStates() - Constructor for class org.apache.beam.sdk.state.ReadableStates
-  
- readAll(Class<T>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- ReadAll() - Constructor for class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadAll
- 
Deprecated. 
- readAll() - Static method in class org.apache.beam.sdk.io.cassandra.CassandraIO
- 
- ReadAll() - Constructor for class org.apache.beam.sdk.io.cassandra.CassandraIO.ReadAll
-  
- readAll(List<String>) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO
- 
Read all HL7v2 Messages from multiple stores. 
- readAll(ValueProvider<List<String>>) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO
- 
Read all HL7v2 Messages from multiple stores. 
- readAll() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO
- 
- ReadAll() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-  
- readAll() - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17
-  
- ReadAll() - Constructor for class org.apache.beam.sdk.io.googleads.GoogleAdsV17.ReadAll
-  
- readAll() - Static method in class org.apache.beam.sdk.io.hbase.HBaseIO
- 
- readAll() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
- 
Like  JdbcIO.read(), but executes multiple instances of the query substituting each element of a
  PCollection as query parameters. 
- ReadAll() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-  
- readAll() - Static method in class org.apache.beam.sdk.io.neo4j.Neo4jIO
- 
Read all rows using a Neo4j Cypher query. 
- ReadAll() - Constructor for class org.apache.beam.sdk.io.neo4j.Neo4jIO.ReadAll
-  
- readAll() - Static method in class org.apache.beam.sdk.io.solr.SolrIO
-  
- ReadAll() - Constructor for class org.apache.beam.sdk.io.solr.SolrIO.ReadAll
-  
- readAll() - Static method in class org.apache.beam.sdk.io.TextIO
- 
- ReadAll() - Constructor for class org.apache.beam.sdk.io.TextIO.ReadAll
- 
Deprecated. 
- readAllGenericRecords(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- readAllGenericRecords(String) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- readAllRequests() - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO
- 
- readAllStreamPartitions() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
- 
Read all the StreamPartition and output PartitionRecord to stream them. 
- ReadAllViaFileBasedSource<T> - Class in org.apache.beam.sdk.io
- 
- ReadAllViaFileBasedSource(long, SerializableFunction<String, ? extends FileBasedSource<T>>, Coder<T>) - Constructor for class org.apache.beam.sdk.io.ReadAllViaFileBasedSource
-  
- ReadAllViaFileBasedSource(long, SerializableFunction<String, ? extends FileBasedSource<T>>, Coder<T>, boolean, ReadAllViaFileBasedSource.ReadFileRangesFnExceptionHandler) - Constructor for class org.apache.beam.sdk.io.ReadAllViaFileBasedSource
-  
- ReadAllViaFileBasedSource.ReadFileRangesFnExceptionHandler - Class in org.apache.beam.sdk.io
- 
A class to handle errors which occur during file reads. 
- ReadAllViaFileBasedSourceTransform<InT,T> - Class in org.apache.beam.sdk.io
-  
- ReadAllViaFileBasedSourceTransform(long, SerializableFunction<String, ? extends FileBasedSource<InT>>, Coder<T>) - Constructor for class org.apache.beam.sdk.io.ReadAllViaFileBasedSourceTransform
-  
- ReadAllViaFileBasedSourceTransform(long, SerializableFunction<String, ? extends FileBasedSource<InT>>, Coder<T>, boolean, ReadAllViaFileBasedSource.ReadFileRangesFnExceptionHandler) - Constructor for class org.apache.beam.sdk.io.ReadAllViaFileBasedSourceTransform
-  
- ReadAllViaFileBasedSourceTransform.AbstractReadFileRangesFn<InT,T> - Class in org.apache.beam.sdk.io
-  
- ReadAllViaFileBasedSourceTransform.SplitIntoRangesFn - Class in org.apache.beam.sdk.io
-  
- ReadAllViaFileBasedSourceWithFilename<T> - Class in org.apache.beam.sdk.io
- 
Reads each file of the input  PCollection and outputs each element as the value of a
  KV, where the key is the filename from which that value came. 
- ReadAllViaFileBasedSourceWithFilename(long, SerializableFunction<String, ? extends FileBasedSource<T>>, Coder<KV<String, T>>) - Constructor for class org.apache.beam.sdk.io.ReadAllViaFileBasedSourceWithFilename
-  
- readAllWithFilter(List<String>, String) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO
- 
Read all HL7v2 Messages from a multiple stores matching a filter. 
- readAllWithFilter(ValueProvider<List<String>>, ValueProvider<String>) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO
- 
Read all HL7v2 Messages from a multiple stores matching a filter. 
- readAsJson() - Static method in class org.apache.beam.io.debezium.DebeziumIO
- 
- readAvroGenericRecords(Schema) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
- readAvros(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that continuously reads binary encoded Avro messages of the given
 type from a Google Cloud Pub/Sub stream. 
- readAvrosWithBeamSchema(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns a  PTransform that continuously reads binary encoded Avro messages of the
 specific type. 
- ReadBuilder() - Constructor for class org.apache.beam.io.debezium.DebeziumTransformRegistrar.ReadBuilder
-  
- ReadBuilder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.ReadBuilder
-  
- ReadBuilder() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.ReadBuilder
-  
- ReadBuilder - Class in org.apache.beam.sdk.io.snowflake.crosslanguage
-  
- ReadBuilder() - Constructor for class org.apache.beam.sdk.io.snowflake.crosslanguage.ReadBuilder
-  
- ReadBuilder.Configuration - Class in org.apache.beam.sdk.io.snowflake.crosslanguage
- 
Parameters class to expose the transform to an external SDK. 
- readBytes() - Static method in class org.apache.beam.sdk.io.kafka.KafkaIO
- 
A specific instance of uninitialized  KafkaIO.read() where key and values are bytes. 
- readCallMetric(TableReference) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
-  
- readChangeStream() - Static method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO
- 
- ReadChangeStream() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
-  
- readChangeStream() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO
- 
- ReadChangeStream() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
-  
- readChangeStreamPartition(PartitionRecord, StreamProgress, Instant, Duration) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.ChangeStreamDao
- 
Streams a partition. 
- readChangeStreamPartitionAction(MetadataTableDao, ChangeStreamDao, ChangeStreamMetrics, ChangeStreamAction, Duration, SizeEstimator<KV<ByteString, ChangeStreamRecord>>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.ActionFactory
- 
- ReadChangeStreamPartitionAction - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.action
- 
- ReadChangeStreamPartitionAction(MetadataTableDao, ChangeStreamDao, ChangeStreamMetrics, ChangeStreamAction, Duration, SizeEstimator<KV<ByteString, ChangeStreamRecord>>) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.ReadChangeStreamPartitionAction
-  
- ReadChangeStreamPartitionDoFn - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn
-  
- ReadChangeStreamPartitionDoFn(DaoFactory, ActionFactory, ChangeStreamMetrics, Duration) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.ReadChangeStreamPartitionDoFn
-  
- ReadChangeStreamPartitionDoFn - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn
- 
A SDF (Splittable DoFn) class which is responsible for performing a change stream query for a
 given partition. 
- ReadChangeStreamPartitionDoFn(DaoFactory, MapperFactory, ActionFactory, ChangeStreamMetrics) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.ReadChangeStreamPartitionDoFn
- 
This class needs a  DaoFactory to build DAOs to access the partition metadata tables and
 to perform the change streams query. 
- ReadChangeStreamPartitionProgressTracker - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction
- 
RestrictionTracker used by  ReadChangeStreamPartitionDoFn to keep
 track of the progress of the stream and to split the restriction for runner initiated
 checkpoints. 
- ReadChangeStreamPartitionProgressTracker(StreamProgress) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.ReadChangeStreamPartitionProgressTracker
- 
Constructs a restriction tracker with the streamProgress. 
- ReadChangeStreamPartitionRangeTracker - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction
- 
- ReadChangeStreamPartitionRangeTracker(PartitionMetadata, TimestampRange) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.ReadChangeStreamPartitionRangeTracker
- 
Receives the partition that will be queried and the timestamp range that belongs to it. 
- ReadDataBuilder() - Constructor for class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.ReadDataBuilder
-  
- readDecompressed(ReadableByteChannel) - Method in enum org.apache.beam.sdk.io.Compression
-  
- readDetectNewPartitionMissingPartitions() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
- 
Read and deserialize missing partition and how long they have been missing from the metadata
 table. 
- readDetectNewPartitionsState() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
- 
Read the low watermark of the pipeline from Detect New Partition row. 
- Reader() - Constructor for class org.apache.beam.sdk.io.Source.Reader
-  
- ReaderInvocationUtil<OutputT,ReaderT extends Source.Reader<OutputT>> - Class in org.apache.beam.runners.flink.metrics
- 
Util for invoking Source.Readermethods that might require aMetricsContainerImplto be active.
 
- ReaderInvocationUtil(String, PipelineOptions, FlinkMetricContainerBase) - Constructor for class org.apache.beam.runners.flink.metrics.ReaderInvocationUtil
-  
- readExternal(ObjectInput) - Method in class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
-  
- readFhirResource(String) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Read fhir resource http body. 
- readFhirResource(String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- readField(Row, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.RowBundle
- 
- readField(Row, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.RowBundles.ArrayOfNestedStringBundle
-  
- readField(Row, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.RowBundles.ArrayOfStringBundle
-  
- readField(Row, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.RowBundles.MapOfIntBundle
-  
- readField(Row, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.RowBundles.MapOfNestedIntBundle
-  
- readField(Row, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.RowBundles.NestedBytesBundle
-  
- readField(Row, Blackhole) - Method in class org.apache.beam.sdk.jmh.schemas.RowBundles.NestedIntBundle
-  
- ReadFileRangesFnExceptionHandler() - Constructor for class org.apache.beam.sdk.io.ReadAllViaFileBasedSource.ReadFileRangesFnExceptionHandler
-  
- readFiles(Class<T>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- ReadFiles() - Constructor for class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadFiles
-  
- readFiles() - Static method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO
- 
- ReadFiles() - Constructor for class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.ReadFiles
-  
- readFiles(Schema) - Static method in class org.apache.beam.sdk.io.parquet.ParquetIO
- 
- ReadFiles() - Constructor for class org.apache.beam.sdk.io.parquet.ParquetIO.ReadFiles
-  
- readFiles() - Static method in class org.apache.beam.sdk.io.TextIO
- 
- ReadFiles() - Constructor for class org.apache.beam.sdk.io.TextIO.ReadFiles
-  
- readFiles() - Static method in class org.apache.beam.sdk.io.TFRecordIO
- 
- ReadFiles() - Constructor for class org.apache.beam.sdk.io.TFRecordIO.ReadFiles
-  
- readFiles(Class<T>) - Static method in class org.apache.beam.sdk.io.thrift.ThriftIO
- 
- ReadFiles() - Constructor for class org.apache.beam.sdk.io.thrift.ThriftIO.ReadFiles
-  
- readFiles() - Static method in class org.apache.beam.sdk.io.xml.XmlIO
- 
- ReadFiles() - Constructor for class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
-  
- readFilesGenericRecords(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- readFilesGenericRecords(String) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- readFrom(InputStream, int, int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
- 
Reads lengthbytes from the specified input stream writing them into the backing data
 store starting atoffset.
 
- ReadFromMySqlSchemaTransformProvider - Class in org.apache.beam.sdk.io.jdbc.providers
-  
- ReadFromMySqlSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.jdbc.providers.ReadFromMySqlSchemaTransformProvider
-  
- ReadFromOracleSchemaTransformProvider - Class in org.apache.beam.sdk.io.jdbc.providers
-  
- ReadFromOracleSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.jdbc.providers.ReadFromOracleSchemaTransformProvider
-  
- readFromPort(BeamFnApi.RemoteGrpcPort, String) - Static method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortRead
-  
- ReadFromPostgresSchemaTransformProvider - Class in org.apache.beam.sdk.io.jdbc.providers
-  
- ReadFromPostgresSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.jdbc.providers.ReadFromPostgresSchemaTransformProvider
-  
- ReadFromPulsarDoFn - Class in org.apache.beam.sdk.io.pulsar
- 
Transform for reading from Apache Pulsar. 
- ReadFromPulsarDoFn(PulsarIO.Read) - Constructor for class org.apache.beam.sdk.io.pulsar.ReadFromPulsarDoFn
-  
- readFromSource(BoundedSource<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
- 
- readFromSplitsOfSource(BoundedSource<T>, long, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
-  
- ReadFromSqlServerSchemaTransformProvider - Class in org.apache.beam.sdk.io.jdbc.providers
-  
- ReadFromSqlServerSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.jdbc.providers.ReadFromSqlServerSchemaTransformProvider
-  
- readFromStartedReader(Source.Reader<T>) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
- 
Reads all elements from the given started Source.Reader.
 
- readFromUnstartedReader(Source.Reader<T>) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
- 
Reads all elements from the given unstarted Source.Reader.
 
- readFullyAsBytes() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
- 
Returns the full contents of the file as bytes. 
- readFullyAsUTF8String() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
- 
Returns the full contents of the file as a Stringdecoded as UTF-8.
 
- readGenericRecords(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
Reads Avro file(s) containing records of the specified schema. 
- readGenericRecords(String) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
Reads Avro file(s) containing records of the specified schema. 
- readKeyPatterns() - Static method in class org.apache.beam.sdk.io.redis.RedisIO
- 
Like  RedisIO.read() but executes multiple instances of the Redis query substituting each
 element of a  PCollection as key pattern. 
- ReadKeyPatterns() - Constructor for class org.apache.beam.sdk.io.redis.RedisIO.ReadKeyPatterns
-  
- readLater() - Method in interface org.apache.beam.sdk.state.BagState
-  
- readLater() - Method in interface org.apache.beam.sdk.state.CombiningState
-  
- readLater() - Method in interface org.apache.beam.sdk.state.GroupingState
-  
- readLater() - Method in interface org.apache.beam.sdk.state.ReadableState
- 
Indicate that the value will be read later. 
- readLater() - Method in interface org.apache.beam.sdk.state.SetState
-  
- readLater() - Method in interface org.apache.beam.sdk.state.ValueState
-  
- readLater() - Method in interface org.apache.beam.sdk.state.WatermarkHoldState
-  
- readMatches() - Static method in class org.apache.beam.sdk.io.FileIO
- 
- ReadMatches() - Constructor for class org.apache.beam.sdk.io.FileIO.ReadMatches
-  
- readMessage() - Static method in class org.apache.beam.sdk.io.jms.JmsIO
-  
- readMessages() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that continuously reads from a Google Cloud Pub/Sub stream. 
- readMessagesWithAttributes() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that continuously reads from a Google Cloud Pub/Sub stream. 
- readMessagesWithAttributesAndMessageId() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that continuously reads from a Google Cloud Pub/Sub stream. 
- readMessagesWithAttributesAndMessageIdAndOrderingKey() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that continuously reads from a Google Cloud Pub/Sub stream. 
- readMessagesWithAttributesWithCoderAndParseFn(Coder<T>, SimpleFunction<PubsubMessage, T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that continuously reads from a Google Cloud Pub/Sub stream,
 mapping each  PubsubMessage, with attributes, into type T using the supplied parse
 function and coder. 
- readMessagesWithCoderAndParseFn(Coder<T>, SimpleFunction<PubsubMessage, T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that continuously reads from a Google Cloud Pub/Sub stream,
 mapping each  PubsubMessage into type T using the supplied parse function and coder. 
- readMessagesWithMessageId() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that continuously reads from a Google Cloud Pub/Sub stream. 
- readNewPartitions() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
-  
- readNewPartitionsIncludingDeleted() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
-  
- readNextBlock() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource.AvroReader
-  
- readNextBlock() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
- 
Read the next block from the input. 
- readNextRecord() - Method in class org.apache.beam.sdk.io.BlockBasedSource.Block
- 
Reads the next record from the block and returns true iff one exists. 
- readNextRecord() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
- 
- readNextRecord() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
- 
Reads the next record via the delegate reader. 
- readNextRecord() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
- 
- readNItemsFromStartedReader(Source.Reader<T>, int) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
- 
Read elements from a Source.Readerthat has already hadSource.Reader#startcalled on it, until n elements are read.
 
- readNItemsFromUnstartedReader(Source.Reader<T>, int) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
- 
Read elements from a Source.Readeruntil n elements are read.
 
- readOnly(String, Map<String, BeamSqlTable>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
- 
This method creates  BeamSqlEnv using empty
 Pipeline Options. 
- ReadOnlyTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider
- 
A ReadOnlyTableProviderprovides in-memory read only set ofBeamSqlTable
 BeamSqlTables.
 
- ReadOnlyTableProvider(String, Map<String, BeamSqlTable>) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.ReadOnlyTableProvider
-  
- ReadOperation - Class in org.apache.beam.sdk.io.gcp.spanner
- 
Encapsulates a spanner read operation. 
- ReadOperation() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- readPrivateKeyFile(String) - Static method in class org.apache.beam.sdk.io.snowflake.KeyPairUtils
-  
- readProtoDynamicMessages(ProtoDomain, String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns a  PTransform that continuously reads binary encoded protobuf messages for the
 type specified by  fullMessageName. 
- readProtoDynamicMessages(Descriptors.Descriptor) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
- readProtos(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that continuously reads binary encoded protobuf messages of the
 given type from a Google Cloud Pub/Sub stream. 
- readRange(Instant, Instant) - Method in interface org.apache.beam.sdk.state.OrderedListState
- 
Read a timestamp-limited subrange of the list. 
- readRangeLater(Instant, Instant) - Method in interface org.apache.beam.sdk.state.OrderedListState
- 
Call to indicate that a specific range will be read from the list, allowing runners to batch
 multiple range reads. 
- readRangesFn() - Method in class org.apache.beam.sdk.io.ReadAllViaFileBasedSource
-  
- readRangesFn() - Method in class org.apache.beam.sdk.io.ReadAllViaFileBasedSourceTransform
-  
- readRangesFn() - Method in class org.apache.beam.sdk.io.ReadAllViaFileBasedSourceWithFilename
-  
- ReadRegistrar() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOTranslation.ReadRegistrar
-  
- ReadRegistrar() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubSubPayloadTranslation.ReadRegistrar
-  
- ReadRegistrar() - Constructor for class org.apache.beam.sdk.io.iceberg.IcebergSchemaTransformTranslation.ReadRegistrar
-  
- ReadRegistrar() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaSchemaTransformTranslation.ReadRegistrar
-  
- ReadRegistrar() - Constructor for class org.apache.beam.sdk.io.kafka.upgrade.KafkaIOTranslation.ReadRegistrar
-  
- readRemainingFromReader(Source.Reader<T>, boolean) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
- 
Read all remaining elements from a Source.Reader.
 
- readResolve() - Method in class org.apache.beam.runners.twister2.translation.wrappers.Twister2BoundedSource
-  
- readResolve() - Method in class org.apache.beam.runners.twister2.translators.functions.AssignWindowsFunction
-  
- readResolve() - Method in class org.apache.beam.runners.twister2.translators.functions.ByteToElemFunction
-  
- readResolve() - Method in class org.apache.beam.runners.twister2.translators.functions.ByteToWindowFunction
-  
- readResolve() - Method in class org.apache.beam.runners.twister2.translators.functions.ByteToWindowFunctionPrimitive
-  
- readResolve() - Method in class org.apache.beam.runners.twister2.translators.functions.DoFnFunction
-  
- readResolve() - Method in class org.apache.beam.runners.twister2.translators.functions.ElemToBytesFunction
-  
- readResolve() - Method in class org.apache.beam.runners.twister2.translators.functions.GroupByWindowFunction
-  
- readResolve() - Method in class org.apache.beam.runners.twister2.translators.functions.internal.SystemReduceFnBuffering
-  
- readResolve() - Method in class org.apache.beam.runners.twister2.translators.functions.MapToTupleFunction
-  
- readResources() - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO
- 
Read resources from a PCollection of resource IDs (e.g. 
- readRows(ReadRowsRequest) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StorageClient
- 
Read rows in the context of a specific read stream. 
- readRows(ReadRowsRequest, String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StorageClient
-  
- readRows(IcebergCatalogConfig) - Static method in class org.apache.beam.sdk.io.iceberg.IcebergIO
-  
- ReadRows() - Constructor for class org.apache.beam.sdk.io.iceberg.IcebergIO.ReadRows
-  
- readRows() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
- 
Read Beam  Rows from a JDBC data source. 
- ReadRows() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
-  
- readRows() - Static method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO
- 
Read Beam  Rows from a SingleStoreDB datasource. 
- readSource(PipelineOptions, TupleTag<T>, DoFn.MultiOutputReceiver, BoundedSource<T>, BigQueryIO.TypedRead.ErrorHandlingParseFn<T>, BadRecordRouter) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-  
- readSourceDescriptors() - Static method in class org.apache.beam.sdk.io.kafka.KafkaIO
- 
- ReadSourceDescriptors() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
-  
- ReadSourceTranslatorBatch<T> - Class in org.apache.beam.runners.twister2.translators.batch
- 
Source translator. 
- ReadSourceTranslatorBatch() - Constructor for class org.apache.beam.runners.twister2.translators.batch.ReadSourceTranslatorBatch
-  
- ReadSourceTranslatorStream<T> - Class in org.apache.beam.runners.twister2.translators.streaming
- 
doc. 
- ReadSourceTranslatorStream() - Constructor for class org.apache.beam.runners.twister2.translators.streaming.ReadSourceTranslatorStream
-  
- ReadSpannerSchema - Class in org.apache.beam.sdk.io.gcp.spanner
- 
This  DoFn reads Cloud Spanner 'information_schema.*' tables to build the  SpannerSchema. 
- ReadSpannerSchema(SpannerConfig, PCollectionView<Dialect>) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.ReadSpannerSchema
- 
Constructor for creating an instance of the ReadSpannerSchema class. 
- ReadSpannerSchema(SpannerConfig, PCollectionView<Dialect>, Set<String>) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.ReadSpannerSchema
- 
Constructor for creating an instance of the ReadSpannerSchema class. 
- readStreamPartitionsWithWatermark() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
- 
Return list of locked StreamPartition and their watermarks. 
- readStrings() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that continuously reads UTF-8 encoded strings from a Google Cloud
 Pub/Sub stream. 
- readStudyMetadata() - Static method in class org.apache.beam.sdk.io.gcp.healthcare.DicomIO
-  
- readTableRows() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
- 
- readTableRowsWithSchema() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
- 
- readTimeout(Integer) - Method in class org.apache.beam.sdk.io.aws2.common.HttpClientConfiguration.Builder
- 
Milliseconds to wait for a read on a socket before an exception is thrown. 
- readTimeout() - Method in class org.apache.beam.sdk.io.aws2.common.HttpClientConfiguration
- 
Milliseconds to wait for a read on a socket before an exception is thrown. 
- readValue - Variable in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
-  
- readValue - Variable in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
-  
- readWithDatumReader(AvroSource.DatumReaderFactory<T>) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
- 
Reads from a BigQuery table or query and returns a  PCollection with one element per
 each row of the table or query result. 
- readWithFilter(String, String) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO
- 
Read all HL7v2 Messages from a single store matching a filter. 
- readWithFilter(ValueProvider<String>, ValueProvider<String>) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO
- 
Read all HL7v2 Messages from a single store matching a filter. 
- readWithMetadata() - Static method in class org.apache.beam.sdk.io.mqtt.MqttIO
-  
- readWithPartitions(TypeDescriptor<PartitionColumnT>) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
- 
Like  JdbcIO.readAll(), but executes multiple instances of the query on the same table
 (subquery) using ranges. 
- readWithPartitions(JdbcReadWithPartitionsHelper<PartitionColumnT>) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
- 
Like  JdbcIO.readAll(), but executes multiple instances of the query on the same table
 (subquery) using ranges. 
- readWithPartitions() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
-  
- ReadWithPartitions() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadWithPartitions
-  
- readWithPartitions() - Static method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO
- 
Like  SingleStoreIO.read(), but executes multiple instances of the query on the same table for each
 database partition. 
- ReadWithPartitions() - Constructor for class org.apache.beam.sdk.io.singlestore.SingleStoreIO.ReadWithPartitions
-  
- readWithPartitionsRows() - Static method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO
- 
Like  SingleStoreIO.readRows(), but executes multiple instances of the query on the same table for each
 database partition. 
- readWithSchema() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO
-  
- ReadWriteRegistrar() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQuerySchemaTransformTranslation.ReadWriteRegistrar
-  
- receive() - Method in interface org.apache.beam.sdk.io.solace.broker.MessageReceiver
- 
Receives a message from the broker. 
- receive() - Method in class org.apache.beam.sdk.io.solace.broker.SolaceMessageReceiver
-  
- ReceiverBuilder<X,T extends org.apache.spark.streaming.receiver.Receiver<X>> - Class in org.apache.beam.sdk.io.sparkreceiver
- 
Class for building an instance for Receiverthat uses Apache Beam mechanisms instead of
 Spark environment.
 
- ReceiverBuilder(Class<T>) - Constructor for class org.apache.beam.sdk.io.sparkreceiver.ReceiverBuilder
-  
- RecommendationAICreateCatalogItem - Class in org.apache.beam.sdk.extensions.ml
- 
A  PTransform using the Recommendations AI API (https://cloud.google.com/recommendations). 
- RecommendationAICreateCatalogItem() - Constructor for class org.apache.beam.sdk.extensions.ml.RecommendationAICreateCatalogItem
-  
- RecommendationAIImportCatalogItems - Class in org.apache.beam.sdk.extensions.ml
- 
A  PTransform connecting to the Recommendations AI API
 (https://cloud.google.com/recommendations) and creating  CatalogItems. 
- RecommendationAIImportCatalogItems() - Constructor for class org.apache.beam.sdk.extensions.ml.RecommendationAIImportCatalogItems
-  
- RecommendationAIImportUserEvents - Class in org.apache.beam.sdk.extensions.ml
- 
A  PTransform connecting to the Recommendations AI API
 (https://cloud.google.com/recommendations) and creating  UserEvents. 
- RecommendationAIImportUserEvents() - Constructor for class org.apache.beam.sdk.extensions.ml.RecommendationAIImportUserEvents
-  
- RecommendationAIIO - Class in org.apache.beam.sdk.extensions.ml
- 
The RecommendationAIIO class acts as a wrapper around the PTransforms that interact with
 the Recommendation AI API (https://cloud.google.com/recommendations).
 
- RecommendationAIIO() - Constructor for class org.apache.beam.sdk.extensions.ml.RecommendationAIIO
-  
- RecommendationAIPredict - Class in org.apache.beam.sdk.extensions.ml
- 
A  PTransform using the Recommendations AI API (https://cloud.google.com/recommendations). 
- RecommendationAIPredict() - Constructor for class org.apache.beam.sdk.extensions.ml.RecommendationAIPredict
-  
- RecommendationAIWriteUserEvent - Class in org.apache.beam.sdk.extensions.ml
- 
A  PTransform using the Recommendations AI API (https://cloud.google.com/recommendations). 
- RecommendationAIWriteUserEvent() - Constructor for class org.apache.beam.sdk.extensions.ml.RecommendationAIWriteUserEvent
-  
- record(List<FieldOperation>) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- RECORD - Static variable in class org.apache.beam.sdk.io.gcp.bigquery.providers.PortableBigQueryDestinations
-  
- Record() - Constructor for class org.apache.beam.sdk.io.solace.data.Solace.Record
-  
- Record() - Constructor for class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Record
-  
- RECORD_NUM - Static variable in class org.apache.beam.sdk.io.contextualtextio.RecordWithMetadata
-  
- RECORD_NUM_IN_OFFSET - Static variable in class org.apache.beam.sdk.io.contextualtextio.RecordWithMetadata
-  
- RECORD_OFFSET - Static variable in class org.apache.beam.sdk.io.contextualtextio.RecordWithMetadata
-  
- RecordAggregation() - Constructor for class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.RecordAggregation
-  
- recordId() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.IncomingMessage
- 
Id to pass to the runner to distinguish this message from all others. 
- recordId() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.OutgoingMessage
- 
If using an id attribute, the record id to associate with this record's metadata so the
 receiver can reject duplicates. 
- RECORDING_ROUTER - Static variable in interface org.apache.beam.sdk.transforms.errorhandling.BadRecordRouter
-  
- RecordingBadRecordRouter() - Constructor for class org.apache.beam.sdk.transforms.errorhandling.BadRecordRouter.RecordingBadRecordRouter
-  
- RecordToPublishResultDoFn - Class in org.apache.beam.sdk.io.solace.write
- 
This class just transforms to PublishResult to be able to capture the windowing with the right
 strategy. 
- RecordToPublishResultDoFn() - Constructor for class org.apache.beam.sdk.io.solace.write.RecordToPublishResultDoFn
-  
- RecordWithMetadata - Class in org.apache.beam.sdk.io.contextualtextio
- 
Helper Class based on  Row, it provides Metadata associated with each Record when reading
 from file(s) using  ContextualTextIO. 
- RecordWithMetadata() - Constructor for class org.apache.beam.sdk.io.contextualtextio.RecordWithMetadata
-  
- recoverRecords(Consumer<HistoryRecord>) - Method in class org.apache.beam.io.debezium.KafkaSourceConsumerFn.DebeziumSDFDatabaseHistory
-  
- RedisConnectionConfiguration - Class in org.apache.beam.sdk.io.redis
- 
RedisConnectionConfigurationdescribes and wraps a connectionConfiguration to Redis
 server or cluster.
 
- RedisConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
-  
- RedisCursor - Class in org.apache.beam.sdk.io.redis
-  
- RedisIO - Class in org.apache.beam.sdk.io.redis
- 
An IO to manipulate Redis key/value database. 
- RedisIO.Read - Class in org.apache.beam.sdk.io.redis
- 
- RedisIO.ReadKeyPatterns - Class in org.apache.beam.sdk.io.redis
- 
- RedisIO.Write - Class in org.apache.beam.sdk.io.redis
- 
- RedisIO.Write.Method - Enum in org.apache.beam.sdk.io.redis
- 
Determines the method used to insert data in Redis. 
- RedisIO.WriteStreams - Class in org.apache.beam.sdk.io.redis
- 
A  PTransform to write stream key pairs (https://redis.io/topics/streams-intro) to a
 Redis server. 
- Redistribute - Class in org.apache.beam.sdk.transforms
- 
A family of  PTransforms that returns a  PCollection equivalent to its
 input but functions as an operational hint to a runner that redistributing the data in some way
 is likely useful. 
- Redistribute() - Constructor for class org.apache.beam.sdk.transforms.Redistribute
-  
- Redistribute.RedistributeArbitrarily<T> - Class in org.apache.beam.sdk.transforms
- 
Noop transform that hints to the runner to try to redistribute the work evenly, or via whatever
 clever strategy the runner comes up with. 
- Redistribute.RedistributeByKey<K,V> - Class in org.apache.beam.sdk.transforms
-  
- Redistribute.Registrar - Class in org.apache.beam.sdk.transforms
- 
Registers translators for the Redistribute family of transforms. 
- ReferenceCountingExecutableStageContextFactory - Class in org.apache.beam.runners.fnexecution.control
- 
- ReferenceCountingExecutableStageContextFactory.Creator - Interface in org.apache.beam.runners.fnexecution.control
- 
Interface for creator which extends Serializable. 
- referencesSingleField() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Returns true if this descriptor references only a single, non-wildcard field. 
- reflect(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns an  AvroCoder instance for the provided element type respecting Avro's Reflect*
 suite for encoding and decoding. 
- reflect(Class<T>) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns an  AvroCoder instance for the provided element type respecting Avro's Reflect*
 suite for encoding and decoding. 
- reflect(Class<T>, Schema) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns an  AvroCoder instance for the provided element type respecting Avro's Reflect*
 suite for encoding and decoding. 
- reflect(Class<T>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory
- 
Returns an  AvroDatumFactory instance for the provided element type respecting Avro's
 Reflect* suite for encoding and decoding. 
- ReflectDatumFactory(Class<T>) - Constructor for class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory.ReflectDatumFactory
-  
- ReflectUtils - Class in org.apache.beam.sdk.schemas.utils
- 
A set of reflection helper methods. 
- ReflectUtils() - Constructor for class org.apache.beam.sdk.schemas.utils.ReflectUtils
-  
- ReflectUtils.ClassWithSchema - Class in org.apache.beam.sdk.schemas.utils
- 
Represents a class and a schema. 
- ReflectUtils.TypeDescriptorWithSchema<T> - Class in org.apache.beam.sdk.schemas.utils
- 
Represents a type descriptor and a schema. 
- refreshSchema(TableReference, BigQueryServices.DatasetService) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableSchemaCache
-  
- refreshThread() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableSchemaCache
-  
- Regex - Class in org.apache.beam.sdk.transforms
- 
PTransforms to use Regular Expressions to process elements in a  PCollection. 
- Regex.AllMatches - Class in org.apache.beam.sdk.transforms
- 
Regex.MatchesName<String>takes aPCollection<String>and returns aPCollection<List<String>>representing the value extracted from all the Regex groups of the
 inputPCollectionto the number of times that element occurs in the input.
 
- Regex.Find - Class in org.apache.beam.sdk.transforms
- 
Regex.Find<String>takes aPCollection<String>and returns aPCollection<String>representing the value extracted from the Regex groups of the inputPCollectionto the number of times that element occurs in the input.
 
- Regex.FindAll - Class in org.apache.beam.sdk.transforms
- 
Regex.Find<String>takes aPCollection<String>and returns aPCollection<List<String>>representing the value extracted from the Regex groups of the inputPCollectionto the number of times that element occurs in the input.
 
- Regex.FindKV - Class in org.apache.beam.sdk.transforms
- 
Regex.MatchesKV<KV<String, String>>takes aPCollection<String>and returns aPCollection<KV<String, String>>representing the key and value extracted from the Regex
 groups of the inputPCollectionto the number of times that element occurs in the
 input.
 
- Regex.FindName - Class in org.apache.beam.sdk.transforms
- 
Regex.Find<String>takes aPCollection<String>and returns aPCollection<String>representing the value extracted from the Regex groups of the inputPCollectionto the number of times that element occurs in the input.
 
- Regex.FindNameKV - Class in org.apache.beam.sdk.transforms
- 
Regex.MatchesKV<KV<String, String>>takes aPCollection<String>and returns aPCollection<KV<String, String>>representing the key and value extracted from the Regex
 groups of the inputPCollectionto the number of times that element occurs in the
 input.
 
- Regex.Matches - Class in org.apache.beam.sdk.transforms
- 
Regex.Matches<String>takes aPCollection<String>and returns aPCollection<String>representing the value extracted from the Regex groups of the inputPCollectionto the number of times that element occurs in the input.
 
- Regex.MatchesKV - Class in org.apache.beam.sdk.transforms
- 
Regex.MatchesKV<KV<String, String>>takes aPCollection<String>and returns aPCollection<KV<String, String>>representing the key and value extracted from the Regex
 groups of the inputPCollectionto the number of times that element occurs in the
 input.
 
- Regex.MatchesName - Class in org.apache.beam.sdk.transforms
- 
Regex.MatchesName<String>takes aPCollection<String>and returns aPCollection<String>representing the value extracted from the Regex groups of the inputPCollectionto the number of times that element occurs in the input.
 
- Regex.MatchesNameKV - Class in org.apache.beam.sdk.transforms
- 
Regex.MatchesNameKV<KV<String, String>>takes aPCollection<String>and returns
 aPCollection<KV<String, String>>representing the key and value extracted from the
 Regex groups of the inputPCollectionto the number of times that element occurs in the
 input.
 
- Regex.ReplaceAll - Class in org.apache.beam.sdk.transforms
- 
Regex.ReplaceAll<String>takes aPCollection<String>and returns aPCollection<String>with all Strings that matched the Regex being replaced with the
 replacement string.
 
- Regex.ReplaceFirst - Class in org.apache.beam.sdk.transforms
- 
Regex.ReplaceFirst<String>takes aPCollection<String>and returns aPCollection<String>with the first Strings that matched the Regex being replaced with the
 replacement string.
 
- Regex.Split - Class in org.apache.beam.sdk.transforms
- 
Regex.Split<String>takes aPCollection<String>and returns aPCollection<String>with the input string split into individual items in a list.
 
- RegexMatcher - Class in org.apache.beam.sdk.testing
- 
Hamcrest matcher to assert a string matches a pattern. 
- RegexMatcher(String) - Constructor for class org.apache.beam.sdk.testing.RegexMatcher
-  
- region(Region) - Method in class org.apache.beam.sdk.io.aws2.common.ClientConfiguration.Builder
- 
- region() - Method in class org.apache.beam.sdk.io.aws2.common.ClientConfiguration
- 
- register(WatchService, WatchEvent.Kind<?>[], WatchEvent.Modifier...) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- register(WatchService, WatchEvent.Kind<?>...) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- register(RelOptPlanner) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
-  
- register(RelOptPlanner) - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-  
- register(Class<? extends PipelineOptions>) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
- 
This registers the interface with this factory. 
- registerBadRecordErrorHandler(PTransform<PCollection<BadRecord>, OutputT>) - Method in class org.apache.beam.sdk.Pipeline
-  
- registerByteSizeObserver(T, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.Coder
- 
Notifies the ElementByteSizeObserverabout the byte size of the encoded value using
 thisCoder.
 
- registerByteSizeObserver(ReadableDuration, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.DurationCoder
-  
- registerByteSizeObserver(IterableT, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
-  
- registerByteSizeObserver(KV<K, V>, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.KvCoder
- 
Notifies ElementByteSizeObserver about the byte size of the encoded value using this coder. 
- registerByteSizeObserver(Map<K, V>, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.MapCoder
-  
- registerByteSizeObserver(T, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.NullableCoder
- 
Overridden to short-circuit the default StructuredCoderbehavior of encoding and
 counting the bytes.
 
- registerByteSizeObserver(SortedMap<K, V>, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.SortedMapCoder
-  
- registerByteSizeObserver(T, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.TimestampPrefixingWindowCoder
-  
- registerByteSizeObserver(RawUnionValue, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
- 
Notifies ElementByteSizeObserver about the byte size of the encoded value using this coder. 
- registerByteSizeObserver(IntervalWindow, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
-  
- registerClasses(Kryo) - Method in class org.apache.beam.runners.spark.coders.SparkRunnerKryoRegistrator
-  
- registerClasses(Kryo) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.SparkSessionFactory.SparkKryoRegistrator
-  
- registerCoderForClass(Class<?>, Coder<?>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
- 
Registers the provided  Coder for the given class. 
- registerCoderForType(TypeDescriptor<?>, Coder<?>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
- 
Registers the provided  Coder for the given type. 
- registerCoderProvider(CoderProvider) - Method in class org.apache.beam.sdk.coders.CoderRegistry
- 
Registers  coderProvider as a potential  CoderProvider which can produce  Coder instances. 
- registerConsumer(String, CloseableFnDataReceiver<BeamFnApi.Elements>) - Method in class org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer
- 
Registers a consumer for the specified instruction id. 
- registerEnvironment(String, RunnerApi.Environment) - Method in class org.apache.beam.runners.fnexecution.provisioning.StaticGrpcProvisionService
-  
- registerFileSystemsOnce(PipelineOptions) - Static method in class org.apache.beam.sdk.io.FileSystems
- 
Register file systems once if never done before. 
- registerForProcessBundleInstructionId(String, StateRequestHandler) - Method in class org.apache.beam.runners.fnexecution.state.GrpcStateService
-  
- registerForProcessBundleInstructionId(String, StateRequestHandler) - Method in interface org.apache.beam.runners.fnexecution.state.StateDelegator
- 
Registers the supplied handler for the given process bundle instruction id for all BeamFnApi.StateRequests with a matching id.
 
- registerJavaBean(Class<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
Register a JavaBean type for automatic schema inference. 
- registerJavaBean(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
Register a JavaBean type for automatic schema inference. 
- registerJob(String, Map<String, List<RunnerApi.ArtifactInformation>>) - Method in class org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService
- 
Registers a set of artifacts to be staged with this service. 
- registerKnownTableNames(List<TableName>) - Method in interface org.apache.beam.sdk.extensions.sql.meta.CustomTableResolver
- 
Register the table names as extracted from the FROM clause. 
- registerKnownTableNames(List<TableName>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.FullNameTableProvider
-  
- registerLineage(String, Schema) - Method in class org.apache.beam.sdk.io.cdap.context.StreamingSourceContextImpl
-  
- registerMetricsForPipelineResult() - Method in class org.apache.beam.runners.flink.metrics.FlinkMetricContainer
- 
This should be called at the end of the Flink job and sets up an accumulator to push the
 metrics to the PipelineResult. 
- registerOutputDataLocation(String, Coder<T>) - Method in class org.apache.beam.sdk.fn.data.BeamFnDataOutboundAggregator
- 
Register the outbound data logical endpoint, returns the FnDataReceiver for processing the
 endpoint's outbound data. 
- registerOutputTimersLocation(String, String, Coder<T>) - Method in class org.apache.beam.sdk.fn.data.BeamFnDataOutboundAggregator
- 
Register the outbound timers logical endpoint, returns the FnDataReceiver for processing the
 endpoint's outbound timers data. 
- registerPOJO(Class<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
Register a POJO type for automatic schema inference. 
- registerPOJO(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
Register a POJO type for automatic schema inference. 
- registerProcessBundleDescriptor(BeamFnApi.ProcessBundleDescriptor) - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClient
-  
- registerProcessBundleDescriptor(BeamFnApi.ProcessBundleDescriptor) - Method in interface org.apache.beam.runners.fnexecution.control.InstructionRequestHandler
-  
- registerProvider(TableProvider) - Method in class org.apache.beam.sdk.extensions.sql.meta.store.InMemoryMetaStore
-  
- registerProvider(TableProvider) - Method in interface org.apache.beam.sdk.extensions.sql.meta.store.MetaStore
- 
Register a table provider. 
- registerReceiver(String, CloseableFnDataReceiver<BeamFnApi.Elements>) - Method in interface org.apache.beam.runners.fnexecution.data.FnDataService
- 
Registers a receiver for the provided instruction id. 
- registerReceiver(String, CloseableFnDataReceiver<BeamFnApi.Elements>) - Method in class org.apache.beam.runners.fnexecution.data.GrpcDataService
-  
- registerSchemaForClass(Class<T>, Schema, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
Register a schema for a specific Classtype.
 
- registerSchemaForType(TypeDescriptor<T>, Schema, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
- registerSchemaProvider(SchemaProvider) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
- registerSchemaProvider(Class<T>, SchemaProvider) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
- registerSchemaProvider(TypeDescriptor<T>, SchemaProvider) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
- 
- registerTables(SchemaPlus, List<List<String>>) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.TableResolution
- 
Registers tables that will be resolved during query analysis, so table providers can eagerly
 pre-load metadata. 
- registerTransformTranslator(Class<TransformT>, TransformTranslator<? extends TransformT>) - Static method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator
- 
Records that instances of the specified PTransform class should be translated by default by the
 corresponding  TransformTranslator. 
- registerUdaf(String, Combine.CombineFn) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
- 
register a Combine.CombineFnas UDAF function used in this query.
 
- registerUdf(String, Class<? extends BeamSqlUdf>) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
- 
register a UDF function used in this query. 
- registerUdf(String, SerializableFunction) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
- 
- Registrar() - Constructor for class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.Registrar
-  
- Registrar() - Constructor for class org.apache.beam.sdk.transforms.Redistribute.Registrar
-  
- Reify - Class in org.apache.beam.sdk.transforms
- 
PTransforms for converting between explicit and implicit form of various Beam
 values. 
- ReifyAsIterable<T> - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
This transforms turns a side input into a singleton PCollection that can be used as the main
 input for another transform. 
- ReifyAsIterable() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.ReifyAsIterable
-  
- rel(int) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
-  
- relativeErrorForPrecision(int) - Static method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct
-  
- relativeFileNaming(ValueProvider<String>, FileIO.Write.FileNaming) - Static method in class org.apache.beam.sdk.io.FileIO.Write
-  
- relativize(Path) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- relBuilder - Variable in class org.apache.beam.sdk.extensions.sql.impl.rel.CalcRelSplitter
-  
- release(ObjectT) - Method in class org.apache.beam.sdk.io.aws2.common.ObjectPool
- 
Release a reference to a shared client instance. 
- releaseByKey(KeyT) - Method in class org.apache.beam.sdk.io.aws2.common.ObjectPool
- 
Release a reference to a shared object instance using KeyT.
 
- releaseDataset(Dataset) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- releaseJobIdLock(Configuration) - Method in interface org.apache.beam.sdk.io.hadoop.format.ExternalSynchronization
- 
Deletes lock ids bounded with given job if any exists. 
- releaseJobIdLock(Configuration) - Method in class org.apache.beam.sdk.io.hadoop.format.HDFSSynchronization
-  
- releaseSavepoint(Savepoint) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- releaseStreamPartitionLockForDeletion(Range.ByteStringRange, String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
- 
This is the 1st step of 2 phase delete of StreamPartition. 
- RelMdNodeStats - Class in org.apache.beam.sdk.extensions.sql.impl.planner
- 
This is the implementation of NodeStatsMetadata. 
- RelMdNodeStats() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.planner.RelMdNodeStats
-  
- RelType(String) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.CalcRelSplitter.RelType
-  
- remerge() - Static method in class org.apache.beam.sdk.transforms.windowing.Window
- 
Creates a  WindowPTransform that does not change assigned windows, but will
 cause windows to be merged again as part of the next  GroupByKey. 
- RemoteBundle - Interface in org.apache.beam.runners.fnexecution.control
- 
A bundle capable of handling input data elements for a bundle descriptorby
 forwarding them to a remote environment for processing.
 
- RemoteEnvironment - Interface in org.apache.beam.runners.fnexecution.environment
- 
A handle to an available remote RunnerApi.Environment.
 
- RemoteEnvironment.SimpleRemoteEnvironment - Class in org.apache.beam.runners.fnexecution.environment
- 
- RemoteEnvironmentOptions - Interface in org.apache.beam.sdk.options
- 
Options that are used to control configuration of the remote environment. 
- RemoteEnvironmentOptions.Options - Class in org.apache.beam.sdk.options
- 
- RemoteGrpcPortRead - Class in org.apache.beam.sdk.fn.data
- 
An execution-time only RunnerApi.PTransformwhich represents an SDK harness reading from aBeamFnApi.RemoteGrpcPort.
 
- RemoteGrpcPortRead() - Constructor for class org.apache.beam.sdk.fn.data.RemoteGrpcPortRead
-  
- RemoteGrpcPortWrite - Class in org.apache.beam.sdk.fn.data
- 
An execution-time only RunnerApi.PTransformwhich represents a write from within an SDK harness to
 aBeamFnApi.RemoteGrpcPort.
 
- RemoteGrpcPortWrite() - Constructor for class org.apache.beam.sdk.fn.data.RemoteGrpcPortWrite
-  
- RemoteInputDestination<T> - Class in org.apache.beam.runners.fnexecution.data
- 
A pair of  Coder and  BeamFnApi.Target which specifies the arguments to a  FnDataService to send data to a remote harness. 
- RemoteInputDestination() - Constructor for class org.apache.beam.runners.fnexecution.data.RemoteInputDestination
-  
- RemoteOutputReceiver<T> - Class in org.apache.beam.runners.fnexecution.control
- 
- RemoteOutputReceiver() - Constructor for class org.apache.beam.runners.fnexecution.control.RemoteOutputReceiver
-  
- remove(Collection<String>) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
-  
- remove() - Method in class org.apache.beam.sdk.fn.stream.DataStreams.DataStreamDecoder
-  
- remove(Object) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
-  
- remove(K) - Method in interface org.apache.beam.sdk.state.MapState
- 
Remove the mapping for a key from this map if it is present. 
- remove(K) - Method in interface org.apache.beam.sdk.state.MultimapState
- 
Removes all values associated with the key from this multimap. 
- remove(T) - Method in interface org.apache.beam.sdk.state.SetState
- 
Removes the specified element from this set if it is present. 
- removeAllPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- removeBucket(Bucket) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
- 
Remove an empty  Bucket in Cloud Storage or propagates an exception. 
- removeMetadata(MetadataEntity) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- removePipelineOption(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- removePrefix() - Method in class org.apache.beam.sdk.extensions.sql.impl.TableName
- 
Remove prefix, e.g. 
- removeProperties(MetadataEntity) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- removeProperties(MetadataEntity, String...) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- removeStagedArtifacts(String) - Method in interface org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService.ArtifactDestinationProvider
-  
- removeStagedArtifacts(String) - Method in class org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService
-  
- removeTags(MetadataEntity) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- removeTags(MetadataEntity, String...) - Method in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
-  
- removeTemporaryFiles(Collection<ResourceId>) - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-  
- rename(Iterable<String>, Iterable<String>, MoveOptions...) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
-  
- rename(List<ClassLoaderFileSystem.ClassLoaderResourceId>, List<ClassLoaderFileSystem.ClassLoaderResourceId>, MoveOptions...) - Method in class org.apache.beam.sdk.io.ClassLoaderFileSystem
-  
- rename(List<ResourceIdT>, List<ResourceIdT>, MoveOptions...) - Method in class org.apache.beam.sdk.io.FileSystem
- 
Renames a Listof file-like resources from one location to another.
 
- rename(List<ResourceId>, List<ResourceId>, MoveOptions...) - Static method in class org.apache.beam.sdk.io.FileSystems
- 
Renames a Listof file-like resources from one location to another.
 
- rename(String, String) - Method in class org.apache.beam.sdk.schemas.transforms.RenameFields.Inner
- 
Rename a specific field. 
- rename(FieldAccessDescriptor, String) - Method in class org.apache.beam.sdk.schemas.transforms.RenameFields.Inner
- 
Rename a specific field. 
- RenameFields - Class in org.apache.beam.sdk.schemas.transforms
- 
A transform for renaming fields inside an existing schema. 
- RenameFields() - Constructor for class org.apache.beam.sdk.schemas.transforms.RenameFields
-  
- RenameFields.Inner<T> - Class in org.apache.beam.sdk.schemas.transforms
- 
The class implementing the actual PTransform. 
- REPEATABLE_ERROR_TYPES - Static variable in class org.apache.beam.io.requestresponse.RequestResponseIO
- 
- Repeatedly - Class in org.apache.beam.sdk.transforms.windowing
- 
A  Trigger that fires according to its subtrigger forever. 
- replace(Class<V>, T) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.utils.ScalaInterop
-  
- replace(String, String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- REPLACE - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- REPLACE_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.BeamBuiltinMethods
-  
- REPLACE_URN - Static variable in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar
-  
- replaceAll(List<PTransformOverride>) - Method in class org.apache.beam.sdk.Pipeline
- 
For internal use only; no backwards-compatibility guarantees. 
- replaceAll(String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
Returns a  Regex.ReplaceAllPTransform that checks if a portion of the line
 matches the Regex and replaces all matches with the replacement String. 
- replaceAll(Pattern, String) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
Returns a  Regex.ReplaceAllPTransform that checks if a portion of the line
 matches the Regex and replaces all matches with the replacement String. 
- ReplaceAll(Pattern, String) - Constructor for class org.apache.beam.sdk.transforms.Regex.ReplaceAll
-  
- ReplaceBuilder() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.ReplaceBuilder
-  
- replaceFirst(String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
Returns a  Regex.ReplaceAllPTransform that checks if a portion of the line
 matches the Regex and replaces the first match with the replacement String. 
- replaceFirst(Pattern, String) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
Returns a  Regex.ReplaceAllPTransform that checks if a portion of the line
 matches the Regex and replaces the first match with the replacement String. 
- ReplaceFirst(Pattern, String) - Constructor for class org.apache.beam.sdk.transforms.Regex.ReplaceFirst
-  
- replaceGcsFilesWithLocalFiles(List<String>) - Static method in class org.apache.beam.runners.dataflow.DataflowRunner
- 
Replaces GCS file paths with local file paths by downloading the GCS files locally. 
- replaceTransforms(Pipeline, StreamingOptions) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
-  
- replaceV1Transforms(Pipeline) - Method in class org.apache.beam.runners.dataflow.DataflowRunner
-  
- ReplicaInfo() - Constructor for class org.apache.beam.sdk.io.solr.SolrIO.ReplicaInfo
-  
- report() - Method in class org.apache.beam.runners.spark.metrics.sink.CsvSink
-  
- report() - Method in class org.apache.beam.runners.spark.metrics.sink.GraphiteSink
-  
- report() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.sink.CodahaleCsvSink
-  
- report() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.sink.CodahaleGraphiteSink
-  
- reportElementSize(long) - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
-  
- reportError(String, Throwable) - Method in class org.apache.beam.sdk.io.sparkreceiver.WrappedSupervisor
-  
- reportFailedRPCMetrics(RetryManager.Operation.Context<?>, BigQuerySinkMetrics.RpcMethod) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySinkMetrics
- 
Records RpcRequests counter and RpcLatency histogram for this RPC call. 
- reportFailedRPCMetrics(RetryManager.Operation.Context<?>, BigQuerySinkMetrics.RpcMethod, String) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySinkMetrics
- 
Records RpcRequests counter and RpcLatency histogram for this RPC call. 
- reportLineage(ResourceIdT, Lineage) - Method in class org.apache.beam.sdk.io.FileSystem
- 
Report  Lineage metrics for resource id at file level. 
- reportLineage(ResourceIdT, Lineage, FileSystem.LineageLevel) - Method in class org.apache.beam.sdk.io.FileSystem
- 
Report  Lineage metrics for resource id to a given level. 
- reportPendingMetrics() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StorageClient
- 
Call this method on Work Item thread to report outstanding metrics. 
- reportSinkLineage(ResourceId) - Static method in class org.apache.beam.sdk.io.FileSystems
- 
Report sink  Lineage metrics for resource id. 
- reportSinkLineage(ResourceId, FileSystem.LineageLevel) - Static method in class org.apache.beam.sdk.io.FileSystems
- 
Report source  Lineage metrics for resource id at given level. 
- reportSourceLineage(ResourceId) - Static method in class org.apache.beam.sdk.io.FileSystems
- 
Report source  Lineage metrics for resource id. 
- reportSourceLineage(ResourceId, FileSystem.LineageLevel) - Static method in class org.apache.beam.sdk.io.FileSystems
- 
Report source  Lineage metrics for resource id at given level. 
- reportSuccessfulRpcMetrics(RetryManager.Operation.Context<?>, BigQuerySinkMetrics.RpcMethod) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySinkMetrics
- 
Records RpcRequests counter and RpcLatency histogram for this RPC call. 
- reportSuccessfulRpcMetrics(RetryManager.Operation.Context<?>, BigQuerySinkMetrics.RpcMethod, String) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySinkMetrics
- 
Records RpcRequests counter and RpcLatency histogram for this RPC call. 
- reportWorkItemStatus(String, ReportWorkItemStatusRequest) - Method in class org.apache.beam.runners.dataflow.DataflowClient
- 
Reports the status of the work item for jobId.
 
- requestProgress() - Method in interface org.apache.beam.runners.fnexecution.control.RemoteBundle
- 
Ask the remote bundle for progress. 
- requestProgress() - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor.ActiveBundle
-  
- RequestResponseIO<RequestT,ResponseT> - Class in org.apache.beam.io.requestresponse
- 
PTransform for reading from and writing to Web APIs. 
- requestsFinalization(String) - Method in interface org.apache.beam.runners.fnexecution.control.BundleFinalizationHandler
- 
This callback is invoked whenever an inflight bundle that is being processed requests
 finalization. 
- requestsFinalization(String) - Method in class org.apache.beam.runners.fnexecution.control.BundleFinalizationHandlers.InMemoryFinalizer
-  
- requestsInProgress() - Method in class org.apache.beam.sdk.io.aws2.common.AsyncBatchWriteHandler
-  
- requestTimeMsSinceEpoch() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.IncomingMessage
- 
Timestamp (in system time) at which we requested the message (ms since epoch). 
- REQUIRED_MEMORY_FOR_DEFAULT_BUFFER_BYTES - Static variable in class org.apache.beam.sdk.io.aws2.options.S3Options.S3UploadBufferSizeBytesFactory
-  
- Requirements - Class in org.apache.beam.sdk.transforms
- 
Describes the run-time requirements of a  Contextful, such as access to side inputs. 
- requiresDataSchema() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSchemaIOProvider
-  
- requiresDataSchema() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySchemaIOProvider
- 
Indicates whether this transform requires a specified data schema. 
- requiresDataSchema() - Method in class org.apache.beam.sdk.io.gcp.datastore.DataStoreV1SchemaIOProvider
-  
- requiresDataSchema() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubSchemaIOProvider
-  
- requiresDataSchema() - Method in class org.apache.beam.sdk.io.jdbc.JdbcSchemaIOProvider
-  
- requiresDataSchema() - Method in interface org.apache.beam.sdk.schemas.io.SchemaIOProvider
- 
Indicates whether the dataSchema value is necessary. 
- requiresDeduping() - Method in class org.apache.beam.sdk.io.solace.read.UnboundedSolaceSource
-  
- requiresDeduping() - Method in class org.apache.beam.sdk.io.UnboundedSource
- 
Returns whether this source requires explicit deduping. 
- requiresSideInputs(Collection<PCollectionView<?>>) - Static method in class org.apache.beam.sdk.transforms.Requirements
- 
Describes the need for access to the given side inputs. 
- requiresSideInputs(PCollectionView<?>...) - Static method in class org.apache.beam.sdk.transforms.Requirements
- 
- reset() - Method in class org.apache.beam.runners.spark.metrics.MetricsContainerStepMapAccumulator
-  
- reset() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsAccumulator
-  
- reset() - Method in class org.apache.beam.sdk.fn.CancellableQueue
- 
Enables the queue to be re-used after it has been cancelled. 
- reset() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
- 
Enables this receiver to be used again for another bundle. 
- reset() - Method in class org.apache.beam.sdk.io.aws2.common.AsyncBatchWriteHandler
-  
- reset() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
-  
- resetCache() - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
- 
Resets the set of interfaces registered with this factory to the default state. 
- resetForNewKey() - Method in class org.apache.beam.runners.fnexecution.state.InMemoryBagUserStateFactory
- 
Prepares previous emitted state handlers for processing a new key. 
- resetLocal() - Method in class org.apache.beam.runners.flink.metrics.MetricsAccumulator
-  
- resetTo(int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
- 
Resets the end of the stream to the specified position. 
- Reshuffle<K,V> - Class in org.apache.beam.sdk.transforms
- 
For internal use only; no backwards compatibility guarantees. 
- Reshuffle.AssignShardFn<T> - Class in org.apache.beam.sdk.transforms
-  
- Reshuffle.ViaRandomKey<T> - Class in org.apache.beam.sdk.transforms
- 
- ReshuffleTrigger<W extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
- 
For internal use only; no backwards compatibility guarantees. 
- ReshuffleTrigger() - Constructor for class org.apache.beam.sdk.transforms.windowing.ReshuffleTrigger
-  
- resolve(Supplier<PipelineOptions>, Dataset<WindowedValue<InT>>) - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator.UnresolvedTranslation
-  
- resolve(String, ResolveOptions) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
-  
- resolve(Path) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- resolve(String) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- resolve(String, ResolveOptions) - Method in class org.apache.beam.sdk.io.ClassLoaderFileSystem.ClassLoaderResourceId
-  
- resolve(String, ResolveOptions) - Method in interface org.apache.beam.sdk.io.fs.ResourceId
- 
Returns a child ResourceIdunderthis.
 
- resolve(Schema) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
- resolveAlias(ResolvedColumn) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
-  
- resolveArtifacts(RunnerApi.Pipeline) - Method in class org.apache.beam.runners.dataflow.DataflowRunner
-  
- resolveArtifacts(ArtifactApi.ResolveArtifactsRequest, StreamObserver<ArtifactApi.ResolveArtifactsResponse>) - Method in class org.apache.beam.runners.fnexecution.artifact.ArtifactRetrievalService
-  
- resolveCalciteTable(SchemaPlus, List<String>) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.TableResolution
- 
Resolves tablePathaccording to the givenschemaPlus.
 
- resolvedTables - Variable in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
-  
- ResolveOptions - Interface in org.apache.beam.sdk.io.fs
- 
- ResolveOptions.StandardResolveOptions - Enum in org.apache.beam.sdk.io.fs
- 
Defines the standard resolve options. 
- resolveSibling(Path) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- resolveSibling(String) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- resolveType(Type) - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
Returns a  TypeDescriptor representing the given type, with type variables resolved
 according to the specialization in this type. 
- RESOURCE_HINTS - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- RESOURCE_ID - Static variable in class org.apache.beam.sdk.io.contextualtextio.RecordWithMetadata
-  
- ResourceHint - Class in org.apache.beam.sdk.transforms.resourcehints
- 
Provides a definition of a resource hint known to the SDK. 
- ResourceHint() - Constructor for class org.apache.beam.sdk.transforms.resourcehints.ResourceHint
-  
- resourceHints - Variable in class org.apache.beam.sdk.transforms.PTransform
-  
- ResourceHints - Class in org.apache.beam.sdk.transforms.resourcehints
- 
Pipeline authors can use resource hints to provide additional information to runners about the
 desired aspects of the execution environment. 
- ResourceHintsOptions - Interface in org.apache.beam.sdk.transforms.resourcehints
- 
Options that are used to control configuration of the remote environment. 
- ResourceHintsOptions.EmptyListDefault - Class in org.apache.beam.sdk.transforms.resourcehints
-  
- ResourceHintsOptions.Options - Class in org.apache.beam.sdk.transforms.resourcehints
- 
- resourceId() - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata
-  
- ResourceId - Interface in org.apache.beam.sdk.io.fs
- 
An identifier which represents a file-like resource. 
- ResourceIdCoder - Class in org.apache.beam.sdk.io.fs
- 
- ResourceIdCoder() - Constructor for class org.apache.beam.sdk.io.fs.ResourceIdCoder
-  
- ResourceIdTester - Class in org.apache.beam.sdk.io.fs
- 
- responseReceivedEx(Object) - Method in class org.apache.beam.sdk.io.solace.broker.PublishResultHandler
-  
- RESTRICTION_CODER - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
- 
- RESTRICTION_ENCODING - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- RestrictionInterrupter<T> - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction
- 
An interrupter for restriction tracker of type T. 
- restrictionTracker(OffsetRange) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.DetectNewPartitionsDoFn
-  
- restrictionTracker(StreamProgress) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.ReadChangeStreamPartitionDoFn
-  
- restrictionTracker(PulsarSourceDescriptor, OffsetRange) - Method in class org.apache.beam.sdk.io.pulsar.ReadFromPulsarDoFn
-  
- RestrictionTracker<RestrictionT,PositionT> - Class in org.apache.beam.sdk.transforms.splittabledofn
- 
Manages access to the restriction and keeps track of its claimed part for a  splittableDoFn. 
- RestrictionTracker() - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker
-  
- RestrictionTracker.HasProgress - Interface in org.apache.beam.sdk.transforms.splittabledofn
- 
All  RestrictionTrackers SHOULD implement this interface to improve auto-scaling and
 splitting performance. 
- RestrictionTracker.IsBounded - Enum in org.apache.beam.sdk.transforms.splittabledofn
-  
- RestrictionTracker.Progress - Class in org.apache.beam.sdk.transforms.splittabledofn
- 
A representation for the amount of known completed and remaining work. 
- RestrictionTracker.TruncateResult<RestrictionT> - Class in org.apache.beam.sdk.transforms.splittabledofn
- 
A representation of the truncate result. 
- RestrictionTrackers - Class in org.apache.beam.sdk.fn.splittabledofn
- 
- RestrictionTrackers() - Constructor for class org.apache.beam.sdk.fn.splittabledofn.RestrictionTrackers
-  
- RestrictionTrackers.ClaimObserver<PositionT> - Interface in org.apache.beam.sdk.fn.splittabledofn
- 
- Result<ResponseT> - Class in org.apache.beam.io.requestresponse
- 
- Result() - Constructor for class org.apache.beam.sdk.schemas.transforms.CoGroup.Result
-  
- Result() - Constructor for class org.apache.beam.sdk.transforms.WithFailures.Result
-  
- resultBuilder() - Static method in class org.apache.beam.sdk.transforms.JsonToRow.ParseResult
-  
- resume() - Static method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
- 
Indicates that there is more work to be done for the current element. 
- resumeDelay() - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
- 
- resumeFromPreviousPipelineAction(ChangeStreamMetrics, MetadataTableDao, Instant, ProcessNewPartitionsAction) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.ActionFactory
-  
- ResumeFromPreviousPipelineAction - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.action
-  
- ResumeFromPreviousPipelineAction(ChangeStreamMetrics, MetadataTableDao, Instant, ProcessNewPartitionsAction) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.ResumeFromPreviousPipelineAction
-  
- retain(AwsOptions, ClientConfiguration) - Method in class org.apache.beam.sdk.io.aws2.common.ObjectPool.ClientPool
- 
Retain a reference to a shared client instance. 
- retain(KeyT) - Method in class org.apache.beam.sdk.io.aws2.common.ObjectPool
- 
Retain a reference to a shared client instance. 
- retrievalToken() - Method in class org.apache.beam.runners.fnexecution.provisioning.JobInfo
-  
- retrieveDicomStudyMetadata(String) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Retrieve DicomStudyMetadata. 
- retrieveDicomStudyMetadata(String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- retrieveFieldNames(List<ResolvedColumn>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
-  
- retrieveRexNode(ResolvedNodes.ResolvedProjectScan, List<RelDataTypeField>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
- 
Extract expressions from a project scan node. 
- retrieveRexNodeFromOrderByScan(RelOptCluster, ResolvedNodes.ResolvedOrderByScan, List<RelDataTypeField>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
- 
Extract expressions from order by scan node. 
- retry(RetryConfiguration) - Method in class org.apache.beam.sdk.io.aws2.common.ClientConfiguration.Builder
- 
- retry(Consumer<RetryConfiguration.Builder>) - Method in class org.apache.beam.sdk.io.aws2.common.ClientConfiguration.Builder
- 
- retry() - Method in class org.apache.beam.sdk.io.aws2.common.ClientConfiguration
- 
- retryCallable(Callable<V>, Set<Class<? extends Exception>>) - Method in class org.apache.beam.sdk.io.solace.RetryCallableManager
- 
Method that executes and repeats the execution of the callable argument, if it throws one of
 the exceptions from the exceptionsToIntercept Set. 
- RetryCallableManager - Class in org.apache.beam.sdk.io.solace
- 
A class that manages retrying of callables based on the exceptions they throw. 
- RetryCallableManager() - Constructor for class org.apache.beam.sdk.io.solace.RetryCallableManager
-  
- RetryCallableManager.Builder - Class in org.apache.beam.sdk.io.solace
-  
- RetryConfiguration - Class in org.apache.beam.sdk.io.aws2.common
- 
Configuration of the retry behavior for AWS SDK clients. 
- RetryConfiguration() - Constructor for class org.apache.beam.sdk.io.aws2.common.RetryConfiguration
-  
- RetryConfiguration() - Constructor for class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.RetryConfiguration
-  
- RetryConfiguration() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.RetryConfiguration
-  
- RetryConfiguration - Class in org.apache.beam.sdk.io.jms
-  
- RetryConfiguration() - Constructor for class org.apache.beam.sdk.io.jms.RetryConfiguration
-  
- RetryConfiguration() - Constructor for class org.apache.beam.sdk.io.solr.SolrIO.RetryConfiguration
-  
- RetryConfiguration.Builder - Class in org.apache.beam.sdk.io.aws2.common
-  
- RetryHttpRequestInitializer - Class in org.apache.beam.sdk.extensions.gcp.util
- 
Implements a request initializer that adds retry handlers to all HttpRequests. 
- RetryHttpRequestInitializer() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer
-  
- RetryHttpRequestInitializer(Collection<Integer>) - Constructor for class org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer
-  
- RetryHttpRequestInitializer(Collection<Integer>, HttpResponseInterceptor) - Constructor for class org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer
-  
- retryTransientErrors() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy
- 
Retry all failures except for known persistent errors. 
- reverse(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- REVERSE - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- REVERSE_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.BeamBuiltinMethods
-  
- reverseArtifactRetrievalService(StreamObserver<ArtifactApi.ArtifactRequestWrapper>) - Method in class org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService
-  
- reverseBytes(byte[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
-  
- Reversed() - Constructor for class org.apache.beam.sdk.transforms.Top.Reversed
-  
- reverseString(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
-  
- revision() - Method in interface org.apache.beam.sdk.options.PipelineOptions
- 
A monotonically increasing revision number of this  PipelineOptions object that can be
 used to detect changes. 
- RHS_TAG - Static variable in class org.apache.beam.sdk.schemas.transforms.Join
-  
- RIGHT - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamJoinPushThroughJoinRule
- 
Instance of the rule that works on logical joins only, and pushes to the right. 
- right(String...) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual.Impl
-  
- right(Integer...) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual.Impl
-  
- right(FieldAccessDescriptor) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual.Impl
-  
- right(String...) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual
-  
- right(Integer...) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual
-  
- right(FieldAccessDescriptor) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual
-  
- rightOuterJoin(String, PCollection<KV<K, V1>>, PCollection<KV<K, V2>>, V1) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
- 
Right Outer Join of two collections of KV elements. 
- rightOuterJoin(PCollection<KV<K, V1>>, PCollection<KV<K, V2>>, V1) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
- 
Right Outer Join of two collections of KV elements. 
- rightOuterJoin(PCollection<RhsT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Join
- 
Perform a right outer join. 
- RingRange - Class in org.apache.beam.sdk.io.cassandra
- 
Models a Cassandra token range. 
- rollback() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- rollback(Savepoint) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- root() - Static method in class org.apache.beam.sdk.transforms.display.DisplayData.Path
- 
Path for display data registered by a top-level component. 
- roundRobinSubList(List<T>, int, int) - Static method in class org.apache.beam.runners.jet.Utils
- 
Assigns the listtocountsublists in a round-robin fashion.
 
- route(DoFn.MultiOutputReceiver, RecordT, Coder<RecordT>, Exception, String) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecordRouter.RecordingBadRecordRouter
-  
- route(DoFn<?, ?>.FinishBundleContext, RecordT, Coder<RecordT>, Exception, String, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecordRouter.RecordingBadRecordRouter
-  
- route(DoFn.MultiOutputReceiver, RecordT, Coder<RecordT>, Exception, String) - Method in interface org.apache.beam.sdk.transforms.errorhandling.BadRecordRouter
-  
- route(DoFn<?, ?>.FinishBundleContext, RecordT, Coder<RecordT>, Exception, String, BoundedWindow) - Method in interface org.apache.beam.sdk.transforms.errorhandling.BadRecordRouter
-  
- route(DoFn.MultiOutputReceiver, RecordT, Coder<RecordT>, Exception, String) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecordRouter.ThrowingBadRecordRouter
-  
- route(DoFn<?, ?>.FinishBundleContext, RecordT, Coder<RecordT>, Exception, String, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecordRouter.ThrowingBadRecordRouter
-  
- ROW - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils
-  
- row(Schema) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
Create a map type for the given key and value types. 
- Row - Class in org.apache.beam.sdk.values
- 
Row is an immutable tuple-like schema to represent one element in a  PCollection. 
- Row.Builder - Class in org.apache.beam.sdk.values
- 
- Row.Equals - Class in org.apache.beam.sdk.values
-  
- Row.FieldValueBuilder - Class in org.apache.beam.sdk.values
- 
Builder for  Row that bases a row on another row. 
- ROW_PROPERTY_MUTATION_INFO - Static variable in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProvider
-  
- ROW_PROPERTY_MUTATION_SQN - Static variable in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProvider
-  
- ROW_PROPERTY_MUTATION_TYPE - Static variable in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProvider
-  
- ROW_SCHEMA - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProvider
-  
- ROW_SCHEMA_MUTATION_INFO - Static variable in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProvider
-  
- rowBag(Schema) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
Create a  StateSpec for a  BagState, optimized for adding values frequently and
 occasionally retrieving all the values that have been added. 
- RowBundle<T> - Class in org.apache.beam.sdk.jmh.schemas
- 
Bundle of rows according to the configured  Factory as input for benchmarks. 
- RowBundle() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundle
-  
- RowBundle(Class<T>) - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundle
-  
- RowBundle.Action - Enum in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles - Interface in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.ArrayOfNestedStringBundle - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.ArrayOfNestedStringBundle.Field - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.ArrayOfStringBundle - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.ArrayOfStringBundle.Field - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.ByteBufferBundle - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.ByteBufferBundle.Field - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.BytesBundle - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.BytesBundle.Field - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.DateTimeBundle - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.DateTimeBundle.Field - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.IntBundle - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.IntBundle.Field - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.MapOfIntBundle - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.MapOfIntBundle.Field - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.MapOfNestedIntBundle - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.MapOfNestedIntBundle.Field - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.NestedBytesBundle - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.NestedBytesBundle.Field - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.NestedIntBundle - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.NestedIntBundle.Field - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.StringBuilderBundle - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.StringBuilderBundle.Field - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.StringBundle - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowBundles.StringBundle.Field - Class in org.apache.beam.sdk.jmh.schemas
-  
- RowCoder - Class in org.apache.beam.sdk.coders
- 
A sub-class of SchemaCoder that can only encode  Row instances. 
- RowCoderCloudObjectTranslator - Class in org.apache.beam.runners.dataflow.util
- 
Translator for row coders. 
- RowCoderCloudObjectTranslator() - Constructor for class org.apache.beam.runners.dataflow.util.RowCoderCloudObjectTranslator
-  
- RowCoderGenerator - Class in org.apache.beam.sdk.coders
- 
A utility for automatically generating a  Coder for  Row objects corresponding to a
 specific schema. 
- RowCoderGenerator() - Constructor for class org.apache.beam.sdk.coders.RowCoderGenerator
-  
- rowFromProto(SchemaApi.Row, Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.SchemaTranslation
-  
- rowMap(Schema, Schema) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
- RowMessages - Class in org.apache.beam.sdk.schemas
-  
- rowMultimap(Schema, Schema) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
- RowMutation - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
- RowMutation() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.RowMutation
-  
- RowMutation.RowMutationCoder - Class in org.apache.beam.sdk.io.gcp.bigquery
-  
- RowMutationCoder() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.RowMutation.RowMutationCoder
-  
- RowMutationInformation - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
This class indicates how to apply a row update to BigQuery. 
- RowMutationInformation() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.RowMutationInformation
-  
- RowMutationInformation.MutationType - Enum in org.apache.beam.sdk.io.gcp.bigquery
-  
- rowOrderedList(Schema) - Static method in class org.apache.beam.sdk.state.StateSpecs
-  
- rowReceiver(DoFn<?, ?>.WindowedContext, TupleTag<T>, SchemaCoder<T>) - Static method in class org.apache.beam.sdk.transforms.DoFnOutputReceivers
- 
Returns a  DoFn.OutputReceiver that automatically converts a  Row to the user's output
 type and delegates to  WindowedContextOutputReceiver. 
- rows() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
- 
- RowSchemaInformationProvider - Class in org.apache.beam.sdk.schemas.utils
-  
- RowSchemaInformationProvider() - Constructor for class org.apache.beam.sdk.schemas.utils.RowSchemaInformationProvider
-  
- RowSelector - Interface in org.apache.beam.sdk.schemas.utils
- 
A selector interface for extracting fields from a row. 
- RowSelectorContainer(Schema, FieldAccessDescriptor, boolean) - Constructor for class org.apache.beam.sdk.schemas.utils.SelectHelpers.RowSelectorContainer
-  
- rowSet(Schema) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
- rowsFromRecordBatch(Schema, VectorSchemaRoot) - Static method in class org.apache.beam.sdk.extensions.arrow.ArrowConversion
- 
- rowsFromSerializedRecordBatch(Schema, InputStream, RootAllocator) - Static method in class org.apache.beam.sdk.extensions.arrow.ArrowConversion
-  
- rowToBytesFn(SchemaProvider, TypeDescriptor<T>, ProcessFunction<? super T, byte[]>) - Static method in class org.apache.beam.sdk.schemas.RowMessages
-  
- rowToBytesFn(SchemaProvider, TypeDescriptor<T>, Coder<? super T>) - Static method in class org.apache.beam.sdk.schemas.RowMessages
-  
- RowToEntity - Class in org.apache.beam.sdk.io.gcp.datastore
- 
A  PTransform to perform a conversion of  Row to  Entity. 
- rowToProto(Row) - Static method in class org.apache.beam.sdk.schemas.SchemaTranslation
-  
- RowUtils - Class in org.apache.beam.sdk.io.gcp.bigtable
-  
- RowUtils() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.RowUtils
-  
- rowValue(Schema) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
Create a  StateSpec for a row value with the specified schema. 
- RowWithGetters<T> - Class in org.apache.beam.sdk.values
- 
- RowWithStorage - Class in org.apache.beam.sdk.values
- 
Concrete subclass of  Row that explicitly stores all fields of the row. 
- rpad(String, Long) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
-  
- rpad(String, Long, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
-  
- rpad(byte[], Long) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
-  
- rpad(byte[], Long, byte[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
-  
- RpcQosOptions - Class in org.apache.beam.sdk.io.gcp.firestore
- 
Quality of Service manager options for Firestore RPCs. 
- RpcQosOptions.Builder - Class in org.apache.beam.sdk.io.gcp.firestore
- 
- rtrim(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- rtrim(String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- RTRIM - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- RTRIM_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.BeamBuiltinMethods
-  
- RULE_arrayQualifier - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- RULE_dotExpression - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- RULE_dotExpressionComponent - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- RULE_fieldSpecifier - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- RULE_mapQualifier - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- RULE_qualifiedComponent - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- RULE_qualifierList - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- ruleNames - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- ruleNames - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- run(Pipeline) - Method in class org.apache.beam.runners.dataflow.DataflowRunner
-  
- run(Pipeline) - Method in class org.apache.beam.runners.dataflow.TestDataflowRunner
-  
- run(Pipeline) - Method in class org.apache.beam.runners.direct.DirectRunner
-  
- run(RunnerApi.Pipeline, JobInfo) - Method in class org.apache.beam.runners.flink.FlinkPipelineRunner
-  
- run(Pipeline) - Method in class org.apache.beam.runners.flink.FlinkRunner
-  
- run(Pipeline) - Method in class org.apache.beam.runners.flink.TestFlinkRunner
-  
- run(Pipeline) - Method in class org.apache.beam.runners.jet.JetRunner
-  
- run(JobApi.RunJobRequest, StreamObserver<JobApi.RunJobResponse>) - Method in class org.apache.beam.runners.jobsubmission.InMemoryJobService
-  
- run() - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver
-  
- run(RunnerApi.Pipeline, JobInfo) - Method in class org.apache.beam.runners.jobsubmission.PortablePipelineJarCreator
- 
Does not actually run the pipeline. Instead bundles the input pipeline along with all
 dependencies, artifacts, etc. 
- run(RunnerApi.Pipeline, JobInfo) - Method in interface org.apache.beam.runners.jobsubmission.PortablePipelineRunner
-  
- run(Pipeline) - Method in class org.apache.beam.runners.portability.PortableRunner
-  
- run(JobApi.RunJobRequest, StreamObserver<JobApi.RunJobResponse>) - Method in class org.apache.beam.runners.portability.testing.TestJobService
-  
- run(Pipeline) - Method in class org.apache.beam.runners.portability.testing.TestPortableRunner
-  
- run(Pipeline) - Method in class org.apache.beam.runners.portability.testing.TestUniversalRunner
-  
- run(Pipeline) - Method in class org.apache.beam.runners.prism.PrismRunner
-  
- run(Pipeline) - Method in class org.apache.beam.runners.prism.TestPrismRunner
-  
- run(RunnerApi.Pipeline, JobInfo) - Method in class org.apache.beam.runners.spark.SparkPipelineRunner
-  
- run(Pipeline) - Method in class org.apache.beam.runners.spark.SparkRunner
-  
- run(Pipeline) - Method in class org.apache.beam.runners.spark.SparkRunnerDebugger
-  
- run(Pipeline) - Method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingRunner
-  
- run(Pipeline) - Method in class org.apache.beam.runners.spark.TestSparkRunner
-  
- run(Pipeline) - Method in class org.apache.beam.runners.twister2.Twister2Runner
-  
- run(Pipeline) - Method in class org.apache.beam.runners.twister2.Twister2TestRunner
-  
- run() - Method in interface org.apache.beam.sdk.function.ThrowingRunnable
-  
- run(PartitionRecord, ChangeStreamRecord, RestrictionTracker<StreamProgress, StreamProgress>, DoFn.OutputReceiver<KV<ByteString, ChangeStreamRecord>>, ManualWatermarkEstimator<Instant>, BytesThroughputEstimator<KV<ByteString, ChangeStreamRecord>>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.ChangeStreamAction
- 
This class processes ReadChangeStreamResponse from bigtable server. 
- run(RestrictionTracker<OffsetRange, Long>, DoFn.OutputReceiver<PartitionRecord>, ManualWatermarkEstimator<Instant>, InitialPipelineState) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.DetectNewPartitionsAction
- 
Perform the necessary steps to manage initial set of partitions and new partitions. 
- run(DoFn.OutputReceiver<PartitionRecord>, Instant) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.GenerateInitialPartitionsAction
- 
The very first step of the pipeline when there are no partitions being streamed yet. 
- run(PartitionRecord, RestrictionTracker<StreamProgress, StreamProgress>, DoFn.OutputReceiver<KV<ByteString, ChangeStreamRecord>>, ManualWatermarkEstimator<Instant>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.ReadChangeStreamPartitionAction
- 
Streams changes from a specific partition. 
- run(DoFn.OutputReceiver<PartitionRecord>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.action.ResumeFromPreviousPipelineAction
- 
Resume from previously drained pipeline. 
- run(PartitionMetadata, ChildPartitionsRecord, RestrictionTracker<TimestampRange, Timestamp>, RestrictionInterrupter<Timestamp>, ManualWatermarkEstimator<Instant>) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.action.ChildPartitionsRecordAction
- 
- run(PartitionMetadata, DataChangeRecord, RestrictionTracker<TimestampRange, Timestamp>, RestrictionInterrupter<Timestamp>, DoFn.OutputReceiver<DataChangeRecord>, ManualWatermarkEstimator<Instant>) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.action.DataChangeRecordAction
- 
- run(RestrictionTracker<TimestampRange, Timestamp>, DoFn.OutputReceiver<PartitionMetadata>, ManualWatermarkEstimator<Instant>) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.action.DetectNewPartitionsAction
- 
Executes the main logic to schedule new partitions. 
- run(PartitionMetadata, HeartbeatRecord, RestrictionTracker<TimestampRange, Timestamp>, RestrictionInterrupter<Timestamp>, ManualWatermarkEstimator<Instant>) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.action.HeartbeatRecordAction
- 
- run(PartitionMetadata, RestrictionTracker<TimestampRange, Timestamp>, DoFn.OutputReceiver<DataChangeRecord>, ManualWatermarkEstimator<Instant>, DoFn.BundleFinalizer) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.action.QueryChangeStreamAction
- 
This method will dispatch a change stream query for the given partition, it delegate the
 processing of the records to one of the corresponding action classes registered and it will
 keep the state of the partition up to date in the Connector's metadata table. 
- run() - Method in class org.apache.beam.sdk.Pipeline
- 
- run(PipelineOptions) - Method in class org.apache.beam.sdk.Pipeline
- 
- run(Pipeline) - Method in class org.apache.beam.sdk.PipelineRunner
- 
Processes the given  Pipeline, potentially asynchronously, returning a runner-specific
 type of result. 
- run(PTransform<PBegin, ?>, PipelineOptions) - Method in class org.apache.beam.sdk.PipelineRunner
- 
- run(PTransform<PBegin, ?>) - Method in class org.apache.beam.sdk.PipelineRunner
- 
- run(Pipeline) - Method in class org.apache.beam.sdk.testing.CrashingRunner
-  
- run() - Method in class org.apache.beam.sdk.testing.TestPipeline
- 
Runs this  TestPipeline, unwrapping any  AssertionError that is raised during
 testing. 
- run(PipelineOptions) - Method in class org.apache.beam.sdk.testing.TestPipeline
- 
- runBeforeProcessing(PipelineOptions) - Static method in class org.apache.beam.sdk.fn.JvmInitializers
- 
Finds all registered implementations of JvmInitializer and executes their beforeProcessingmethods.
 
- RunInference<OutputT> - Class in org.apache.beam.sdk.extensions.python.transforms
- 
Wrapper for invoking external Python RunInference.
 
- runInTransaction(Function<PartitionMetadataDao.InTransactionContext, T>) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao
- 
Runs a given function in a transaction context. 
- runInTransaction(Function<PartitionMetadataDao.InTransactionContext, T>, String) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao
-  
- Runner() - Constructor for class org.apache.beam.runners.dataflow.DataflowPipelineRegistrar.Runner
-  
- Runner() - Constructor for class org.apache.beam.runners.direct.DirectRegistrar.Runner
-  
- Runner() - Constructor for class org.apache.beam.runners.flink.FlinkRunnerRegistrar.Runner
-  
- Runner() - Constructor for class org.apache.beam.runners.jet.JetRunnerRegistrar.Runner
-  
- Runner() - Constructor for class org.apache.beam.runners.prism.PrismRegistrar.Runner
-  
- Runner() - Constructor for class org.apache.beam.runners.spark.SparkRunnerRegistrar.Runner
-  
- Runner() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingRunnerRegistrar.Runner
-  
- Runner() - Constructor for class org.apache.beam.runners.twister2.Twister2RunnerRegistrar.Runner
-  
- RunnerRegistrar() - Constructor for class org.apache.beam.runners.portability.testing.TestUniversalRunner.RunnerRegistrar
-  
- runOnStartup() - Static method in class org.apache.beam.sdk.fn.JvmInitializers
- 
Finds all registered implementations of JvmInitializer and executes their onStartupmethods.
 
- runQuery() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.Read
- 
Factory method to create a new type safe builder for RunQueryRequestoperations.
 
- runResourceIdBattery(ResourceId) - Static method in class org.apache.beam.sdk.io.fs.ResourceIdTester
- 
- runTest(Pipeline) - Method in class org.apache.beam.runners.twister2.Twister2Runner
-  
- runWindowFn(WindowFn<T, W>, List<Long>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
- 
Runs the  WindowFn over the provided input, returning a map of windows to the timestamps
 in those windows. 
- runWindowFnWithValue(WindowFn<T, W>, List<TimestampedValue<T>>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
- 
Runs the  WindowFn over the provided input, returning a map of windows to the timestamps
 in those windows. 
- runWithAdditionalOptionArgs(List<String>) - Method in class org.apache.beam.sdk.testing.TestPipeline
- 
Runs this  TestPipeline with additional cmd pipeline option args. 
- S3ClientBuilderFactory - Interface in org.apache.beam.sdk.io.aws2.options
- 
Construct S3ClientBuilder from S3 pipeline options. 
- S3FileSystemConfiguration - Class in org.apache.beam.sdk.io.aws2.s3
- 
Object used to configure S3FileSystem.
 
- S3FileSystemConfiguration() - Constructor for class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration
-  
- S3FileSystemConfiguration.Builder - Class in org.apache.beam.sdk.io.aws2.s3
-  
- S3FileSystemRegistrar - Class in org.apache.beam.sdk.io.aws2.s3
- 
AutoServiceregistrar for theS3FileSystem.
 
- S3FileSystemRegistrar() - Constructor for class org.apache.beam.sdk.io.aws2.s3.S3FileSystemRegistrar
-  
- S3FileSystemSchemeRegistrar - Interface in org.apache.beam.sdk.io.aws2.s3
- 
- S3Options - Interface in org.apache.beam.sdk.io.aws2.options
- 
Options used to configure Amazon Web Services S3. 
- S3Options.S3UploadBufferSizeBytesFactory - Class in org.apache.beam.sdk.io.aws2.options
- 
Provide the default s3 upload buffer size in bytes: 64MB if more than 512MB in RAM are
 available and 5MB otherwise. 
- S3Options.SSECustomerKeyFactory - Class in org.apache.beam.sdk.io.aws2.options
-  
- S3UploadBufferSizeBytesFactory() - Constructor for class org.apache.beam.sdk.io.aws2.options.S3Options.S3UploadBufferSizeBytesFactory
-  
- Sample - Class in org.apache.beam.sdk.transforms
- 
PTransforms for taking samples of the elements in aPCollection, or samples of
 the values associated with each key in aPCollectionofKVs.
 
- Sample() - Constructor for class org.apache.beam.sdk.transforms.Sample
-  
- Sample.FixedSizedSampleFn<T> - Class in org.apache.beam.sdk.transforms
- 
CombineFnthat computes a fixed-size sample of a collection of values.
 
- SAMPLE_PARTITION - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamsConstants
- 
We use a bogus partition here to estimate the average size of a partition metadata record. 
- SampleAllFiles() - Constructor for class org.apache.beam.sdk.io.TextRowCountEstimator.SampleAllFiles
-  
- satisfies(RelTrait) - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-  
- satisfies(SerializableFunction<Iterable<T>, Void>) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
- 
Applies the provided checking function (presumably containing assertions) to the iterable in
 question. 
- satisfies(SerializableFunction<Iterable<T>, Void>) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
-  
- satisfies(SerializableFunction<Iterable<T>, Void>) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionListContentsAssert
- 
- satisfies(List<SerializableFunction<Iterable<T>, Void>>) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionListContentsAssert
- 
- satisfies(SerializableFunction<T, Void>) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
- 
Applies the provided checking function (presumably containing assertions) to the value in
 question. 
- saveAsync(T) - Method in interface org.apache.beam.sdk.io.cassandra.Mapper
- 
This method is called for each save event. 
- SbeLogicalTypes - Class in org.apache.beam.sdk.extensions.sbe
- 
Classes that represent various SBE semantic types. 
- SbeLogicalTypes.LocalMktDate - Class in org.apache.beam.sdk.extensions.sbe
- 
Representation of SBE's LocalMktDate. 
- SbeLogicalTypes.TZTimeOnly - Class in org.apache.beam.sdk.extensions.sbe
- 
Represents SBE's TimeOnly composite type. 
- SbeLogicalTypes.TZTimestamp - Class in org.apache.beam.sdk.extensions.sbe
- 
Represents SBE's TZTimestamp composite type. 
- SbeLogicalTypes.Uint16 - Class in org.apache.beam.sdk.extensions.sbe
- 
Represents SBE's uint16 type. 
- SbeLogicalTypes.Uint32 - Class in org.apache.beam.sdk.extensions.sbe
- 
Represents SBE's uint32 type. 
- SbeLogicalTypes.Uint64 - Class in org.apache.beam.sdk.extensions.sbe
- 
Represents SBE's uint64 type. 
- SbeLogicalTypes.Uint8 - Class in org.apache.beam.sdk.extensions.sbe
- 
Represents SBE's uint8 type. 
- SbeLogicalTypes.UTCDateOnly - Class in org.apache.beam.sdk.extensions.sbe
- 
Representation of SBE's UTCDateOnly. 
- SbeLogicalTypes.UTCTimeOnly - Class in org.apache.beam.sdk.extensions.sbe
- 
Represents SBE's UTCTimeOnly composite type. 
- SbeLogicalTypes.UTCTimestamp - Class in org.apache.beam.sdk.extensions.sbe
- 
Represents SBE's UTCTimestamp composite type. 
- SbeSchema - Class in org.apache.beam.sdk.extensions.sbe
- 
Represents an SBE schema. 
- SbeSchema.IrOptions - Class in org.apache.beam.sdk.extensions.sbe
- 
Options for configuring schema generation from an  Ir. 
- SbeSchema.IrOptions.Builder - Class in org.apache.beam.sdk.extensions.sbe
- 
- ScalaInterop - Class in org.apache.beam.runners.spark.structuredstreaming.translation.utils
- 
Utilities for easier interoperability with the Spark Scala API. 
- ScalaInterop.Fun1<T,V> - Interface in org.apache.beam.runners.spark.structuredstreaming.translation.utils
-  
- ScalaInterop.Fun2<T1,T2,V> - Interface in org.apache.beam.runners.spark.structuredstreaming.translation.utils
-  
- scalaIterator(Iterable<T>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.utils.ScalaInterop
- 
Scala Iteratorof JavaIterable.
 
- scalaIterator(Iterator<T>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.utils.ScalaInterop
- 
Scala Iteratorof JavaIterator.
 
- SCALAR_FIELD_NAME - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- ScalarFn - Class in org.apache.beam.sdk.extensions.sql.udf
- 
A scalar function that can be executed as part of a SQL query. 
- ScalarFn() - Constructor for class org.apache.beam.sdk.extensions.sql.udf.ScalarFn
-  
- ScalarFn.ApplyMethod - Annotation Type in org.apache.beam.sdk.extensions.sql.udf
- 
Annotates the single method in a  ScalarFn implementation that is to be applied to SQL
 function arguments. 
- ScalarFnReflector - Class in org.apache.beam.sdk.extensions.sql.impl
- 
Reflection-based implementation logic for  ScalarFn. 
- ScalarFnReflector() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.ScalarFnReflector
-  
- ScalarFunctionImpl - Class in org.apache.beam.sdk.extensions.sql.impl
- 
Beam-customized version from ScalarFunctionImpl, to
 address BEAM-5921.
 
- ScalarFunctionImpl(Method, CallImplementor, String) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
-  
- ScalarFunctionImpl(Method, CallImplementor) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
-  
- ScheduledExecutorServiceFactory() - Constructor for class org.apache.beam.sdk.options.ExecutorOptions.ScheduledExecutorServiceFactory
-  
- schema() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
-  
- schema - Variable in class org.apache.beam.sdk.extensions.sql.meta.SchemaBaseBeamTable
-  
- schema(Schema) - Method in class org.apache.beam.sdk.extensions.sql.meta.Table.Builder
-  
- SCHEMA - Static variable in class org.apache.beam.sdk.io.ClassLoaderFileSystem
-  
- schema() - Method in class org.apache.beam.sdk.io.gcp.datastore.DataStoreV1SchemaIOProvider.DataStoreV1SchemaIO
-  
- SCHEMA - Static variable in class org.apache.beam.sdk.io.iceberg.SnapshotInfo
-  
- SCHEMA - Static variable in class org.apache.beam.sdk.managed.testing.TestSchemaTransformProvider
-  
- schema() - Method in interface org.apache.beam.sdk.schemas.io.SchemaIO
- 
Returns the schema of the data. 
- Schema - Class in org.apache.beam.sdk.schemas
- 
- Schema(List<Schema.Field>) - Constructor for class org.apache.beam.sdk.schemas.Schema
-  
- Schema(List<Schema.Field>, Schema.Options) - Constructor for class org.apache.beam.sdk.schemas.Schema
-  
- schema - Variable in class org.apache.beam.sdk.schemas.SchemaCoder
-  
- Schema.Builder - Class in org.apache.beam.sdk.schemas
- 
Builder class for building  Schema objects. 
- Schema.EquivalenceNullablePolicy - Enum in org.apache.beam.sdk.schemas
- 
Control whether nullable is included in equivalence check. 
- Schema.Field - Class in org.apache.beam.sdk.schemas
- 
Field of a row. 
- Schema.Field.Builder - Class in org.apache.beam.sdk.schemas
- 
- Schema.FieldType - Class in org.apache.beam.sdk.schemas
- 
A descriptor of a single field type. 
- Schema.LogicalType<InputT,BaseT> - Interface in org.apache.beam.sdk.schemas
- 
A LogicalType allows users to define a custom schema type. 
- Schema.Options - Class in org.apache.beam.sdk.schemas
-  
- Schema.Options.Builder - Class in org.apache.beam.sdk.schemas
-  
- Schema.TypeName - Enum in org.apache.beam.sdk.schemas
- 
An enumerated list of type constructors. 
- SchemaAndRecord - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
A wrapper for a  GenericRecord and the  TableSchema representing the schema of the
 table (or query) it was generated from. 
- SchemaAndRecord(GenericRecord, TableSchema) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.SchemaAndRecord
-  
- SchemaBaseBeamTable - Class in org.apache.beam.sdk.extensions.sql.meta
- 
- SchemaBaseBeamTable(Schema) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.SchemaBaseBeamTable
-  
- SchemaCaseFormat - Annotation Type in org.apache.beam.sdk.schemas.annotations
- 
When used on a  POJO,  Java Bean, or  AutoValue class the specified case format will be
 used for all the generated Schema fields. 
- schemaCoder(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Returns an SchemaCoderinstance for the provided element type.
 
- schemaCoder(Class<T>) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Returns an SchemaCoderinstance for the provided element class.
 
- schemaCoder(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Returns an SchemaCoderinstance for the Avro schema.
 
- schemaCoder(Class<T>, Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Returns an SchemaCoderinstance for the provided element type using the provided Avro
 schema.
 
- schemaCoder(AvroCoder<T>) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Returns an SchemaCoderinstance based on the provided AvroCoder for the element type.
 
- SchemaCoder<T> - Class in org.apache.beam.sdk.schemas
- 
SchemaCoder is used as the coder for types that have schemas registered. 
- SchemaCoder(Schema, TypeDescriptor<T>, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Constructor for class org.apache.beam.sdk.schemas.SchemaCoder
-  
- SchemaCoderCloudObjectTranslator - Class in org.apache.beam.runners.dataflow.util
- 
Translator for Schema coders. 
- SchemaCoderCloudObjectTranslator() - Constructor for class org.apache.beam.runners.dataflow.util.SchemaCoderCloudObjectTranslator
-  
- SchemaConversionOptions() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.SchemaConversionOptions
-  
- SchemaCreate - Annotation Type in org.apache.beam.sdk.schemas.annotations
- 
Can be put on a constructor or a static method, in which case that constructor or method will be
 used to created instance of the class by Beam's schema code. 
- SchemaFieldDescription - Annotation Type in org.apache.beam.sdk.schemas.annotations
- 
When used on a  POJO field, a  Java Bean getter, or an  AutoValue getter, the specified description is used
 for the generated schema field. 
- SchemaFieldName - Annotation Type in org.apache.beam.sdk.schemas.annotations
- 
When used on a  POJO field, a  Java Bean getter, or an  AutoValue getter, the specified name is used for the
 generated schema field. 
- SchemaFieldNumber - Annotation Type in org.apache.beam.sdk.schemas.annotations
- 
When used on a  POJO field, a  Java Bean getter, or an  AutoValue getter, the generated field will have the
 specified index. 
- schemaFor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.extensions.avro.schemas.AvroRecordSchema
-  
- schemaFor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoMessageSchema
-  
- schemaFor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.io.aws2.schemas.AwsSchemaProvider
-  
- schemaFor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.io.thrift.ThriftSchema
-  
- schemaFor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.annotations.DefaultSchema.DefaultSchemaProvider
-  
- schemaFor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.AutoValueSchema
-  
- schemaFor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema
-  
- schemaFor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.JavaFieldSchema
-  
- schemaFor(TypeDescriptor<T>) - Method in interface org.apache.beam.sdk.schemas.SchemaProvider
- 
Lookup a schema for the given type. 
- schemaFromClass(TypeDescriptor<?>, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.StaticSchemaInference
- 
Infer a schema from a Java class. 
- schemaFromJavaBeanClass(TypeDescriptor<?>, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
- 
Create a  Schema for a Java Bean class. 
- schemaFromPojoClass(TypeDescriptor<?>, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
-  
- schemaFromProto(SchemaApi.Schema) - Static method in class org.apache.beam.sdk.schemas.SchemaTranslation
-  
- SchemaIgnore - Annotation Type in org.apache.beam.sdk.schemas.annotations
- 
When used on a POJO field or a JavaBean getter, that field or getter is ignored from the inferred
 schema. 
- SchemaInformationProvider - Interface in org.apache.beam.sdk.schemas.utils
- 
- SchemaIO - Interface in org.apache.beam.sdk.schemas.io
- 
An abstraction to create schema capable and aware IOs. 
- SchemaIOProvider - Interface in org.apache.beam.sdk.schemas.io
- 
Provider to create  SchemaIO instances for use in Beam SQL and other SDKs. 
- SchemaIOTableProviderWrapper - Class in org.apache.beam.sdk.extensions.sql.meta.provider
- 
- SchemaIOTableProviderWrapper() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.SchemaIOTableProviderWrapper
-  
- SchemaLogicalType - Class in org.apache.beam.sdk.schemas.logicaltypes
- 
A schema represented as a serialized proto bytes. 
- SchemaLogicalType() - Constructor for class org.apache.beam.sdk.schemas.logicaltypes.SchemaLogicalType
-  
- schemaPathFromId(String, String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
-  
- schemaPathFromPath(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
-  
- SchemaProvider - Interface in org.apache.beam.sdk.schemas
- 
Concrete implementations of this class allow creation of schema service objects that vend a
  Schema for a specific type. 
- SchemaProviderRegistrar - Interface in org.apache.beam.sdk.schemas
- 
SchemaProvider creators have the ability to automatically have their  schemaProvider registered with this SDK by creating a  ServiceLoader entry
 and a concrete implementation of this interface. 
- SchemaRegistry - Class in org.apache.beam.sdk.schemas
- 
- schemaToProto(Schema, boolean) - Static method in class org.apache.beam.sdk.schemas.SchemaTranslation
-  
- schemaToProto(Schema, boolean, boolean) - Static method in class org.apache.beam.sdk.schemas.SchemaTranslation
-  
- schemaToProtoTableSchema(TableSchema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProto
-  
- SchemaTransform - Class in org.apache.beam.sdk.schemas.transforms
- 
An abstraction representing schema capable and aware transforms. 
- SchemaTransform() - Constructor for class org.apache.beam.sdk.schemas.transforms.SchemaTransform
-  
- SchemaTransformConfiguration - Class in org.apache.beam.sdk.io.iceberg
-  
- SchemaTransformConfiguration() - Constructor for class org.apache.beam.sdk.io.iceberg.SchemaTransformConfiguration
-  
- SchemaTransformConfiguration.Builder - Class in org.apache.beam.sdk.io.iceberg
-  
- SchemaTransformPayloadTranslator() - Constructor for class org.apache.beam.sdk.schemas.transforms.SchemaTransformTranslation.SchemaTransformPayloadTranslator
-  
- SchemaTransformProvider - Interface in org.apache.beam.sdk.schemas.transforms
- 
Provider to create  SchemaTransform instances for use in Beam SQL and other SDKs. 
- SchemaTransformTranslation - Class in org.apache.beam.sdk.schemas.transforms
- 
A  PTransformTranslation.TransformPayloadTranslator implementation that translates between a Java  SchemaTransform and a protobuf payload for that transform. 
- SchemaTransformTranslation() - Constructor for class org.apache.beam.sdk.schemas.transforms.SchemaTransformTranslation
-  
- SchemaTransformTranslation.SchemaTransformPayloadTranslator<T extends SchemaTransform> - Class in org.apache.beam.sdk.schemas.transforms
-  
- SchemaTranslation - Class in org.apache.beam.sdk.schemas
- 
Utility methods for translating schemas. 
- SchemaTranslation() - Constructor for class org.apache.beam.sdk.schemas.SchemaTranslation
-  
- schemaTypeCreator(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.extensions.avro.schemas.AvroRecordSchema
-  
- schemaTypeCreator(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoMessageSchema
-  
- schemaTypeCreator(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.io.aws2.schemas.AwsSchemaProvider
-  
- schemaTypeCreator(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.io.thrift.ThriftSchema
-  
- schemaTypeCreator(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.schemas.AutoValueSchema
-  
- schemaTypeCreator(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
- 
- schemaTypeCreator(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
- 
Deprecated. 
- schemaTypeCreator(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProviderV2
-  
- schemaTypeCreator(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProviderV2
-  
- schemaTypeCreator(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema
-  
- schemaTypeCreator(TypeDescriptor<?>, Schema) - Method in class org.apache.beam.sdk.schemas.JavaFieldSchema
-  
- SchemaUserTypeCreator - Interface in org.apache.beam.sdk.schemas
- 
A creator interface for user types that have schemas. 
- SchemaUtil - Class in org.apache.beam.sdk.io.jdbc
- 
Provides utility functions for working with Beam  Schema types. 
- SchemaUtil() - Constructor for class org.apache.beam.sdk.io.jdbc.SchemaUtil
-  
- SchemaUtil.BeamRowMapper - Class in org.apache.beam.sdk.io.jdbc
- 
- SchemaUtils - Class in org.apache.beam.sdk.schemas
- 
A set of utility functions for schemas. 
- SchemaUtils() - Constructor for class org.apache.beam.sdk.schemas.SchemaUtils
-  
- SchemaVerification - Class in org.apache.beam.sdk.values
-  
- SchemaVerification() - Constructor for class org.apache.beam.sdk.values.SchemaVerification
-  
- SchemaZipFold<T> - Class in org.apache.beam.sdk.schemas.utils
- 
Visitor that zips schemas, and accepts pairs of fields and their types. 
- SchemaZipFold() - Constructor for class org.apache.beam.sdk.schemas.utils.SchemaZipFold
-  
- SchemaZipFold.Context - Class in org.apache.beam.sdk.schemas.utils
- 
Context referring to a current position in a schema. 
- SCHEME - Static variable in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- scopedMetricsContainer(MetricsContainer) - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
- 
- SDF_PREFIX - Static variable in class org.apache.beam.runners.fnexecution.control.BundleCheckpointHandlers.StateAndTimerBundleCheckpointHandler
-  
- SdkCoreByteStringOutputStream() - Constructor for class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark.SdkCoreByteStringOutputStream
-  
- sdkFields() - Method in class org.apache.beam.sdk.io.aws2.schemas.AwsBuilderFactory
-  
- SdkHarnessClient - Class in org.apache.beam.runners.fnexecution.control
- 
A high-level client for an SDK harness. 
- SdkHarnessClient.BundleProcessor - Class in org.apache.beam.runners.fnexecution.control
- 
A processor capable of creating bundles for some registered BeamFnApi.ProcessBundleDescriptor.
 
- SdkHarnessClient.BundleProcessor.ActiveBundle - Class in org.apache.beam.runners.fnexecution.control
- 
An active bundle for a particular BeamFnApi.ProcessBundleDescriptor.
 
- SdkHarnessLogLevelOverrides() - Constructor for class org.apache.beam.sdk.options.SdkHarnessOptions.SdkHarnessLogLevelOverrides
-  
- SdkHarnessOptions - Interface in org.apache.beam.sdk.options
- 
Options that are used to control configuration of the SDK harness. 
- SdkHarnessOptions.BundleProcessorCacheTimeoutFactory - Class in org.apache.beam.sdk.options
-  
- SdkHarnessOptions.DefaultMaxCacheMemoryUsageMb - Class in org.apache.beam.sdk.options
- 
The default implementation which detects how much memory to use for a process wide cache. 
- SdkHarnessOptions.DefaultMaxCacheMemoryUsageMbFactory - Class in org.apache.beam.sdk.options
- 
- SdkHarnessOptions.LogLevel - Enum in org.apache.beam.sdk.options
- 
The set of log levels that can be used in the SDK harness. 
- SdkHarnessOptions.MaxCacheMemoryUsageMb - Interface in org.apache.beam.sdk.options
- 
Specifies the maximum amount of memory to use within the current SDK harness instance. 
- SdkHarnessOptions.SdkHarnessLogLevelOverrides - Class in org.apache.beam.sdk.options
- 
Defines a log level override for a specific class, package, or name. 
- searchFhirResource(String, String, Map<String, Object>, String) - Method in interface org.apache.beam.sdk.io.gcp.healthcare.HealthcareApiClient
- 
Search fhir resource http body. 
- searchFhirResource(String, String, Map<String, Object>, String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient
-  
- searchResources(String) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO
- 
Search resources from a Fhir store with String parameter values. 
- searchResourcesWithGenericParameters(String) - Static method in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO
- 
Search resources from a Fhir store with any type of parameter values. 
- secretManagerProjectId(String) - Method in class org.apache.beam.sdk.io.solace.broker.GCPSecretSessionServiceFactory.Builder
- 
Optional for Dataflow or VMs running on Google Cloud. 
- secretManagerProjectId() - Method in class org.apache.beam.sdk.io.solace.broker.GCPSecretSessionServiceFactory
-  
- seekable(BeamRelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
- 
check if BeamRelNodeimplementsBeamSeekableTable.
 
- seekableInputIndex() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
-  
- seekRow(Row) - Method in interface org.apache.beam.sdk.extensions.sql.BeamSqlSeekableTable
- 
return a list of Rowwith given key set.
 
- Select - Class in org.apache.beam.sdk.schemas.transforms
- 
A  PTransform for selecting a subset of fields from a schema type. 
- Select() - Constructor for class org.apache.beam.sdk.schemas.transforms.Select
-  
- select(Row) - Method in interface org.apache.beam.sdk.schemas.utils.RowSelector
-  
- select(Row) - Method in class org.apache.beam.sdk.schemas.utils.SelectHelpers.RowSelectorContainer
-  
- Select.Fields<T> - Class in org.apache.beam.sdk.schemas.transforms
-  
- Select.Flattened<T> - Class in org.apache.beam.sdk.schemas.transforms
- 
- SelectHelpers - Class in org.apache.beam.sdk.schemas.utils
- 
Helper methods to select subrows out of rows. 
- SelectHelpers() - Constructor for class org.apache.beam.sdk.schemas.utils.SelectHelpers
-  
- SelectHelpers.RowSelectorContainer - Class in org.apache.beam.sdk.schemas.utils
-  
- Semp - Class in org.apache.beam.sdk.io.solace.data
-  
- Semp() - Constructor for class org.apache.beam.sdk.io.solace.data.Semp
-  
- Semp.Queue - Class in org.apache.beam.sdk.io.solace.data
-  
- Semp.QueueData - Class in org.apache.beam.sdk.io.solace.data
-  
- SempBasicAuthClientExecutor - Class in org.apache.beam.sdk.io.solace.broker
- 
A class to execute requests to SEMP v2 with Basic Auth authentication. 
- SempBasicAuthClientExecutor(String, String, String, String, HttpRequestFactory) - Constructor for class org.apache.beam.sdk.io.solace.broker.SempBasicAuthClientExecutor
-  
- SempClient - Interface in org.apache.beam.sdk.io.solace.broker
- 
This interface defines methods for interacting with a Solace message broker using the Solace
 Element Management Protocol (SEMP). 
- SempClientFactory - Interface in org.apache.beam.sdk.io.solace.broker
- 
This interface serves as a blueprint for creating SempClient objects, which are used to interact
 with a Solace message broker using the Solace Element Management Protocol (SEMP). 
- sendElements(BeamFnApi.Elements) - Method in class org.apache.beam.sdk.fn.data.BeamFnDataOutboundAggregator
-  
- SENDER_TIMESTAMP_FUNCTION - Static variable in class org.apache.beam.sdk.io.solace.SolaceIO
-  
- sendOrCollectBufferedDataAndFinishOutboundStreams() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataOutboundAggregator
- 
Closes the streams for all registered outbound endpoints. 
- seqOf(T...) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.utils.ScalaInterop
-  
- SequenceDefinition() - Constructor for class org.apache.beam.sdk.transforms.PeriodicSequence.SequenceDefinition
-  
- SequenceDefinition(Instant, Instant, Duration) - Constructor for class org.apache.beam.sdk.transforms.PeriodicSequence.SequenceDefinition
-  
- SequenceDefinition(Instant, Instant, Duration, boolean) - Constructor for class org.apache.beam.sdk.transforms.PeriodicSequence.SequenceDefinition
- 
catchUpToNow is experimental; no backwards-compatibility guarantees. 
- SequenceRangeAccumulator - Class in org.apache.beam.sdk.extensions.ordered.combiner
- 
Default accumulator used to combine sequence ranges. 
- SequenceRangeAccumulator() - Constructor for class org.apache.beam.sdk.extensions.ordered.combiner.SequenceRangeAccumulator
-  
- SequenceRangeAccumulator.SequenceRangeAccumulatorCoder - Class in org.apache.beam.sdk.extensions.ordered.combiner
-  
- serde(T) - Static method in class org.apache.beam.runners.jet.Utils
- 
Returns a deep clone of an object by serializing and deserializing it (ser-de). 
- SerializableBiConsumer<FirstInputT,SecondInputT> - Interface in org.apache.beam.sdk.transforms
- 
A union of the BiConsumerandSerializableinterfaces.
 
- SerializableBiFunction<FirstInputT,SecondInputT,OutputT> - Interface in org.apache.beam.sdk.transforms
- 
A union of the BiFunctionandSerializableinterfaces.
 
- SerializableCoder<T extends java.io.Serializable> - Class in org.apache.beam.sdk.coders
- 
A  Coder for Java classes that implement  Serializable. 
- SerializableCoder(Class<T>, TypeDescriptor<T>) - Constructor for class org.apache.beam.sdk.coders.SerializableCoder
-  
- SerializableCoder.SerializableCoderProviderRegistrar - Class in org.apache.beam.sdk.coders
- 
- SerializableCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.coders.SerializableCoder.SerializableCoderProviderRegistrar
-  
- SerializableComparator<T> - Interface in org.apache.beam.sdk.transforms
- 
A Comparatorthat is alsoSerializable.
 
- SerializableConfiguration - Class in org.apache.beam.sdk.io.hadoop
- 
A wrapper to allow Hadoop Configurations to be serialized using Java's standard
 serialization mechanisms.
 
- SerializableConfiguration() - Constructor for class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
-  
- SerializableConfiguration(Configuration) - Constructor for class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
-  
- SerializableFunction<InputT,OutputT> - Interface in org.apache.beam.sdk.transforms
- 
A function that computes an output value of type OutputTfrom an input value of typeInputT, isSerializable, and does not allow checked exceptions to be declared.
 
- SerializableFunctions - Class in org.apache.beam.sdk.transforms
- 
- SerializableFunctions() - Constructor for class org.apache.beam.sdk.transforms.SerializableFunctions
-  
- SerializableIr - Class in org.apache.beam.sdk.extensions.sbe
- 
A wrapper around  Ir that fulfils Java's  Serializable contract. 
- SerializableMatcher<T> - Interface in org.apache.beam.sdk.testing
- 
A  Matcher that is also  Serializable. 
- SerializableMatchers - Class in org.apache.beam.sdk.testing
- 
- SerializableRexFieldAccess - Class in org.apache.beam.sdk.extensions.sql.impl.utils
- 
SerializableRexFieldAccess. 
- SerializableRexFieldAccess(RexFieldAccess) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.SerializableRexFieldAccess
-  
- SerializableRexInputRef - Class in org.apache.beam.sdk.extensions.sql.impl.utils
- 
SerializableRexInputRef. 
- SerializableRexInputRef(RexInputRef) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.SerializableRexInputRef
-  
- SerializableRexNode - Class in org.apache.beam.sdk.extensions.sql.impl.utils
- 
SerializableRexNode. 
- SerializableRexNode() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.SerializableRexNode
-  
- SerializableRexNode.Builder - Class in org.apache.beam.sdk.extensions.sql.impl.utils
- 
SerializableRexNode.Builder. 
- SerializableSplit() - Constructor for class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.SerializableSplit
-  
- SerializableSplit(InputSplit) - Constructor for class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.SerializableSplit
-  
- serialize(String, Instant) - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantSerializer
-  
- serialize(ValueProvider<?>, JsonGenerator, SerializerProvider) - Method in class org.apache.beam.sdk.options.ValueProvider.Serializer
-  
- serialize(Row) - Method in interface org.apache.beam.sdk.schemas.io.payloads.PayloadSerializer
-  
- serializeAwsCredentialsProvider(AwsCredentialsProvider) - Static method in class org.apache.beam.sdk.io.aws2.options.AwsSerializableUtils
-  
- SERIALIZED_FN - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- SERIALIZED_TEST_STREAM - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- serializeOneOf(Expression, List<Encoder<T>>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers
-  
- Serializer() - Constructor for class org.apache.beam.sdk.options.ValueProvider.Serializer
-  
- serializeTimers(Collection<TimerInternals.TimerData>, TimerInternals.TimerDataCoderV2) - Static method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-  
- serialVersionUID - Static variable in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
-  
- serialVersionUID - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
-  
- serialVersionUID - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
-  
- serialVersionUID - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoDynamicMessageSchema
-  
- serialVersionUID - Static variable in interface org.apache.beam.sdk.schemas.io.payloads.PayloadSerializer
-  
- seriesId - Variable in class org.apache.beam.sdk.io.gcp.healthcare.WebPathParser.DicomWebPath
-  
- ServerConfiguration() - Constructor for class org.apache.beam.runners.jobsubmission.JobServerDriver.ServerConfiguration
-  
- serverDirect() - Static method in class org.apache.beam.sdk.fn.stream.OutboundObserverFactory
- 
- ServerFactory - Class in org.apache.beam.sdk.fn.server
- 
A gRPC serverfactory.
 
- ServerFactory() - Constructor for class org.apache.beam.sdk.fn.server.ServerFactory
-  
- ServerFactory.InetSocketAddressServerFactory - Class in org.apache.beam.sdk.fn.server
- 
Creates a gRPC Serverusing the default server factory.
 
- ServerFactory.UrlFactory - Interface in org.apache.beam.sdk.fn.server
- 
Factory that constructs client-accessible URLs from a local server address and port. 
- ServerInfo() - Constructor for class org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory.ServerInfo
-  
- SESSION_WINDOW_TVF - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.TVFStreamingUtils
-  
- sessionBuilder(String) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.SparkSessionFactory
- 
Creates Spark session builder with some optimizations for local mode, e.g. 
- Sessions - Class in org.apache.beam.sdk.transforms.windowing
- 
- SessionService - Class in org.apache.beam.sdk.io.solace.broker
- 
The SessionService interface provides a set of methods for managing a session with the Solace
 messaging system. 
- SessionService() - Constructor for class org.apache.beam.sdk.io.solace.broker.SessionService
-  
- SessionServiceFactory - Class in org.apache.beam.sdk.io.solace.broker
- 
This abstract class serves as a blueprint for creating `SessionServiceFactory` objects. 
- SessionServiceFactory() - Constructor for class org.apache.beam.sdk.io.solace.broker.SessionServiceFactory
-  
- set(long) - Method in class org.apache.beam.runners.jet.metrics.GaugeImpl
-  
- set(long) - Method in class org.apache.beam.sdk.metrics.DelegatingGauge
- 
Set the gauge. 
- set(long) - Method in interface org.apache.beam.sdk.metrics.Gauge
- 
Set current value for this gauge. 
- set(long) - Method in class org.apache.beam.sdk.metrics.NoOpGauge
-  
- set(ObjectT, ValueT) - Method in interface org.apache.beam.sdk.schemas.FieldValueSetter
- 
Sets the specified field on object to value. 
- set() - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
- set(Coder<T>) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
- set(Instant) - Method in interface org.apache.beam.sdk.state.Timer
- 
Sets or resets the time in the timer's  TimeDomain at which it should fire. 
- set(String, Instant) - Method in interface org.apache.beam.sdk.state.TimerMap
-  
- set(long...) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
- 
- setAccessKey(String) - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- setAccountName(String) - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- setActiveWorkRefreshPeriodMillis(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setAllowDuplicates(Boolean) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setAllowNonRestoredState(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setAlsoStartLoopbackWorker(boolean) - Method in interface org.apache.beam.sdk.expansion.service.ExpansionServiceOptions
-  
- setApiRootUrl(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setAppend(Boolean) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider.Configuration.Builder
-  
- setAppName(String) - Method in interface org.apache.beam.sdk.options.ApplicationNameOptions
-  
- setArtifactPort(int) - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver.ServerConfiguration
-  
- setAttachedMode(boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setAttachmentBytes(byte[]) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record.Builder
-  
- setAttributeId(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration.Builder
-  
- setAttributeId(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration.Builder
-  
- setAttributeMap(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration.Builder
-  
- setAttributes(List<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration.Builder
-  
- setAttributes(List<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration.Builder
-  
- setAttributes(List<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration.Builder
-  
- setAttributes(List<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration.Builder
-  
- setAttributesMap(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration.Builder
-  
- setAttributesMap(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration.Builder
-  
- setAuthenticator(String) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setAutoBalanceWriteFilesShardingEnabled(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setAutoCommit(boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- setAutoOffsetResetConfig(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration.Builder
-  
- setAutoscalingAlgorithm(DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-  
- setAutoSharding(Boolean) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.Builder
-  
- setAutosharding(Boolean) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration.Builder
-  
- setAutoWatermarkInterval(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setAveragePartitionBytesSize(long) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.DetectNewPartitionsDoFn
- 
Sets the average partition bytes size to estimate the backlog of this DoFn. 
- setAwsCredentialsProvider(AwsCredentialsProvider) - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
-  
- setAwsRegion(Region) - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
-  
- setAzureConnectionString(String) - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- setAzureCredentialsProvider(TokenCredential) - Method in interface org.apache.beam.sdk.io.azure.options.AzureOptions
-  
- setBatching(Boolean) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration.Builder
-  
- setBatching(Boolean) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.ReadBuilder.Configuration
-  
- setBatchIntervalMillis(Long) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- setBatchRequestSupplier(Supplier<GcsUtil.BatchInterface>) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
-  
- setBatchSize(Long) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration.Builder
-  
- setBatchSize(Integer) - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteConfiguration.Builder
-  
- setBatchSizeBytes(Integer) - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText.Builder
-  
- setBatchSizeBytes(Integer) - Method in class org.apache.beam.sdk.extensions.ml.DLPInspectText.Builder
-  
- setBatchSizeBytes(Integer) - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText.Builder
-  
- setBeamVersion(String) - Method in class org.apache.beam.sdk.transformservice.launcher.TransformServiceLauncher
- 
Specifies the Beam version to get containers for the transform service. 
- setBigQueryEndpoint(String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setBigQueryLocation(String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.TestBigQueryOptions
-  
- setBigQueryProject(String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setBigQueryServices(BigQueryServices) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransform
-  
- setBigQueryServices(BigQueryServices) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform
-  
- setBigtableChangeStreamInstanceId(String) - Method in interface org.apache.beam.sdk.io.gcp.bigtable.changestreams.BigtableChangeStreamTestOptions
-  
- setBlobServiceEndpoint(String) - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- setBlobstoreClientFactoryClass(Class<? extends BlobstoreClientBuilderFactory>) - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- setBlockOnRun(boolean) - Method in interface org.apache.beam.runners.direct.DirectOptions
-  
- setBlockOnRun(boolean) - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-  
- setBootstrapServers(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration.Builder
- 
Sets the bootstrap servers for the Kafka consumer. 
- setBootstrapServers(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration.Builder
-  
- setBqStreamingApiLoggingFrequencySec(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setBranch(String) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setBucketKeyEnabled(boolean) - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- setBucketKeyEnabled(boolean) - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration.Builder
-  
- setBundleProcessorCacheTimeout(Duration) - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
-  
- setBundleSize(Long) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- setCacheDisabled(boolean) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- setCallable(String) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaRowUdf.Configuration.Builder
-  
- setCaseSensitive(Boolean) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setCatalog(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- setCatalogConfig(IcebergCatalogConfig) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setCatalogName(String) - Method in class org.apache.beam.sdk.io.iceberg.IcebergCatalogConfig.Builder
-  
- setCatalogName(String) - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration.Builder
-  
- setCatalogName(String) - Method in class org.apache.beam.sdk.io.iceberg.SchemaTransformConfiguration.Builder
-  
- setCatalogProperties(Map<String, String>) - Method in class org.apache.beam.sdk.io.iceberg.IcebergCatalogConfig.Builder
-  
- setCatalogProperties(Map<String, String>) - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration.Builder
-  
- setCatalogProperties(Map<String, String>) - Method in class org.apache.beam.sdk.io.iceberg.SchemaTransformConfiguration.Builder
-  
- setChangeStreamName(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration.Builder
-  
- setChannelzShowOnlyWindmillServiceChannels(boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setCharset(String) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.XmlConfiguration.Builder
- 
The charset used to write the file. 
- setCheckpoint(Long) - Method in interface org.apache.beam.sdk.io.sparkreceiver.HasOffset
- 
Some Receiversupport mechanism of checkpoint (e.g.
 
- setCheckpointDir(String) - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
-  
- setCheckpointDurationMillis(Long) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- setCheckpointingInterval(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setCheckpointingInterval(Long) - Method in interface org.apache.beam.sdk.io.kafka.KafkaIO.Read.FakeFlinkPipelineOptions
-  
- setCheckpointingMode(String) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setCheckpointTimeoutMillis(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setChecksum(String) - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
-  
- setClientBuilderFactory(Class<? extends ClientBuilderFactory>) - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
-  
- setClientFactory(PubsubTestClient.PubsubTestClientFactory) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration.Builder
-  
- setClientInfo(String, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- setClientInfo(Properties) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- setClock(Clock) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration.Builder
-  
- setClock(PubsubIO.Read<T>, Clock) - Method in interface org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient.PubsubTestClientFactory
-  
- setCloningBehavior(DoFnTester.CloningBehavior) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- setClusterName(String) - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
-  
- setClusterType(String) - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- setCodeJarPathname(String) - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
-  
- SetCoder<T> - Class in org.apache.beam.sdk.coders
- 
- SetCoder(Coder<T>) - Constructor for class org.apache.beam.sdk.coders.SetCoder
-  
- setCoder(String) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Record.Builder
-  
- setCoder(Coder<T>) - Method in class org.apache.beam.sdk.values.PCollection
- 
Sets the  Coder used by this  PCollection to encode and decode the values stored
 in it. 
- setCoderRegistry(CoderRegistry) - Method in class org.apache.beam.sdk.Pipeline
- 
- setColumnDelimiter(String) - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText.Builder
-  
- setColumnDelimiter(String) - Method in class org.apache.beam.sdk.extensions.ml.DLPInspectText.Builder
-  
- setColumnDelimiter(String) - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText.Builder
-  
- setColumns(List<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration.Builder
-  
- setColumns(SnowflakeColumn[]) - Method in class org.apache.beam.sdk.io.snowflake.data.SnowflakeTableSchema
-  
- setCommitOffsetInFinalize(Boolean) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setCompression(String) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.Builder
- 
- setCompression(Compression) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
-  
- setCompressionCodecName(String) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.ParquetConfiguration.Builder
- 
Specifies compression codec. 
- setConfig(byte[]) - Method in class org.apache.beam.sdk.extensions.schemaio.expansion.ExternalSchemaIOTransformRegistrar.Configuration
-  
- setConfigProperties(Map<String, String>) - Method in class org.apache.beam.sdk.io.iceberg.IcebergCatalogConfig.Builder
-  
- setConfigProperties(Map<String, String>) - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration.Builder
-  
- setConfigProperties(Map<String, String>) - Method in class org.apache.beam.sdk.io.iceberg.SchemaTransformConfiguration.Builder
-  
- setConfluentSchemaRegistrySubject(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration.Builder
-  
- setConfluentSchemaRegistryUrl(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration.Builder
-  
- setConnectionInitSql(List<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration.Builder
-  
- setConnectionInitSql(List<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration.Builder
-  
- setConnectionProperties(List<String>) - Method in class org.apache.beam.io.debezium.DebeziumTransformRegistrar.ReadBuilder.Configuration
-  
- setConnectionProperties(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration.Builder
-  
- setConnectionProperties(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration.Builder
-  
- setConsumerConfig(Map<String, String>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setConsumerConfigUpdates(Map<String, String>) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration.Builder
-  
- setConsumerPollingTimeout(Long) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setContentType(String) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.CreateOptions.Builder
-  
- setContext(BatchContextImpl) - Method in class org.apache.beam.sdk.io.cdap.Plugin.Builder
-  
- setCosmosClientBuilder(CosmosClientBuilder) - Method in interface org.apache.beam.sdk.io.azure.cosmos.CosmosOptions
-  
- setCosmosKey(String) - Method in interface org.apache.beam.sdk.io.azure.cosmos.CosmosOptions
-  
- setCosmosServiceEndpoint(String) - Method in interface org.apache.beam.sdk.io.azure.cosmos.CosmosOptions
-  
- setCountBackoffs(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountCacheReadFailures(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountCacheReadNonNulls(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountCacheReadNulls(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountCacheReadRequests(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountCacheWriteFailures(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountCacheWriteRequests(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountCacheWriteSuccesses(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountCalls(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountFailures(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountRequests(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountResponses(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountryOfResidence(String) - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
-  
- setCountSetup(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountShouldBackoff(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountSleeps(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCountTeardown(Boolean) - Method in class org.apache.beam.io.requestresponse.Monitoring.Builder
-  
- setCreatedAt(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata.Builder
- 
Sets the time at which the partition was created. 
- setCreateDisposition(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.Builder
-  
- setCreateDisposition(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.WriteBuilder.Configuration
-  
- setCreateFromSnapshot(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setCredentialFactoryClass(Class<? extends CredentialFactory>) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-  
- setCrossProduct(Boolean) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaExplodeTransformProvider.Configuration.Builder
-  
- setCsvConfiguration(FileWriteSchemaTransformConfiguration.CsvConfiguration) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.Builder
- 
Configures extra details related to writing CSV formatted files. 
- setCurrentBundleTimestamp(Instant) - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- setCurrentContainer(MetricsContainer) - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
- 
- setCurrentSchema(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
- 
Set the current (default) schema. 
- setCurrentTransform(AppliedPTransform<?, ?, ?>) - Method in class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- setCustomBeamRequirement(String) - Method in interface org.apache.beam.sdk.extensions.python.PythonExternalTransformOptions
- 
Set custom Beam version for bootstrap Beam venv. 
- setCustomerId(int) - Method in class org.apache.beam.sdk.extensions.sql.example.model.Order
-  
- setCustomerProvidedKey(CustomerProvidedKey) - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- setCustomErrors(CustomHttpErrors) - Method in class org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer
-  
- setDatabase(String) - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration.Builder
-  
- setDatabase(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setDatabase(ValueProvider<String>) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration.Builder
-  
- setDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration.Builder
-  
- setDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.CrossLanguageConfiguration
-  
- setDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration.Builder
-  
- setDataCatalogEndpoint(String) - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogPipelineOptions
-  
- setDataflowClient(Dataflow) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setDataflowEndpoint(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setDataflowEndpoint(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setDataflowJobFile(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setDataflowKmsKey(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-  
- setDataflowServiceOptions(List<String>) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setDataflowWorkerJar(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setDataSchema(byte[]) - Method in class org.apache.beam.sdk.extensions.schemaio.expansion.ExternalSchemaIOTransformRegistrar.Configuration
-  
- setDataSourceConfiguration(SingleStoreIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadConfiguration.Builder
-  
- setDataSourceConfiguration(SingleStoreIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteConfiguration.Builder
-  
- setDataType(SnowflakeDataType) - Method in class org.apache.beam.sdk.io.snowflake.data.SnowflakeColumn
-  
- setDebeziumConnectionProperties(List<String>) - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration.Builder
-  
- setDeduplicate(Deduplicate.KeyedValues<Uuid, SequencedMessage>) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.UuidDeduplicationOptions.Builder
- 
Set the deduplication transform. 
- setDefaultEnvironmentConfig(String) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- setDefaultEnvironmentType(String) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- setDefaultPipelineOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.io.FileSystems
- 
Sets the default configuration in workers. 
- setDefaultPipelineOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.metrics.Metrics
- 
Initialize metrics flags if not already done so. 
- setDefaultSdkHarnessLogLevel(SdkHarnessOptions.LogLevel) - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
-  
- setDefaultTimezone(String) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
-  
- setDefaultWorkerLogLevel(DataflowWorkerLoggingOptions.Level) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
- 
Deprecated. 
- setDeidentifyConfig(DeidentifyConfig) - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText.Builder
-  
- setDeidentifyTemplateName(String) - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText.Builder
-  
- setDelimiter(String) - Method in class org.apache.beam.sdk.io.csv.providers.CsvWriteTransformProvider.CsvWriteConfiguration.Builder
-  
- setDelimiters(byte[]) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
-  
- setDescription(String) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
-  
- setDescription(String) - Method in class org.apache.beam.sdk.schemas.Schema.Field.Builder
-  
- setDescription(String) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Failure.Builder
-  
- setDesiredNumUnboundedSourceSplits(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setDestination(Solace.Destination) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record.Builder
-  
- setDialect(String) - Method in class org.apache.beam.sdk.extensions.sql.expansion.ExternalSqlTransformRegistrar.Configuration
-  
- setDirectoryTreatment(FileIO.ReadMatches.DirectoryTreatment) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
-  
- setDisableAutoCommit(Boolean) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration.Builder
-  
- setDisableMetrics(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setDiskSizeGb(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-  
- setDisplayData(List<DisplayData.ItemSpec<?>>) - Method in class org.apache.beam.sdk.transforms.PTransform
- 
Set display data for your PTransform. 
- setDriverClassName(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration.Builder
-  
- setDriverClassName(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration.Builder
-  
- setDriverJars(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration.Builder
-  
- setDriverJars(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration.Builder
-  
- setDrop(List<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.Builder
-  
- setDrop(List<String>) - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration.Builder
-  
- setDrop(List<String>) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider.Configuration.Builder
-  
- setDumpHeapOnOOM(boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setDuplicateCount(long) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus.Builder
-  
- setEarliestBufferedSequence(Long) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus.Builder
-  
- setElasticsearchHttpPort(Integer) - Method in interface org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOITCommon.ElasticsearchPipelineOptions
-  
- setElasticsearchServer(String) - Method in interface org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOITCommon.ElasticsearchPipelineOptions
-  
- setElements(Long) - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration.Rate.Builder
-  
- setElementsPerPeriod(Long) - Method in class org.apache.beam.sdk.io.GenerateSequence.External.ExternalConfiguration
-  
- setElementType(FieldValueTypeInformation) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
-  
- setEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
-  
- setEmulatorHost(String) - Method in interface org.apache.beam.sdk.io.gcp.firestore.FirestoreOptions
- 
Define a host port pair to allow connecting to a Cloud Firestore emulator instead of the live
 service. 
- setEmulatorHost(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.CrossLanguageConfiguration
-  
- setEnableBucketReadMetricCounter(Boolean) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- setEnableBucketWriteMetricCounter(Boolean) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- setEnableHeapDumps(boolean) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- setEnableLogViaFnApi(boolean) - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
-  
- setEnableSparkMetricSinks(Boolean) - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
-  
- setEnableStableInputDrain(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setEnableStorageReadApiV2(Boolean) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setEnableStreamingEngine(boolean) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-  
- setEnableWebUI(Boolean) - Method in interface org.apache.beam.runners.prism.PrismPipelineOptions
-  
- setEncodedRecord(byte[]) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Record.Builder
-  
- setEncodingPositions(Map<String, Integer>) - Method in class org.apache.beam.sdk.schemas.Schema
- 
Sets the encoding positions for this schema. 
- setEnd(Long) - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration.Builder
-  
- setEndAtTimestamp(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration.Builder
-  
- setEndpoint(URI) - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
-  
- setEndTime(Instant) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.PartitionRecord
-  
- setEndTimestamp(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata.Builder
- 
Sets the end time of the partition. 
- setEnforceEncodability(boolean) - Method in interface org.apache.beam.runners.direct.DirectOptions
-  
- setEnforceImmutability(boolean) - Method in interface org.apache.beam.runners.direct.DirectOptions
-  
- setEnvironmentCacheMillis(int) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- setEnvironmentExpirationMillis(int) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- setEnvironmentOptions(List<String>) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- setError(String) - Method in class org.apache.beam.sdk.io.solace.data.Solace.PublishResult.Builder
-  
- setError(String) - Method in class org.apache.beam.sdk.schemas.io.Failure.Builder
-  
- setErrorField(String) - Method in class org.apache.beam.sdk.transforms.JsonToRow.JsonToRowWithErrFn
- 
Adds the error message to the returned error Row. 
- setErrorHandling(BigQueryWriteConfiguration.ErrorHandling) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.Builder
-  
- setErrorHandling(PubsubReadSchemaTransformConfiguration.ErrorHandling) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration.Builder
-  
- setErrorHandling(PubsubWriteSchemaTransformConfiguration.ErrorHandling) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration.Builder
-  
- setErrorHandling(ErrorHandling) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration.Builder
-  
- setErrorHandling(ErrorHandling) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration.Builder
-  
- setErrorHandling(ErrorHandling) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration.Builder
-  
- setErrorHandling(ErrorHandling) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration.Builder
-  
- setErrorHandling(ErrorHandling) - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration.Builder
-  
- setErrorHandling(ErrorHandling) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaFilterTransformProvider.Configuration.Builder
-  
- setErrorHandling(ErrorHandling) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider.Configuration.Builder
-  
- setException(String) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Failure.Builder
-  
- setExceptionStacktrace(String) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Failure.Builder
-  
- setExecutionModeForBatch(String) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setExecutionRetryDelay(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setExecutorService(ExecutorService) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
- 
- setExpansionPort(int) - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver.ServerConfiguration
-  
- setExpansionServiceConfig(ExpansionServiceConfig) - Method in interface org.apache.beam.sdk.expansion.service.ExpansionServiceOptions
-  
- setExpansionServiceConfigFile(String) - Method in interface org.apache.beam.sdk.expansion.service.ExpansionServiceOptions
-  
- setExpectedAssertions(Integer) - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
-  
- setExpectFileToNotExist(boolean) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.CreateOptions.Builder
-  
- setExpectFileToNotExist(Boolean) - Method in class org.apache.beam.sdk.io.fs.CreateOptions.Builder
-  
- setExperiments(List<String>) - Method in interface org.apache.beam.sdk.options.ExperimentalOptions
-  
- setExpiration(long) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record.Builder
-  
- setExpression(String) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaRowUdf.Configuration.Builder
-  
- setExternalizedCheckpointsEnabled(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setExtraInteger(Integer) - Method in class org.apache.beam.sdk.managed.testing.TestSchemaTransformProvider.Config.Builder
-  
- setExtraString(String) - Method in class org.apache.beam.sdk.managed.testing.TestSchemaTransformProvider.Config.Builder
-  
- setFailOnCheckpointingErrors(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setFailsafeTableRowPayload(byte[]) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWritePayload.Builder
-  
- setFailToLock(boolean) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.StreamProgress
-  
- setFailure(BadRecord.Failure) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Builder
-  
- setFasterCopy(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setFeatures(AnnotateTextRequest.Features) - Method in class org.apache.beam.sdk.extensions.ml.AnnotateText.Builder
-  
- setFetchSize(Integer) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration.Builder
-  
- setField(Field) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
-  
- setFieldId(Integer) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Builder
-  
- setFieldName(String) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Builder
-  
- setFieldRename(String) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Builder
-  
- setFields(List<String>) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaExplodeTransformProvider.Configuration.Builder
-  
- setFields(Map<String, JavaRowUdf.Configuration>) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider.Configuration.Builder
-  
- setFileDescriptorPath(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration.Builder
-  
- setFileDescriptorPath(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration.Builder
-  
- setFileDescriptorPath(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration.Builder
-  
- setFileDescriptorPath(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration.Builder
-  
- setFileFormat(FileFormat) - Method in class org.apache.beam.sdk.io.iceberg.IcebergDestination.Builder
-  
- setFileInputSplitMaxSizeMB(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setFilenamePrefix(String) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.Builder
- 
A common prefix to use for all generated filenames. 
- setFilenameSuffix(String) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.Builder
- 
Configures the filename suffix for written files. 
- setFilepattern(String) - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformConfiguration.Builder
-  
- setFilePattern(String) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
-  
- setFilesToStage(List<String>) - Method in interface org.apache.beam.sdk.options.FileStagingOptions
-  
- setFileSystem(FileSystem) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- setFilter(Expression) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setFinishBundleBeforeCheckpointing(boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setFinishedAt(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata.Builder
- 
Sets the time at which the partition finished running. 
- setFirestoreDb(String) - Method in interface org.apache.beam.sdk.io.gcp.firestore.FirestoreOptions
- 
Set the Firestore database ID to connect to. 
- setFirestoreHost(String) - Method in interface org.apache.beam.sdk.io.gcp.firestore.FirestoreOptions
- 
Define a host port pair to allow connecting to a Cloud Firestore instead of the default live
 service. 
- setFirestoreProject(String) - Method in interface org.apache.beam.sdk.io.gcp.firestore.FirestoreOptions
- 
- setFlexRSGoal(DataflowPipelineOptions.FlexResourceSchedulingGoal) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setFlinkConfDir(String) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setFlinkMaster(String) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setForceSlotSharingGroup(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setForceStreaming(boolean) - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
-  
- setForceUnalignedCheckpointEnabled(boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setFormat(String) - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformConfiguration.Builder
-  
- setFormat(String) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.Builder
- 
The format of the file content. 
- setFormat(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration.Builder
-  
- setFormat(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration.Builder
-  
- setFormat(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration.Builder
-  
- setFormat(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration.Builder
-  
- setFormat(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration.Builder
-  
- setFormat(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration.Builder
-  
- setFormatClass(Class<?>) - Method in class org.apache.beam.sdk.io.cdap.Plugin.Builder
-  
- setFormatProviderClass(Class<?>) - Method in class org.apache.beam.sdk.io.cdap.Plugin.Builder
-  
- setFromSnapshotExclusive(Long) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setFromSnapshotInclusive(Long) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setFromSnapshotRefExclusive(String) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setFromSnapshotRefInclusive(String) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setGcpCredential(Credentials) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-  
- setGcpOauthScopes(List<String>) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-  
- setGcpTempLocation(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-  
- setGcsEndpoint(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- setGcsHttpRequestReadTimeout(Integer) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- setGcsHttpRequestWriteTimeout(Integer) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- setGcsPerformanceMetrics(Boolean) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- setGcsReadCounterPrefix(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- setGcsRewriteDataOpBatchLimit(Integer) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- setGcsUploadBufferSizeBytes(Integer) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- setGcsUploadBufferSizeBytes(Integer) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions.Builder
-  
- setGcsUtil(GcsUtil) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- setGcsWriteCounterPrefix(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- setGCThrashingPercentagePerPeriod(Double) - Method in interface org.apache.beam.sdk.options.MemoryMonitorOptions
-  
- setGetOffsetFn(SerializableFunction<V, Long>) - Method in class org.apache.beam.sdk.io.cdap.Plugin.Builder
-  
- setGetReceiverArgsFromConfigFn(SerializableFunction<PluginConfig, Object[]>) - Method in class org.apache.beam.sdk.io.cdap.Plugin.Builder
-  
- setGlobalConfigRefreshPeriod(Duration) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setGoogleAdsClientId(String) - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsOptions
-  
- setGoogleAdsClientSecret(String) - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsOptions
-  
- setGoogleAdsCredential(Credentials) - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsOptions
-  
- setGoogleAdsCredentialFactoryClass(Class<? extends CredentialFactory>) - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsOptions
-  
- setGoogleAdsDeveloperToken(String) - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsOptions
-  
- setGoogleAdsEndpoint(String) - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsOptions
-  
- setGoogleAdsRefreshToken(String) - Method in interface org.apache.beam.sdk.io.googleads.GoogleAdsOptions
-  
- setGoogleApiTrace(GoogleApiDebugOptions.GoogleApiTracer) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions
-  
- setGoogleCloudStorageReadOptions(GoogleCloudStorageReadOptions) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- setGroupFilesFileLoad(Boolean) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setGroupingTableMaxSizeMb(int) - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
-  
- setHdfsConfiguration(List<Configuration>) - Method in interface org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptions
-  
- setHeaderColumns(PCollectionView<List<String>>) - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText.Builder
-  
- setHeaderColumns(PCollectionView<List<String>>) - Method in class org.apache.beam.sdk.extensions.ml.DLPInspectText.Builder
-  
- setHeaderColumns(PCollectionView<List<String>>) - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText.Builder
-  
- setHeartbeatMillis(long) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata.Builder
- 
Sets the heartbeat interval in millis. 
- setHoldability(int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- setHooks(DataflowRunnerHooks) - Method in class org.apache.beam.runners.dataflow.DataflowRunner
- 
Sets callbacks to invoke during execution see DataflowRunnerHooks.
 
- setHost(String) - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration.Builder
-  
- setHost(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.CrossLanguageConfiguration
-  
- setHotKeyLoggingEnabled(boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setHttpClient(HttpClient) - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- setHttpClientConfiguration(HttpClientConfiguration) - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
-  
- setHttpHeaders(Map<String, String>) - Method in class org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer
-  
- setHttpPipeline(HttpPipeline) - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- setHTTPReadTimeout(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setHTTPWriteTimeout(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setHumanReadableJsonRecord(String) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Record.Builder
-  
- setId(int) - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
-  
- setId(int) - Method in class org.apache.beam.sdk.extensions.sql.example.model.Order
-  
- setIdAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration.Builder
-  
- setIdAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration.Builder
-  
- setIdLabel(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.Configuration
-  
- setIdLabel(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite.Configuration
-  
- setIdleShutdownTimeout(String) - Method in interface org.apache.beam.runners.prism.PrismPipelineOptions
-  
- setImpersonateServiceAccount(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-  
- setIndex(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration.Builder
-  
- setInferMaps(boolean) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.SchemaConversionOptions.Builder
-  
- setInitialPositionInStream(String) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.ReadDataBuilder.Configuration
-  
- setInitialTimestampInStream(Long) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.ReadDataBuilder.Configuration
-  
- setInput(Input) - Method in class org.apache.beam.sdk.io.cdap.context.BatchSourceContextImpl
- 
Overrides the input configuration of this Batch job to the specified  Input. 
- setInputFile(String) - Method in interface org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.WordCountOptions
-  
- setInsertBundleParallelism(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setInspectConfig(InspectConfig) - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText.Builder
-  
- setInspectConfig(InspectConfig) - Method in class org.apache.beam.sdk.extensions.ml.DLPInspectText.Builder
-  
- setInspectConfig(InspectConfig) - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText.Builder
-  
- setInspectTemplateName(String) - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText.Builder
-  
- setInspectTemplateName(String) - Method in class org.apache.beam.sdk.extensions.ml.DLPInspectText.Builder
-  
- setInspectTemplateName(String) - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText.Builder
-  
- setInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProvider.BigtableReadSchemaTransformConfiguration.Builder
-  
- setInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProvider.BigtableWriteSchemaTransformConfiguration.Builder
-  
- setInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration.Builder
-  
- setInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration.Builder
-  
- setInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.CrossLanguageConfiguration
-  
- setInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration.Builder
-  
- setIsBoundedInternal(PCollection.IsBounded) - Method in class org.apache.beam.sdk.values.PCollection
- 
For internal use only; no backwards-compatibility guarantees. 
- setIsReadSeekEfficient(boolean) - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
-  
- setIsWindmillServiceDirectPathEnabled(boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setJavaAggregateFunctions(ImmutableMap<List<String>, Combine.CombineFn<?, ?, ?>>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.UserFunctionDefinitions.Builder
-  
- setJavaClassLookupAllowlist(JavaClassLookupTransformProvider.AllowList) - Method in interface org.apache.beam.sdk.expansion.service.ExpansionServiceOptions
-  
- setJavaClassLookupAllowlistFile(String) - Method in interface org.apache.beam.sdk.expansion.service.ExpansionServiceOptions
-  
- setJavaScalarFunctions(ImmutableMap<List<String>, UserFunctionDefinitions.JavaScalarFunction>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.UserFunctionDefinitions.Builder
-  
- setJdbcType(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration.Builder
-  
- setJdbcType(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration.Builder
-  
- setJdbcUrl(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration.Builder
-  
- setJdbcUrl(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration.Builder
-  
- setJdkAddOpenModules(List<String>) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setJdkAddOpenModules(List<String>) - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
-  
- setJetDefaultParallelism(Integer) - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
-  
- setJetLocalMode(Integer) - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
-  
- setJetProcessorsCooperative(Boolean) - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
-  
- setJetServers(String) - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
-  
- setJfrRecordingDurationSec(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setJobCheckIntervalInSecs(int) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setJobEndpoint(String) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- setJobFileZip(String) - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- setJobId(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
-  
- setJobLabelsMap(Map<String, String>) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setJobName(String) - Method in interface org.apache.beam.sdk.options.PipelineOptions
-  
- setJobServerConfig(String...) - Method in interface org.apache.beam.runners.portability.testing.TestPortablePipelineOptions
-  
- setJobServerDriver(Class<JobServerDriver>) - Method in interface org.apache.beam.runners.portability.testing.TestPortablePipelineOptions
-  
- setJobServerTimeout(int) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- setJobType(String) - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- setJsonToRowWithErrFn(JsonToRow.JsonToRowWithErrFn) - Method in class org.apache.beam.sdk.transforms.JsonToRow.JsonToRowWithErrFn.ParseWithError.Builder
-  
- setKeep(List<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.Builder
-  
- setKeep(List<String>) - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration.Builder
-  
- setKeep(JavaRowUdf.Configuration) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaFilterTransformProvider.Configuration.Builder
-  
- setKeyDeserializer(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setKeySerializer(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write.External.Configuration
-  
- setKinesisIOConsumerArns(Map<String, String>) - Method in interface org.apache.beam.sdk.io.aws2.kinesis.KinesisIOOptions
-  
- setKmsKey(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransformConfiguration.Builder
-  
- setKmsKey(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.Builder
-  
- setLabels(Map<String, String>) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setLanguage(String) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaFilterTransformProvider.Configuration.Builder
-  
- setLanguage(String) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaMapToFieldsTransformProvider.Configuration.Builder
-  
- setLanguageHint(String) - Method in class org.apache.beam.sdk.extensions.ml.AnnotateText.Builder
-  
- setLastContiguousSequenceRange(ContiguousSequenceRange) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus.Builder
-  
- setLastEventReceived(boolean) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus.Builder
-  
- setLastModifiedMillis(long) - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
-  
- setLastProcessedSequence(Long) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus.Builder
-  
- setLatencyNanos(Long) - Method in class org.apache.beam.sdk.io.solace.data.Solace.PublishResult.Builder
-  
- setLatencyTrackingInterval(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setLatestBufferedSequence(Long) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus.Builder
-  
- setLength(Long) - Method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeVarchar
-  
- setLevel(String) - Method in class org.apache.beam.sdk.schemas.transforms.providers.LoggingTransformProvider.Configuration.Builder
-  
- setLineField(String) - Method in class org.apache.beam.sdk.transforms.JsonToRow.JsonToRowWithErrFn
- 
Sets the field name for the line field in the returned Row. 
- setListeners(List<JavaStreamingListener>) - Method in interface org.apache.beam.runners.spark.SparkContextOptions
-  
- setLoadBalanceBundles(boolean) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- setLocalJobServicePortFile(String) - Method in interface org.apache.beam.runners.portability.testing.TestUniversalRunner.Options
-  
- setLocalWindmillHostport(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setLocation(String) - Method in class org.apache.beam.sdk.extensions.schemaio.expansion.ExternalSchemaIOTransformRegistrar.Configuration
-  
- setLocation(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration.Builder
-  
- setLocation(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration.Builder
-  
- setLocation(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration.Builder
-  
- setLocation(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration.Builder
-  
- setLoginTimeout(String) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setLogMdc(boolean) - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
-  
- setManifestListLocation(String) - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo.Builder
-  
- setMapKeyType(FieldValueTypeInformation) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
-  
- setMapping(Contextful<Contextful.Fn<T, Long>>) - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.Globally.Builder
-  
- setMapping(Contextful<Contextful.Fn<KV<K, V>, KV<K, Long>>>) - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.PerKey.Builder
-  
- setMapValueType(FieldValueTypeInformation) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
-  
- setMaxBufferingDurationMilliSec(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setMaxBundlesFromWindmillOutstanding(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setMaxBundleSize(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setMaxBundleTimeMills(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setMaxBytesFromWindmillOutstanding(long) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setMaxCacheMemoryUsageMb(int) - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
-  
- setMaxCacheMemoryUsageMbClass(Class<? extends SdkHarnessOptions.MaxCacheMemoryUsageMb>) - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
-  
- setMaxCacheMemoryUsagePercent(float) - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
-  
- setMaxCapacityPerShard(Long) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.ReadDataBuilder.Configuration
-  
- setMaxConnectionPoolConnections(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setMaxNumberOfRecords(Long) - Method in class org.apache.beam.io.debezium.DebeziumTransformRegistrar.ReadBuilder.Configuration
-  
- setMaxNumRecords(Long) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.ReadDataBuilder.Configuration
-  
- setMaxNumRecords(Long) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setMaxNumWorkers(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-  
- setMaxOutputElementsPerBundle(int) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler
- 
Overrides the default value. 
- setMaxParallelism(Integer) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setMaxReadTime(Long) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.ReadDataBuilder.Configuration
-  
- setMaxReadTime(Long) - Method in class org.apache.beam.sdk.io.GenerateSequence.External.ExternalConfiguration
-  
- setMaxReadTime(Long) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setMaxReadTimeSeconds(Integer) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration.Builder
-  
- setMaxRecordsPerBatch(Long) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- setMaxStackTraceDepthToReport(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setMaxStreamingBatchSize(Long) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setMaxStreamingRowsToBatch(Long) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setMemoryMB(int) - Method in class org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options
- 
Sets the size of the memory buffer in megabytes. 
- setMessageId(Long) - Method in class org.apache.beam.sdk.extensions.sbe.SbeSchema.IrOptions.Builder
-  
- setMessageId(String) - Method in class org.apache.beam.sdk.io.solace.data.Solace.CorrelationKey.Builder
-  
- setMessageId(String) - Method in class org.apache.beam.sdk.io.solace.data.Solace.PublishResult.Builder
-  
- setMessageId(String) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record.Builder
-  
- setMessageName(String) - Method in class org.apache.beam.sdk.extensions.sbe.SbeSchema.IrOptions.Builder
-  
- setMessageName(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration.Builder
-  
- setMessageName(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration.Builder
-  
- setMessageName(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration.Builder
-  
- setMessageName(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration.Builder
-  
- setMessageRecord(Object) - Method in class org.apache.beam.sdk.io.pulsar.PulsarMessage
-  
- setMetadataTable(String) - Method in interface org.apache.beam.sdk.io.gcp.spanner.SpannerIO.SpannerChangeStreamOptions
- 
Specifies the name of the metadata table. 
- setMethod(Method) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
-  
- setMetricsGraphiteHost(String) - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
-  
- setMetricsGraphitePort(Integer) - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
-  
- setMetricsHttpSinkUrl(String) - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
-  
- setMetricsPushPeriod(Long) - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
-  
- setMetricsSink(Class<? extends MetricsSink>) - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
-  
- setMetricsSupported(boolean) - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
- 
Called by the run to indicate whether metrics reporting is supported. 
- setMimeType(String) - Method in class org.apache.beam.sdk.io.fs.CreateOptions.Builder
-  
- setMinConnectionPoolConnections(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setMinCpuPlatform(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-  
- setMinPauseBetweenCheckpoints(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setMinReadTimeMillis(Long) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- setName(String) - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
-  
- setName(String) - Method in class org.apache.beam.sdk.io.snowflake.data.SnowflakeColumn
-  
- setName(String) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Destination.Builder
-  
- setName(String) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
-  
- setName(String) - Method in class org.apache.beam.sdk.schemas.Schema.Field.Builder
-  
- setName(String) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaRowUdf.Configuration.Builder
-  
- setName(String) - Method in class org.apache.beam.sdk.values.PCollection
- 
- setName(String) - Method in class org.apache.beam.sdk.values.PValueBase
- 
- setNetwork(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-  
- setNetworkTimeout(Executor, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- setNullable(boolean) - Method in class org.apache.beam.sdk.io.snowflake.data.SnowflakeColumn
-  
- setNullable(boolean) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
-  
- setNumber(Integer) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
-  
- setNumberOfBufferedEvents(long) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus.Builder
-  
- setNumberOfExecutionRetries(Integer) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setNumberOfReceivedEvents(long) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus.Builder
-  
- setNumberOfWorkerHarnessThreads(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setNumConcurrentCheckpoints(int) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setNumFailuresExpected(int) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
-  
- setNumSampledBytesPerFile(long) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
-  
- setNumShards(Integer) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.Builder
- 
The number of output shards produced; a value of 1 disables sharding. 
- setNumStorageWriteApiStreamAppendClients(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setNumStorageWriteApiStreams(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setNumStreamingKeys(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setNumStreams(Integer) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.Builder
-  
- setNumWorkers(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-  
- setOAuthToken(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setOauthToken(String) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setObjectReuse(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setOffsetDeduplication(Boolean) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setOnCreateMatcher(SerializableMatcher<PipelineResult>) - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-  
- setOneOfTypes(Map<String, FieldValueTypeInformation>) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
-  
- setOnly(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.Builder
-  
- setOnly(String) - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration.Builder
-  
- setOnSuccessMatcher(SerializableMatcher<PipelineResult>) - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-  
- setOperation(String) - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo.Builder
-  
- setOperatorChaining(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- SetOperatorFilteringDoFn(String, String, BeamSetOperatorRelBase.OpType, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSetOperatorsTransforms.SetOperatorFilteringDoFn
-  
- setOption(String, Row) - Method in class org.apache.beam.sdk.schemas.Schema.Options.Builder
-  
- setOption(String, Schema.FieldType, Object) - Method in class org.apache.beam.sdk.schemas.Schema.Options.Builder
-  
- setOption(String, Schema.FieldType, Object) - Static method in class org.apache.beam.sdk.schemas.Schema.Options
-  
- setOption(String, Row) - Static method in class org.apache.beam.sdk.schemas.Schema.Options
-  
- setOptions(ImmutableMap<String, String>) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setOptions(Schema.Options) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
- 
Returns a copy of the Field with isNullable set. 
- setOptions(Schema.Options.Builder) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
-  
- setOptions(Schema.Options) - Method in class org.apache.beam.sdk.schemas.Schema.Field.Builder
-  
- setOptions(Schema.Options.Builder) - Method in class org.apache.beam.sdk.schemas.Schema.Field.Builder
-  
- setOptionsId(long) - Method in interface org.apache.beam.sdk.options.PipelineOptions
-  
- setOutput(String) - Method in interface org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.WordCountOptions
-  
- setOutput(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.ErrorHandling.Builder
-  
- setOutput(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration.ErrorHandling.Builder
-  
- setOutput(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration.ErrorHandling.Builder
-  
- setOutput(String) - Method in class org.apache.beam.sdk.schemas.transforms.providers.ErrorHandling.Builder
-  
- setOutputDataSet(PCollection<T>, TSet<WindowedValue<T>>) - Method in class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- setOutputExecutablePath(String) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- setOutputFilePrefix(String) - Method in interface org.apache.beam.sdk.extensions.sql.example.BeamSqlDataCatalogExample.DCExamplePipelineOptions
-  
- setOutputParallelization(Boolean) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration.Builder
-  
- setOutputParallelization(Boolean) - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadConfiguration.Builder
-  
- setOverrideWindmillBinary(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setParallelism(Integer) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setParallelism(int) - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- setParam(String, Object) - Method in class org.apache.beam.sdk.io.cdap.ConfigWrapper
- 
Sets a  Plugin single parameter. 
- setParameters(T, PreparedStatement) - Method in interface org.apache.beam.sdk.io.jdbc.JdbcIO.PreparedStatementSetter
-  
- setParameters(PreparedStatement) - Method in interface org.apache.beam.sdk.io.jdbc.JdbcIO.StatementPreparator
-  
- setParameters(KV<PartitionT, PartitionT>, PreparedStatement) - Method in interface org.apache.beam.sdk.io.jdbc.JdbcReadWithPartitionsHelper
-  
- setParameters(PreparedStatement) - Method in interface org.apache.beam.sdk.io.singlestore.SingleStoreIO.StatementPreparator
-  
- setParentId(Long) - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo.Builder
-  
- setParentTokens(HashSet<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata.Builder
- 
Sets the collection of parent partition identifiers. 
- setParquetConfiguration(FileWriteSchemaTransformConfiguration.ParquetConfiguration) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.Builder
- 
Configures extra details related to writing Parquet formatted files. 
- setPartitionKey(String) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.WriteBuilder.Configuration
-  
- setPartitionSpec(PartitionSpec) - Method in class org.apache.beam.sdk.io.iceberg.IcebergTableCreateConfig.Builder
-  
- setPartitionToken(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata.Builder
- 
Sets the unique partition identifier. 
- setPassword(String) - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration.Builder
-  
- setPassword(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration.Builder
-  
- setPassword(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration.Builder
-  
- setPassword(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setPassword(ValueProvider<String>) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setPath(String) - Method in class org.apache.beam.sdk.io.csv.providers.CsvWriteTransformProvider.CsvWriteConfiguration.Builder
-  
- setPath(String) - Method in class org.apache.beam.sdk.io.json.providers.JsonWriteTransformProvider.JsonWriteConfiguration.Builder
-  
- setPath(String) - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaRowUdf.Configuration.Builder
-  
- setPathValidator(PathValidator) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- setPathValidatorClass(Class<? extends PathValidator>) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
-  
- setPayload(byte[]) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWritePayload.Builder
-  
- setPayload(byte[]) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record.Builder
-  
- setPayload(byte[]) - Method in class org.apache.beam.sdk.schemas.io.Failure.Builder
-  
- setPeriod(Long) - Method in class org.apache.beam.sdk.io.GenerateSequence.External.ExternalConfiguration
-  
- setPeriodicStatusPageOutputDirectory(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setPerWorkerMetricsUpdateReportingPeriodMillis(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setPipelineOption(String, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- setPipelineOptions(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
-  
- setPipelineOptions(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.impl.JdbcConnection
-  
- setPipelineOptionsMap(Map<String, String>) - Method in class org.apache.beam.sdk.extensions.sql.impl.JdbcConnection
- 
- setPipelinePolicy(HttpPipelinePolicy) - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- setPipelineUrl(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setPlannerName(String) - Method in interface org.apache.beam.sdk.extensions.sql.impl.BeamSqlPipelineOptions
-  
- setPluginClass(Class<?>) - Method in class org.apache.beam.sdk.io.cdap.Plugin.Builder
-  
- setPluginType(PluginConstants.PluginType) - Method in class org.apache.beam.sdk.io.cdap.Plugin.Builder
-  
- setPollIntervalMillis(Long) - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformConfiguration.Builder
-  
- setPort(Integer) - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration.Builder
-  
- setPort(int) - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver.ServerConfiguration
-  
- setPortNumber(String) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setPrecision(int) - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.Globally.Builder
-  
- setPrecision(Integer) - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.PerKey.Builder
-  
- setPrecision(int) - Method in class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeNumber
-  
- setPredefinedCsvFormat(String) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.CsvConfiguration.Builder
- 
The CSVFormat.Predefined#name()of the written CSV file.
 
- setPreferGroupByKeyToHandleHugeValues(Boolean) - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
-  
- setPrefix(String) - Method in class org.apache.beam.sdk.schemas.transforms.providers.LoggingTransformProvider.Configuration.Builder
-  
- setPrimaryKey(List<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.Builder
-  
- setPrimaryKey(TableReference, List<String>) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- setPriority(int) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record.Builder
-  
- setPrismLocation(String) - Method in interface org.apache.beam.runners.prism.PrismPipelineOptions
-  
- setPrismLogLevel(String) - Method in interface org.apache.beam.runners.prism.PrismPipelineOptions
-  
- setPrismVersionOverride(String) - Method in interface org.apache.beam.runners.prism.PrismPipelineOptions
-  
- setPrivateKeyPassphrase(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setPrivateKeyPassphrase(ValueProvider<String>) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setPrivateKeyPath(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setPrivateKeyPath(String) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setProcessWideContainer(MetricsContainer) - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
- 
- setProducerConfig(Map<String, String>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write.External.Configuration
-  
- setProducerConfigUpdates(Map<String, String>) - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration.Builder
-  
- setProduceStatusUpdateOnEveryEvent(boolean) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler
- 
Sets the indicator of whether the status notification needs to be produced on every event. 
- setProfilingAgentConfiguration(DataflowProfilingOptions.DataflowProfilingAgentConfiguration) - Method in interface org.apache.beam.runners.dataflow.options.DataflowProfilingOptions
-  
- setProject(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setProject(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-  
- setProject(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration.Builder
-  
- setProject(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration.Builder
-  
- setProjectId(String) - Method in class org.apache.beam.sdk.extensions.ml.DLPDeidentifyText.Builder
-  
- setProjectId(String) - Method in class org.apache.beam.sdk.extensions.ml.DLPInspectText.Builder
-  
- setProjectId(String) - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText.Builder
-  
- setProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProvider.BigtableReadSchemaTransformConfiguration.Builder
-  
- setProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProvider.BigtableWriteSchemaTransformConfiguration.Builder
-  
- setProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration.Builder
-  
- setProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration.Builder
-  
- setProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.CrossLanguageConfiguration
-  
- setProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration.Builder
-  
- setProvidedSparkContext(JavaSparkContext) - Method in interface org.apache.beam.runners.spark.SparkContextOptions
-  
- setProviderRuntimeValues(ValueProvider<Map<String, Object>>) - Method in interface org.apache.beam.sdk.testing.TestPipeline.TestValueProviderOptions
-  
- setProxyConfiguration(ProxyConfiguration) - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
-  
- setPublished(Boolean) - Method in class org.apache.beam.sdk.io.solace.data.Solace.PublishResult.Builder
-  
- setPublishMonotonicNanos(long) - Method in class org.apache.beam.sdk.io.solace.data.Solace.CorrelationKey.Builder
-  
- setPubsubRootUrl(String) - Method in interface org.apache.beam.sdk.io.gcp.pubsub.PubsubOptions
-  
- setQualifiers(List<FieldAccessDescriptor.FieldDescriptor.Qualifier>) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Builder
-  
- setQuery(String) - Method in class org.apache.beam.sdk.extensions.sql.expansion.ExternalSqlTransformRegistrar.Configuration
-  
- setQuery(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryExportReadSchemaTransformConfiguration.Builder
- 
Configures the BigQuery read job with the SQL query. 
- setQuery(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransformConfiguration.Builder
-  
- setQuery(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration.Builder
-  
- setQuery(String) - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadConfiguration.Builder
-  
- setQuery(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setQuery(ValueProvider<String>) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setQueryLocation(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryExportReadSchemaTransformConfiguration.Builder
- 
BigQuery geographic location where the query job will be executed. 
- setQueryPlannerClassName(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
-  
- setQueryString(String) - Method in interface org.apache.beam.sdk.extensions.sql.example.BeamSqlDataCatalogExample.DCExamplePipelineOptions
-  
- setQueue(Queue) - Method in class org.apache.beam.sdk.io.solace.broker.SessionServiceFactory
- 
- setRamMegaBytes(int) - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- setRate(GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration.Rate) - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration.Builder
-  
- setRateLimit(Long) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.ReadDataBuilder.Configuration
-  
- setRawPrivateKey(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setRawPrivateKey(ValueProvider<String>) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setRawType(Class<?>) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
-  
- setReadChangeStreamTimeout(Duration) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.BigtableChangeStreamAccessor
-  
- setReadChangeStreamTimeout(Duration) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.DaoFactory
-  
- setReaderCacheTimeoutSec(Integer) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setReadOnly(boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- setReadQuery(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration.Builder
-  
- setReadTimeout(int) - Method in class org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer
-  
- setReadTimePercentage(Double) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- setReadTimestamp(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.ReadBuilder.Configuration
-  
- setReceiverClass(Class<? extends Receiver<V>>) - Method in class org.apache.beam.sdk.io.cdap.Plugin.Builder
-  
- setReceiveTimestamp(Long) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record.Builder
-  
- setRecord(BadRecord.Record) - Method in class org.apache.beam.sdk.transforms.errorhandling.BadRecord.Builder
-  
- setRecordJfrOnGcThrashing(boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setRedelivered(boolean) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record.Builder
-  
- setRedistribute(Boolean) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setRedistributeNumKeys(Integer) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setRegion(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setReidentifyConfig(DeidentifyConfig) - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText.Builder
-  
- setReidentifyTemplateName(String) - Method in class org.apache.beam.sdk.extensions.ml.DLPReidentifyText.Builder
-  
- setReIterableGroupByKeyResult(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setRelative() - Method in interface org.apache.beam.sdk.state.Timer
- 
Sets the timer relative to the current time, according to any offset and alignment specified. 
- setReplicationGroupMessageId(String) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record.Builder
-  
- setReplyTo(Solace.Destination) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record.Builder
-  
- setReportCheckpointDuration(String) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setRequestRecordsLimit(Long) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.ReadDataBuilder.Configuration
-  
- setResourceHints(ResourceHints) - Method in class org.apache.beam.sdk.transforms.PTransform
- 
Sets resource hints for the transform. 
- setResourceHints(List<String>) - Method in interface org.apache.beam.sdk.transforms.resourcehints.ResourceHintsOptions
-  
- setResourceId(ResourceId) - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
-  
- setResultCount(long) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus.Builder
-  
- setRetainDockerContainers(boolean) - Method in interface org.apache.beam.sdk.options.ManualDockerEnvironmentOptions
-  
- setRetainExternalizedCheckpointsOnCancellation(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setRole(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setRole(ValueProvider<String>) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setRootElement(String) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.XmlConfiguration.Builder
- 
Sets the enclosing root element for the generated XML files. 
- setRowGroupSize(Integer) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.ParquetConfiguration.Builder
- 
Specify row-group size; if not set or zero, a default is used by the underlying writer. 
- setRowRestriction(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransformConfiguration.Builder
-  
- setRowSchema(Schema) - Method in class org.apache.beam.sdk.values.PCollection
- 
Sets a schema on this PCollection. 
- setRuleSets(Collection<RuleSet>) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
- 
Set the ruleSet used for query optimizer. 
- setRunner(Class<? extends PipelineRunner<?>>) - Method in interface org.apache.beam.sdk.options.PipelineOptions
-  
- setRunnerDeterminedSharding(boolean) - Method in interface org.apache.beam.runners.direct.DirectTestOptions
-  
- setRunningAt(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata.Builder
- 
Sets the time at which the partition started running. 
- Sets - Class in org.apache.beam.sdk.transforms
- 
The  PTransforms that allow to compute different set functions across  PCollections. 
- Sets() - Constructor for class org.apache.beam.sdk.transforms.Sets
-  
- sets(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
- 
- setS3ClientBuilder(S3ClientBuilder) - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration.Builder
-  
- setS3ClientFactoryClass(Class<? extends S3ClientBuilderFactory>) - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- setS3StorageClass(String) - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- setS3StorageClass(String) - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration.Builder
-  
- setS3ThreadPoolSize(int) - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- setS3ThreadPoolSize(int) - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration.Builder
-  
- setS3UploadBufferSizeBytes(Integer) - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- setS3UploadBufferSizeBytes(int) - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration.Builder
-  
- setSamplingStrategy(TextRowCountEstimator.SamplingStrategy) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
-  
- setSasToken(String) - Method in interface org.apache.beam.sdk.io.azure.options.BlobstoreOptions
-  
- setSaveHeapDumpsToGcsPath(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setSavepoint() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- setSavepoint(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- setSavepointPath(String) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setSaveProfilesToGcs(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowProfilingOptions
-  
- setScale(int) - Method in class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeNumber
-  
- setScanType(IcebergScanConfig.ScanType) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setScheduledAt(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata.Builder
- 
Sets the time at which the partition was scheduled. 
- setScheduledExecutorService(ScheduledExecutorService) - Method in interface org.apache.beam.sdk.options.ExecutorOptions
-  
- setSchema(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- setSchema(String) - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformConfiguration.Builder
-  
- setSchema(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration.Builder
-  
- setSchema(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration.Builder
-  
- setSchema(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration.Builder
-  
- setSchema(byte[]) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.ReadBuilder.Configuration
-  
- setSchema(Schema) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setSchema(Schema) - Method in class org.apache.beam.sdk.io.iceberg.IcebergTableCreateConfig.Builder
-  
- setSchema(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration.Builder
-  
- setSchema(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration.Builder
-  
- setSchema(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setSchema(ValueProvider<String>) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setSchema(Schema, TypeDescriptor<T>, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Method in class org.apache.beam.sdk.values.PCollection
- 
- setSchemaId(Integer) - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo.Builder
-  
- setSchemaIfNotPresent(String, Schema) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
-  
- setSchematizedData(String) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2Message
-  
- setScheme(String) - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration.Builder
-  
- setSdkContainerImage(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-  
- setSdkHarnessContainerImageOverrides(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setSdkHarnessLogLevelOverrides(SdkHarnessOptions.SdkHarnessLogLevelOverrides) - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
-  
- setSdkWorkerParallelism(int) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
-  
- setSeconds(Long) - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration.Rate.Builder
-  
- setSelectedFields(List<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransformConfiguration.Builder
-  
- setSemiPersistDir(String) - Method in interface org.apache.beam.sdk.options.RemoteEnvironmentOptions
-  
- setSenderTimestamp(Long) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record.Builder
-  
- setSequenceNumber(long) - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo.Builder
-  
- setSequenceNumber(Long) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record.Builder
-  
- setSerializedWindowingStrategy(byte[]) - Method in class org.apache.beam.sdk.expansion.service.WindowIntoTransformProvider.Configuration.Builder
-  
- setServerName(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setServerName(ValueProvider<String>) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setServiceAccount(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setShardNameTemplate(String) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.Builder
- 
- setShouldFailRow(Function<TableRow, Boolean>) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- setShutdownSourcesAfterIdleMs(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setSideInput(PCollectionView<T>, BoundedWindow, T) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- setSideInputDataSet(String, BatchTSet<WindowedValue<ElemT>>) - Method in class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- setSideInputs(Map<PCollectionView<?>, Map<BoundedWindow, ?>>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- setSize(Long) - Method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeBinary
-  
- setSizeBytes(long) - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
-  
- setSizeEstimator(CoderSizeEstimator<KV<ByteString, ChangeStreamRecord>>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.ReadChangeStreamPartitionDoFn
- 
Sets the estimator to track throughput for each DoFn instance. 
- setSkipHeaderLines(int) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
-  
- setSnapshot(Long) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setSnapshotId(long) - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo.Builder
-  
- setSnowPipe(ValueProvider<String>) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setSorterType(ExternalSorter.Options.SorterType) - Method in class org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options
- 
Sets the sorter type. 
- setSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction.Builder
-  
- setSparkMaster(String) - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
-  
- setSql(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.ReadBuilder.Configuration
-  
- setSqlScalarFunctions(ImmutableMap<List<String>, ResolvedNodes.ResolvedCreateFunctionStmt>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.UserFunctionDefinitions.Builder
-  
- setSqlTableValuedFunctions(ImmutableMap<List<String>, ResolvedNode>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.UserFunctionDefinitions.Builder
-  
- setSSEAlgorithm(String) - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- setSSEAlgorithm(String) - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration.Builder
-  
- setSSECustomerKey(SSECustomerKey) - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- setSSECustomerKey(SSECustomerKey) - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration.Builder
-  
- setSSEKMSKeyId(String) - Method in interface org.apache.beam.sdk.io.aws2.options.S3Options
-  
- setSSEKMSKeyId(String) - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration.Builder
-  
- setStableUniqueNames(PipelineOptions.CheckEnabled) - Method in interface org.apache.beam.sdk.options.PipelineOptions
-  
- setStager(Stager) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setStagerClass(Class<? extends Stager>) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setStagingBucketName(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setStagingBucketName(ValueProvider<String>) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setStagingLocation(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setStaleness(Long) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.ReadBuilder.Configuration
-  
- setStart(Long) - Method in class org.apache.beam.sdk.io.GenerateSequence.External.ExternalConfiguration
-  
- setStart(Long) - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration.Builder
-  
- setStartAtTimestamp(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration.Builder
-  
- setStartOffset(Long) - Method in interface org.apache.beam.sdk.io.sparkreceiver.HasOffset
-  
- setStartReadTime(Long) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setStartTimestamp(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata.Builder
- 
Sets the start time of the partition. 
- setState(PipelineResult.State) - Method in class org.apache.beam.runners.twister2.Twister2PipelineResult
-  
- setState(PartitionMetadata.State) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata.Builder
- 
Sets the current state of the partition. 
- SetState<T> - Interface in org.apache.beam.sdk.state
- 
- setStateBackend(String) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setStateBackendFactory(Class<? extends FlinkStateBackendFactory>) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
- 
- setStateBackendStoragePath(String) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setStatistics(BeamTableStatistics) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestUnboundedTable
-  
- setStatusDate(Instant) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingStatus.Builder
-  
- setStatusUpdateFrequency(Duration) - Method in class org.apache.beam.sdk.extensions.ordered.OrderedProcessingHandler
- 
Changes the default status update frequency. 
- setStop(Long) - Method in class org.apache.beam.sdk.io.GenerateSequence.External.ExternalConfiguration
-  
- setStopPipelineWatermark(Long) - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
-  
- setStopReadTime(Long) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setStorageApiAppendThresholdBytes(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setStorageApiAppendThresholdRecordCount(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setStorageClient(Storage) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
-  
- setStorageIntegrationName(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setStorageIntegrationName(ValueProvider<String>) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setStorageLevel(String) - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
-  
- setStorageWriteApiMaxRequestSize(Long) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setStorageWriteApiMaxRetries(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setStorageWriteApiTriggeringFrequencySec(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setStorageWriteMaxInflightBytes(Long) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setStorageWriteMaxInflightRequests(Long) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setStreaming(boolean) - Method in interface org.apache.beam.sdk.options.StreamingOptions
-  
- setStreamingSideInputCacheExpirationMillis(Integer) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setStreamingSideInputCacheMb(Integer) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setStreamingTimeoutMs(Long) - Method in interface org.apache.beam.runners.spark.SparkPortableStreamingPipelineOptions
-  
- setStuckCommitDurationMillis(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setSubmissionMode(SolaceIO.SubmissionMode) - Method in class org.apache.beam.sdk.io.solace.broker.SessionServiceFactory
- 
Called by the write connector to set the submission mode used to create the message producers. 
- setSubnetwork(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-  
- setSubscription(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.Configuration
-  
- setSubscription(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration.Builder
-  
- setSubscriptionName(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration.Builder
-  
- setSubscriptionPath(SubscriptionPath) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.SubscriberOptions.Builder
-  
- setSummary(Map<String, String>) - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo.Builder
-  
- setSupportKafkaMetrics(boolean) - Static method in class org.apache.beam.sdk.io.kafka.KafkaSinkMetrics
-  
- setSupportMetricsDeletion(boolean) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySinkMetrics
-  
- setSupportStreamingInsertsMetrics(boolean) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySinkMetrics
-  
- setTable(String) - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration.Builder
-  
- setTable(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.Builder
-  
- setTable(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration.Builder
-  
- setTable(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.ReadBuilder.Configuration
-  
- setTable(String) - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration.Builder
-  
- setTable(String) - Method in class org.apache.beam.sdk.io.iceberg.SchemaTransformConfiguration.Builder
-  
- setTable(String) - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadConfiguration.Builder
-  
- setTable(String) - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteConfiguration.Builder
-  
- setTable(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setTable(ValueProvider<String>) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- settableArguments - Variable in class org.apache.beam.sdk.io.cdap.context.BatchContextImpl
- 
- setTableCreateConfig(IcebergTableCreateConfig) - Method in class org.apache.beam.sdk.io.iceberg.IcebergDestination.Builder
-  
- setTableId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProvider.BigtableReadSchemaTransformConfiguration.Builder
-  
- setTableId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProvider.BigtableWriteSchemaTransformConfiguration.Builder
-  
- setTableId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration.Builder
-  
- setTableId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration.Builder
-  
- setTableIdentifier(TableIdentifier) - Method in class org.apache.beam.sdk.io.iceberg.IcebergDestination.Builder
-  
- setTableIdentifier(String) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setTableIdentifier(TableIdentifier) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setTableIdentifier(String...) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setTableSchema(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.WriteBuilder.Configuration
-  
- setTableSpec(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryExportReadSchemaTransformConfiguration.Builder
- 
Specifies a table for a BigQuery read job. 
- setTableSpec(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransformConfiguration.Builder
-  
- setTag(String) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setTargetDataset(String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.TestBigQueryOptions
-  
- setTargetParallelism(int) - Method in interface org.apache.beam.runners.direct.DirectOptions
-  
- setTempDatasetId(String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setTemplateLocation(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setTempLocation(String) - Method in class org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options
- 
Sets the path to a temporary location where the sorter writes intermediate files. 
- setTempLocation(String) - Method in interface org.apache.beam.sdk.options.PipelineOptions
-  
- setTempRoot(String) - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-  
- SETTER_WITH_NULL_METHOD_ERROR - Static variable in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
-  
- setTerminateAfterSecondsSinceNewOutput(Long) - Method in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformConfiguration.Builder
-  
- SetterTypeSupplier() - Constructor for class org.apache.beam.sdk.schemas.JavaBeanSchema.SetterTypeSupplier
-  
- setTestMode(boolean) - Method in interface org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineOptions
-  
- setTestTimeoutSeconds(Long) - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
-  
- setThroughputEstimator(BytesThroughputEstimator<DataChangeRecord>) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.ReadChangeStreamPartitionDoFn
- 
Sets the estimator to calculate the backlog of this function. 
- setTimer(TimerInternals.TimerData) - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-  
- setTimer(StateNamespace, String, String, Instant, Instant, TimeDomain) - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-  
- setTimer(Instant, TimeDomain) - Method in interface org.apache.beam.sdk.state.Timers
- 
Sets a timer to fire when the event time watermark, the current processing time, or the
 synchronized processing time watermark surpasses a given timestamp. 
- setTimer(Instant, Instant, TimeDomain) - Method in interface org.apache.beam.sdk.state.Timers
- 
Set a timer with outputTimestamp. 
- setTimestamp(Instant) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWritePayload.Builder
-  
- setTimestamp(Long) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setTimestampAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.Configuration
-  
- setTimestampAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite.Configuration
-  
- setTimestampAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration.Builder
-  
- setTimestampAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration.Builder
-  
- setTimestampBound(TimestampBound) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction.Builder
-  
- setTimestampBoundMode(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.ReadBuilder.Configuration
-  
- setTimestampMillis(long) - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo.Builder
-  
- setTimestampPolicy(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setTimeSupplier(Supplier<Timestamp>) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRangeTracker
-  
- setTimeToLive(long) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Record.Builder
-  
- setTimeUnit(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar.ReadBuilder.Configuration
-  
- setTopic(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.Configuration
-  
- setTopic(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite.Configuration
-  
- setTopic(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformConfiguration.Builder
-  
- setTopic(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformConfiguration.Builder
-  
- setTopic(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write.External.Configuration
-  
- setTopic(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration.Builder
- 
Sets the topic from which to read. 
- setTopic(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider.KafkaWriteSchemaTransformConfiguration.Builder
-  
- setTopicName(String) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration.Builder
-  
- setTopicPath(TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PublisherOptions.Builder
-  
- setTopics(List<String>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setToSnapshot(Long) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setToSnapshotRef(String) - Method in class org.apache.beam.sdk.io.iceberg.IcebergScanConfig.Builder
-  
- setTransactionIsolation(int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- setTransformNameMapping(Map<String, String>) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setTriggeringFrequencySeconds(Long) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.Builder
-  
- setTriggeringFrequencySeconds(Integer) - Method in class org.apache.beam.sdk.io.iceberg.IcebergWriteSchemaTransformProvider.Configuration.Builder
-  
- setTruncateTimestamps(boolean) - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogPipelineOptions
-  
- setTruncateTimestamps(BigQueryUtils.ConversionOptions.TruncateTimestamps) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.ConversionOptions.Builder
-  
- setTSetEnvironment(TSetEnvironment) - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- setTwister2Home(String) - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- setType(Solace.DestinationType) - Method in class org.apache.beam.sdk.io.solace.data.Solace.Destination.Builder
-  
- setType(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
-  
- setType(Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Field.Builder
-  
- setTypeDescriptor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.values.PCollection
- 
- setTypeMap(Map<String, Class<?>>) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
-  
- setUint16Behavior(UnsignedOptions.Behavior) - Method in class org.apache.beam.sdk.extensions.sbe.UnsignedOptions.Builder
-  
- setUint32Behavior(UnsignedOptions.Behavior) - Method in class org.apache.beam.sdk.extensions.sbe.UnsignedOptions.Builder
-  
- setUint64Behavior(UnsignedOptions.Behavior) - Method in class org.apache.beam.sdk.extensions.sbe.UnsignedOptions.Builder
-  
- setUint8Behavior(UnsignedOptions.Behavior) - Method in class org.apache.beam.sdk.extensions.sbe.UnsignedOptions.Builder
-  
- setUnalignedCheckpointEnabled(boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
-  
- setUnboundedReaderMaxElements(Integer) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setUnboundedReaderMaxReadTimeMs(Integer) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setUnboundedReaderMaxReadTimeSec(Integer) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setUnboundedReaderMaxWaitForElementsMs(Integer) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setUnknownFieldsPayload(byte[]) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWritePayload.Builder
-  
- setup() - Method in interface org.apache.beam.io.requestresponse.SetupTeardown
- 
Called during the  DoFn's setup lifecycle method. 
- setup(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides.BatchStatefulDoFn
-  
- setUp(Schema) - Method in interface org.apache.beam.sdk.extensions.sql.BeamSqlSeekableTable
- 
prepare the instance. 
- setup() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.DetectNewPartitionsDoFn
-  
- setup() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dofn.ReadChangeStreamPartitionDoFn
-  
- setup() - Method in class org.apache.beam.sdk.io.gcp.datastore.RampupThrottlingFn
-  
- setup() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.DetectNewPartitionsDoFn
- 
- setup() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dofn.ReadChangeStreamPartitionDoFn
- 
- setup() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadSpannerSchema
-  
- setUp() - Static method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
-  
- setUp() - Static method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
-  
- setup() - Method in interface org.apache.beam.sdk.io.kafka.CheckStopReadingFn
-  
- setup() - Method in class org.apache.beam.sdk.io.pulsar.WriteToPulsarDoFn
-  
- setup() - Method in class org.apache.beam.sdk.jmh.schemas.RowBundle
-  
- setup() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark.NewVsCopy.ArrayCopyState
-  
- setup() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark.NewVsCopy.ArrayNewState
-  
- setup(Blackhole) - Method in class org.apache.beam.sdk.jmh.util.VarIntBenchmark.BlackholeOutput
-  
- setup() - Method in class org.apache.beam.sdk.jmh.util.VarIntBenchmark.Bytes
-  
- setup() - Method in class org.apache.beam.sdk.jmh.util.VarIntBenchmark.Longs
-  
- setup() - Method in class org.apache.beam.sdk.transforms.Reshuffle.AssignShardFn
-  
- setUpdate(boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
-  
- setUpdateCompatibilityVersion(String) - Method in interface org.apache.beam.sdk.options.StreamingOptions
-  
- setUploadBufferSizeBytes(int) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.CreateOptions.Builder
-  
- setupModule(Module.SetupContext) - Method in class org.apache.beam.sdk.io.aws2.options.AwsModule
-  
- SetupTeardown - Interface in org.apache.beam.io.requestresponse
- 
Provided by user and called within  DoFn.Setup and @{link
 org.apache.beam.sdk.transforms.DoFn.Teardown} lifecycle methods of  Call's  DoFn. 
- setUpToDateThreshold(Long) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.ReadDataBuilder.Configuration
-  
- setUrl(String) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setUseActiveSparkSession(boolean) - Method in interface org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineOptions
-  
- setUseAltsServer(boolean) - Method in interface org.apache.beam.sdk.expansion.service.ExpansionServiceOptions
-  
- setUseAtLeastOnceSemantics(Boolean) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.Builder
-  
- setUseCdcWrites(Boolean) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.Builder
-  
- setUseDataStreamForBatch(Boolean) - Method in interface org.apache.beam.runners.flink.VersionDependentFlinkPipelineOptions
-  
- setUsePublicIps(Boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-  
- setUserAgent(String) - Method in interface org.apache.beam.sdk.options.PipelineOptions
-  
- setUsername(String) - Method in class org.apache.beam.io.debezium.DebeziumReadSchemaTransformProvider.DebeziumReadSchemaTransformConfiguration.Builder
-  
- setUsername(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration.Builder
-  
- setUsername(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration.Builder
-  
- setUsername(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setUsername(ValueProvider<String>) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setUseSeparateWindmillHeartbeatStreams(Boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setUsesProvidedSparkContext(boolean) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
-  
- setUseStandardSql(Boolean) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryExportReadSchemaTransformConfiguration.Builder
- 
Enables BigQuery's Standard SQL dialect when reading from a query. 
- setUseStorageApiConnectionPool(Boolean) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setUseStorageWriteApi(Boolean) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setUseStorageWriteApiAtLeastOnce(Boolean) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
-  
- setUseTransformService(boolean) - Method in interface org.apache.beam.sdk.extensions.python.PythonExternalTransformOptions
-  
- setUseWindmillIsolatedChannels(Boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setUuid(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.PartitionRecord
-  
- setUUID(UUID) - Method in class org.apache.beam.sdk.schemas.Schema
- 
Set this schema's UUID. 
- setUuidExtractor(SerializableFunction<SequencedMessage, Uuid>) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.UuidDeduplicationOptions.Builder
-  
- SetUuidFn(String) - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.SetUuidFromPubSubMessage.SetUuidFn
-  
- SetUuidFromPubSubMessage(String) - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.SetUuidFromPubSubMessage
-  
- setValueDeserializer(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
-  
- setValueSerializer(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write.External.Configuration
-  
- setVerifyRowValues(Boolean) - Method in interface org.apache.beam.sdk.extensions.sql.impl.BeamSqlPipelineOptions
-  
- setWarehouse(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.CrossLanguageConfiguration
-  
- setWarehouse(ValueProvider<String>) - Method in interface org.apache.beam.sdk.io.snowflake.SnowflakePipelineOptions
-  
- setWatermark(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata.Builder
- 
Sets the watermark (last processed timestamp) for the partition. 
- setWatermark(Instant) - Method in interface org.apache.beam.sdk.transforms.splittabledofn.ManualWatermarkEstimator
- 
Sets a timestamp before or at the timestamps of all future elements produced by the associated
 DoFn. 
- setWatermark(Instant) - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.Manual
-  
- setWatermarkIdleDurationThreshold(Long) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.ReadDataBuilder.Configuration
-  
- setWatermarkPolicy(String) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.ReadDataBuilder.Configuration
-  
- setWindmillGetDataStreamCount(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setWindmillHarnessUpdateReportingPeriod(Duration) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setWindmillMessagesBetweenIsReadyChecks(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setWindmillRequestBatchedGetWorkResponse(boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setWindmillServiceCommitThreads(Integer) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setWindmillServiceEndpoint(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setWindmillServicePort(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setWindmillServiceRpcChannelAliveTimeoutSec(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setWindmillServiceStreamingLogEveryNStreamFailures(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setWindmillServiceStreamingRpcBatchLimit(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setWindmillServiceStreamingRpcHealthCheckPeriodMs(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setWindmillServiceStreamMaxBackoffMillis(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions
-  
- setWindowedWrites() - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
- 
Indicates that the operation will be performing windowed writes. 
- setWindowingStrategyInternal(WindowingStrategy<?, ?>) - Method in class org.apache.beam.sdk.values.PCollection
- 
For internal use only; no backwards-compatibility guarantees. 
- setWithAttributes(Boolean) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.Configuration
-  
- setWithPartitions(Boolean) - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadConfiguration.Builder
-  
- setWorkerCacheMb(Integer) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
-  
- setWorkerCPUs(int) - Method in interface org.apache.beam.runners.twister2.Twister2PipelineOptions
-  
- setWorkerDiskType(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-  
- setWorkerHarnessContainerImage(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
- 
- setWorkerId(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
-  
- setWorkerLogLevelOverrides(DataflowWorkerLoggingOptions.WorkerLogLevelOverrides) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
- 
Deprecated. 
- setWorkerMachineType(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
-  
- setWorkerPool(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
-  
- setWorkerRegion(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-  
- setWorkerSystemErrMessageLevel(DataflowWorkerLoggingOptions.Level) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
- 
Deprecated. 
- setWorkerSystemOutMessageLevel(DataflowWorkerLoggingOptions.Level) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
- 
Deprecated. 
- setWorkerZone(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
-  
- setWriteDisposition(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration.Builder
-  
- setWriteDisposition(String) - Method in class org.apache.beam.sdk.io.snowflake.crosslanguage.WriteBuilder.Configuration
-  
- setWriteStatement(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration.Builder
-  
- setWriteTimeout(int) - Method in class org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer
-  
- setXmlConfiguration(FileWriteSchemaTransformConfiguration.XmlConfiguration) - Method in class org.apache.beam.sdk.io.fileschematransform.FileWriteSchemaTransformConfiguration.Builder
- 
Configures extra details related to writing XML formatted files. 
- setZetaSqlDefaultTimezone(String) - Method in interface org.apache.beam.sdk.extensions.sql.impl.BeamSqlPipelineOptions
-  
- setZone(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
- 
- sha1Bytes(byte[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinHashFunctions
- 
SHA1(X) 
- sha1String(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinHashFunctions
- 
SHA1(X) 
- sha256Bytes(byte[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinHashFunctions
- 
SHA256(X) 
- sha256String(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinHashFunctions
- 
SHA256(X) 
- sha512Bytes(byte[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinHashFunctions
- 
SHA512(X) 
- sha512String(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinHashFunctions
- 
SHA512(X) 
- ShardedKey<K> - Class in org.apache.beam.sdk.values
- 
A key and a shard number. 
- ShardedKeyCoder<KeyT> - Class in org.apache.beam.sdk.coders
- 
- ShardedKeyCoder(Coder<KeyT>) - Constructor for class org.apache.beam.sdk.coders.ShardedKeyCoder
-  
- ShardingFunction<UserT,DestinationT> - Interface in org.apache.beam.sdk.io
- 
- ShardNameTemplate - Class in org.apache.beam.sdk.io
- 
Standard shard naming templates. 
- ShardNameTemplate() - Constructor for class org.apache.beam.sdk.io.ShardNameTemplate
-  
- shardRefreshInterval(Duration) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.RecordAggregation.Builder
- 
Refresh interval for shards. 
- shortCircuitReturnNull(StackManipulation, StackManipulation) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
-  
- shorts() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
- 
- shouldConvertRaggedUnionTypesToVarying() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelDataTypeSystem
-  
- shouldDefer(TransformHierarchy.Node) - Method in class org.apache.beam.runners.spark.SparkRunner.Evaluator
-  
- shouldPublishLatencyMetrics() - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- shouldRepeat() - Method in exception org.apache.beam.io.requestresponse.UserCodeExecutionException
- 
Reports whether when thrown warrants repeat execution. 
- shouldRepeat() - Method in exception org.apache.beam.io.requestresponse.UserCodeQuotaException
- 
Reports that quota errors should be repeated. 
- shouldRepeat() - Method in exception org.apache.beam.io.requestresponse.UserCodeRemoteSystemException
- 
Reports that remote system errors should be repeated. 
- shouldRepeat() - Method in exception org.apache.beam.io.requestresponse.UserCodeTimeoutException
- 
Reports that timeouts should be repeated. 
- shouldResume() - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
- 
If false, the  DoFn promises that there is no more work remaining for the current
 element, so the runner should not resume the  DoFn.ProcessElement call. 
- shouldRetry(InsertRetryPolicy.Context) - Method in class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy
- 
Return true if this failure should be retried. 
- shutdown() - Method in class org.apache.beam.sdk.transformservice.launcher.TransformServiceLauncher
-  
- sideInput(PCollectionView<SideInputT>) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
- 
Returns the value of a given side input. 
- sideInput(PCollectionView<SideInputT>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
- 
Returns the value of a given side input. 
- sideInput(PCollectionView<T>) - Method in interface org.apache.beam.sdk.state.StateContext
- 
Returns the value of the side input for the corresponding state window. 
- sideInput(PCollectionView<T>) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.Context
- 
Returns the value of the side input for the window corresponding to the main input's window
 in which values are being combined. 
- sideInput(PCollectionView<T>) - Method in class org.apache.beam.sdk.transforms.Contextful.Fn.Context
- 
Accesses the given side input. 
- sideInput(PCollectionView<T>) - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContext
- 
Returns the value of the side input. 
- SideInputBroadcast<T> - Class in org.apache.beam.runners.spark.util
- 
Broadcast helper for side inputs. 
- sideInputId() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.SideInputSpec
-  
- sideInputJoin(PCollection<Row>, PCollection<Row>, FieldAccessDescriptor, FieldAccessDescriptor) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSideInputJoinRel
-  
- SideInputSpec() - Constructor for class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.SideInputSpec
-  
- SideInputValues<T> - Interface in org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions
- 
SideInputValues serves as a Kryo serializable container that contains a materialized view
 of side inputs. 
- SideInputValues.BaseSideInputValues<BinaryT,ValuesT,T> - Class in org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions
-  
- SideInputValues.ByWindow<T> - Class in org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions
- 
- SideInputValues.Global<T> - Class in org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions
- 
- SideInputValues.Loader<T> - Interface in org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions
- 
- signalStart() - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal
- 
Outputs a message that the pipeline has started. 
- signalSuccessWhen(Coder<T>, SerializableFunction<T, String>, SerializableFunction<Set<T>, Boolean>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal
- 
Outputs a success message when successPredicateis evaluated to true.
 
- signalSuccessWhen(Coder<T>, SerializableFunction<Set<T>, Boolean>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal
- 
- SimpleCombineFn(SerializableFunction<Iterable<V>, V>) - Constructor for class org.apache.beam.sdk.transforms.Combine.SimpleCombineFn
- 
Deprecated. 
- SimpleFunction<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
- 
- SimpleFunction() - Constructor for class org.apache.beam.sdk.transforms.SimpleFunction
-  
- SimpleFunction(SerializableFunction<InputT, OutputT>) - Constructor for class org.apache.beam.sdk.transforms.SimpleFunction
-  
- SimpleIdentifierContext(FieldSpecifierNotationParser.DotExpressionComponentContext) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.SimpleIdentifierContext
-  
- SimpleRateLimitPolicy(double) - Constructor for class org.apache.beam.sdk.io.googleads.GoogleAdsV17.SimpleRateLimitPolicy
-  
- SimpleRateLimitPolicy(double, long, TimeUnit) - Constructor for class org.apache.beam.sdk.io.googleads.GoogleAdsV17.SimpleRateLimitPolicy
-  
- SimpleRemoteEnvironment() - Constructor for class org.apache.beam.runners.fnexecution.environment.RemoteEnvironment.SimpleRemoteEnvironment
-  
- singleByteEncodeDoLoopByteString(VarIntBenchmark.Bytes, VarIntBenchmark.ByteStringOutput) - Method in class org.apache.beam.sdk.jmh.util.VarIntBenchmark
-  
- singleByteEncodeDoLoopTwiddleByteString(VarIntBenchmark.Bytes, VarIntBenchmark.ByteStringOutput) - Method in class org.apache.beam.sdk.jmh.util.VarIntBenchmark
-  
- singleByteEncodeLoopByteString(VarIntBenchmark.Bytes, VarIntBenchmark.ByteStringOutput) - Method in class org.apache.beam.sdk.jmh.util.VarIntBenchmark
-  
- singleByteEncodeUnrolledByteString(VarIntBenchmark.Bytes, VarIntBenchmark.ByteStringOutput) - Method in class org.apache.beam.sdk.jmh.util.VarIntBenchmark
-  
- SingleEnvironmentInstanceJobBundleFactory - Class in org.apache.beam.runners.fnexecution.control
- 
- singleOutputOverrideFactory() - Static method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides
- 
Returns a  PTransformOverrideFactory that replaces a single-output  ParDo with a
 composite transform specialized for the  DataflowRunner. 
- SingleStoreIO - Class in org.apache.beam.sdk.io.singlestore
- 
IO to read and write data on SingleStoreDB. 
- SingleStoreIO() - Constructor for class org.apache.beam.sdk.io.singlestore.SingleStoreIO
-  
- SingleStoreIO.DataSourceConfiguration - Class in org.apache.beam.sdk.io.singlestore
- 
A POJO describing a SingleStoreDB DataSourceby providing all properties needed to
 create it.
 
- SingleStoreIO.Read<T> - Class in org.apache.beam.sdk.io.singlestore
- 
- SingleStoreIO.Read.SingleStoreRowMapperInitializationException - Exception in org.apache.beam.sdk.io.singlestore
-  
- SingleStoreIO.ReadWithPartitions<T> - Class in org.apache.beam.sdk.io.singlestore
- 
- SingleStoreIO.RowMapper<T> - Interface in org.apache.beam.sdk.io.singlestore
- 
- SingleStoreIO.RowMapperWithCoder<T> - Interface in org.apache.beam.sdk.io.singlestore
- 
A RowMapper that provides a Coder for resulting PCollection. 
- SingleStoreIO.RowMapperWithInit<T> - Interface in org.apache.beam.sdk.io.singlestore
- 
A RowMapper that requires initialization. 
- SingleStoreIO.StatementPreparator - Interface in org.apache.beam.sdk.io.singlestore
- 
An interface used by the SingleStoreIO  SingleStoreIO.Read to set the parameters of the  PreparedStatement. 
- SingleStoreIO.UserDataMapper<T> - Interface in org.apache.beam.sdk.io.singlestore
- 
- SingleStoreIO.Write<T> - Class in org.apache.beam.sdk.io.singlestore
- 
- SingleStoreSchemaTransformReadConfiguration - Class in org.apache.beam.sdk.io.singlestore.schematransform
- 
Configuration for reading from SingleStoreDB. 
- SingleStoreSchemaTransformReadConfiguration() - Constructor for class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadConfiguration
-  
- SingleStoreSchemaTransformReadConfiguration.Builder - Class in org.apache.beam.sdk.io.singlestore.schematransform
-  
- SingleStoreSchemaTransformReadProvider - Class in org.apache.beam.sdk.io.singlestore.schematransform
- 
- SingleStoreSchemaTransformReadProvider() - Constructor for class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadProvider
-  
- SingleStoreSchemaTransformWriteConfiguration - Class in org.apache.beam.sdk.io.singlestore.schematransform
- 
Configuration for writing to SingleStoreDB. 
- SingleStoreSchemaTransformWriteConfiguration() - Constructor for class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteConfiguration
-  
- SingleStoreSchemaTransformWriteConfiguration.Builder - Class in org.apache.beam.sdk.io.singlestore.schematransform
-  
- SingleStoreSchemaTransformWriteProvider - Class in org.apache.beam.sdk.io.singlestore.schematransform
- 
- SingleStoreSchemaTransformWriteProvider() - Constructor for class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteProvider
-  
- singleTable(TableIdentifier, Schema) - Static method in interface org.apache.beam.sdk.io.iceberg.DynamicDestinations
-  
- singletonView(PCollection<T>, PCollectionViews.TypeDescriptorSupplier<T>, WindowingStrategy<?, W>, boolean, T, Coder<T>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
- 
Returns a  PCollectionView<T> capable of processing elements windowed using the provided
  WindowingStrategy. 
- singletonViewUsingVoidKey(TupleTag<Materializations.MultimapView<Void, T>>, PCollection<KV<Void, T>>, PCollectionViews.TypeDescriptorSupplier<T>, WindowingStrategy<?, W>, boolean, T, Coder<T>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
- 
- sinh(Double) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinTrigonometricFunctions
- 
SINH(X) 
- sink(Class<ElementT>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- sink(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- sink(String) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- Sink() - Constructor for class org.apache.beam.sdk.extensions.avro.io.AvroIO.Sink
-  
- sink - Variable in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
- 
The Sink that this WriteOperation will write to. 
- sink(Schema) - Static method in class org.apache.beam.sdk.io.parquet.ParquetIO
- 
- Sink() - Constructor for class org.apache.beam.sdk.io.parquet.ParquetIO.Sink
-  
- sink() - Static method in class org.apache.beam.sdk.io.TextIO
- 
- Sink() - Constructor for class org.apache.beam.sdk.io.TextIO.Sink
-  
- sink() - Static method in class org.apache.beam.sdk.io.TFRecordIO
- 
- Sink() - Constructor for class org.apache.beam.sdk.io.TFRecordIO.Sink
-  
- sink(TProtocolFactory) - Static method in class org.apache.beam.sdk.io.thrift.ThriftIO
- 
- Sink() - Constructor for class org.apache.beam.sdk.io.thrift.ThriftIO.Sink
-  
- sink(Class<T>) - Static method in class org.apache.beam.sdk.io.xml.XmlIO
- 
Outputs records as XML-formatted elements using JAXB. 
- Sink() - Constructor for class org.apache.beam.sdk.io.xml.XmlIO.Sink
-  
- SinkMetrics - Class in org.apache.beam.sdk.metrics
- 
Standard Sink Metrics. 
- SinkMetrics() - Constructor for class org.apache.beam.sdk.metrics.SinkMetrics
-  
- sinkViaGenericRecords(Schema, AvroIO.RecordFormatter<ElementT>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- size() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
- 
Returns the number of bytes in the backing array that are valid. 
- size() - Method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingSeekableByteChannel
-  
- size() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedList
-  
- size() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedRow
-  
- size() - Method in class org.apache.beam.sdk.fn.data.WeightedList
-  
- size() - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
-  
- size() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
-  
- size() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
- 
Returns the number of columns for this schema. 
- size() - Method in class org.apache.beam.sdk.values.PCollectionList
- 
- size() - Method in class org.apache.beam.sdk.values.TupleTagList
- 
Returns the number of TupleTags in this TupleTagList. 
- sizeBytes() - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata
-  
- SizeEstimator<T> - Interface in org.apache.beam.sdk.io.gcp.bigtable.changestreams.estimator
-  
- SizeEstimator<T> - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.estimator
- 
This class is used to estimate the size in bytes of a given element. 
- SizeEstimator(Coder<T>) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.estimator.SizeEstimator
-  
- sizeOf(T) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.estimator.CoderSizeEstimator
- 
Estimates the size in bytes of the given element with the configured  Coder . 
- sizeOf(T) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.estimator.NullSizeEstimator
-  
- sizeOf(T) - Method in interface org.apache.beam.sdk.io.gcp.bigtable.changestreams.estimator.SizeEstimator
-  
- sizeOf(T) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.estimator.SizeEstimator
- 
Estimates the size in bytes of the given element with the configured  Coder . 
- Sketch() - Constructor for class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.Sketch
-  
- SketchFrequencies - Class in org.apache.beam.sdk.extensions.sketching
- 
PTransforms to compute the estimate frequency of each element in a stream.
 
- SketchFrequencies() - Constructor for class org.apache.beam.sdk.extensions.sketching.SketchFrequencies
-  
- SketchFrequencies.CountMinSketchFn<InputT> - Class in org.apache.beam.sdk.extensions.sketching
- 
- SketchFrequencies.GlobalSketch<InputT> - Class in org.apache.beam.sdk.extensions.sketching
- 
- SketchFrequencies.PerKeySketch<K,V> - Class in org.apache.beam.sdk.extensions.sketching
- 
- SketchFrequencies.Sketch<T> - Class in org.apache.beam.sdk.extensions.sketching
- 
Wrap StreamLib's Count-Min Sketch to support counting all user types by hashing the encoded
 user type using the supplied deterministic coder. 
- skipCertificateVerification(boolean) - Method in class org.apache.beam.sdk.io.aws2.common.ClientConfiguration.Builder
- 
Optional flag to skip certificate verification. 
- skipCertificateVerification() - Method in class org.apache.beam.sdk.io.aws2.common.ClientConfiguration
- 
Optional flag to skip certificate verification. 
- SkipCertificateVerificationTrustManagerProvider() - Constructor for class org.apache.beam.sdk.io.aws2.common.ClientBuilderFactory.SkipCertificateVerificationTrustManagerProvider
-  
- skipIfEmpty() - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Don't write any output files if the PCollection is empty. 
- skipInvalidRows() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Insert all valid rows of a request, even if invalid rows exist. 
- Slf4jLogWriter - Class in org.apache.beam.runners.fnexecution.logging
- 
- SLIDING_WINDOW_TVF - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.TVFStreamingUtils
-  
- SlidingWindows - Class in org.apache.beam.sdk.transforms.windowing
- 
A  WindowFn that windows values into possibly overlapping fixed-size timestamp-based
 windows. 
- SMALL_INT - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- smallest(int) - Static method in class org.apache.beam.sdk.transforms.Top
- 
Returns a PTransformthat takes an inputPCollection<T>and returns aPCollection<List<T>>with a single element containing the smallestcountelements of
 the inputPCollection<T>, in increasing order, sorted according to their natural order.
 
- Smallest() - Constructor for class org.apache.beam.sdk.transforms.Top.Smallest
- 
Deprecated. 
- smallestDoublesFn(int) - Static method in class org.apache.beam.sdk.transforms.Top
- 
- smallestFn(int) - Static method in class org.apache.beam.sdk.transforms.Top
- 
- smallestIntsFn(int) - Static method in class org.apache.beam.sdk.transforms.Top
- 
- smallestLongsFn(int) - Static method in class org.apache.beam.sdk.transforms.Top
- 
- smallestPerKey(int) - Static method in class org.apache.beam.sdk.transforms.Top
- 
Returns a PTransformthat takes an inputPCollection<KV<K, V>>and returns aPCollection<KV<K, List<V>>>that contains an output element mapping each distinct key
 in the inputPCollectionto the smallestcountvalues associated with that key
 in the inputPCollection<KV<K, V>>, in increasing order, sorted according to their
 natural order.
 
- SnappyCoder<T> - Class in org.apache.beam.sdk.coders
- 
Wraps an existing coder with Snappy compression. 
- snapshot(SchemaVersion) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
-  
- SnapshotInfo - Class in org.apache.beam.sdk.io.iceberg
- 
This is an AutoValue representation of an Iceberg Snapshot.
 
- SnapshotInfo() - Constructor for class org.apache.beam.sdk.io.iceberg.SnapshotInfo
-  
- SnapshotInfo.Builder - Class in org.apache.beam.sdk.io.iceberg
-  
- SnowflakeArray - Class in org.apache.beam.sdk.io.snowflake.data.structured
-  
- SnowflakeArray() - Constructor for class org.apache.beam.sdk.io.snowflake.data.structured.SnowflakeArray
-  
- SnowflakeBatchServiceConfig - Class in org.apache.beam.sdk.io.snowflake.services
- 
Class for preparing configuration for batch write and read. 
- SnowflakeBatchServiceConfig(SerializableFunction<Void, DataSource>, String, String, String, String, String, String, String) - Constructor for class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceConfig
- 
Creating a batch configuration for reading. 
- SnowflakeBatchServiceConfig(SerializableFunction<Void, DataSource>, List<String>, SnowflakeTableSchema, String, String, String, String, CreateDisposition, WriteDisposition, String, String, String) - Constructor for class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceConfig
- 
Creating a batch configuration for writing. 
- SnowflakeBatchServiceImpl - Class in org.apache.beam.sdk.io.snowflake.services
- 
- SnowflakeBatchServiceImpl() - Constructor for class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceImpl
-  
- SnowflakeBinary - Class in org.apache.beam.sdk.io.snowflake.data.text
-  
- SnowflakeBinary() - Constructor for class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeBinary
-  
- SnowflakeBinary(long) - Constructor for class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeBinary
-  
- SnowflakeBoolean - Class in org.apache.beam.sdk.io.snowflake.data.logical
-  
- SnowflakeBoolean() - Constructor for class org.apache.beam.sdk.io.snowflake.data.logical.SnowflakeBoolean
-  
- SnowflakeChar - Class in org.apache.beam.sdk.io.snowflake.data.text
-  
- SnowflakeChar() - Constructor for class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeChar
-  
- SnowflakeColumn - Class in org.apache.beam.sdk.io.snowflake.data
- 
POJO describing single Column within Snowflake Table. 
- SnowflakeColumn() - Constructor for class org.apache.beam.sdk.io.snowflake.data.SnowflakeColumn
-  
- SnowflakeColumn(String, SnowflakeDataType) - Constructor for class org.apache.beam.sdk.io.snowflake.data.SnowflakeColumn
-  
- SnowflakeColumn(String, SnowflakeDataType, boolean) - Constructor for class org.apache.beam.sdk.io.snowflake.data.SnowflakeColumn
-  
- SnowflakeDataType - Interface in org.apache.beam.sdk.io.snowflake.data
- 
Interface for data types to provide SQLs for themselves. 
- SnowflakeDate - Class in org.apache.beam.sdk.io.snowflake.data.datetime
-  
- SnowflakeDate() - Constructor for class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeDate
-  
- SnowflakeDateTime - Class in org.apache.beam.sdk.io.snowflake.data.datetime
-  
- SnowflakeDateTime() - Constructor for class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeDateTime
-  
- SnowflakeDecimal - Class in org.apache.beam.sdk.io.snowflake.data.numeric
-  
- SnowflakeDecimal() - Constructor for class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeDecimal
-  
- SnowflakeDecimal(int, int) - Constructor for class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeDecimal
-  
- SnowflakeDouble - Class in org.apache.beam.sdk.io.snowflake.data.numeric
-  
- SnowflakeDouble() - Constructor for class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeDouble
-  
- SnowflakeFloat - Class in org.apache.beam.sdk.io.snowflake.data.numeric
-  
- SnowflakeFloat() - Constructor for class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeFloat
-  
- SnowflakeGeography - Class in org.apache.beam.sdk.io.snowflake.data.geospatial
-  
- SnowflakeGeography() - Constructor for class org.apache.beam.sdk.io.snowflake.data.geospatial.SnowflakeGeography
-  
- SnowflakeInteger - Class in org.apache.beam.sdk.io.snowflake.data.numeric
-  
- SnowflakeInteger() - Constructor for class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeInteger
-  
- SnowflakeIO - Class in org.apache.beam.sdk.io.snowflake
- 
IO to read and write data on Snowflake. 
- SnowflakeIO() - Constructor for class org.apache.beam.sdk.io.snowflake.SnowflakeIO
-  
- SnowflakeIO.Concatenate - Class in org.apache.beam.sdk.io.snowflake
- 
Combines list of Stringto provide oneStringwith paths where files were
 staged for write.
 
- SnowflakeIO.CsvMapper<T> - Interface in org.apache.beam.sdk.io.snowflake
- 
Interface for user-defined function mapping parts of CSV line into T. 
- SnowflakeIO.DataSourceConfiguration - Class in org.apache.beam.sdk.io.snowflake
- 
A POJO describing a DataSource, providing all properties allowing to create aDataSource.
 
- SnowflakeIO.DataSourceProviderFromDataSourceConfiguration - Class in org.apache.beam.sdk.io.snowflake
- 
- SnowflakeIO.Read<T> - Class in org.apache.beam.sdk.io.snowflake
- 
- SnowflakeIO.Read.CleanTmpFilesFromGcsFn - Class in org.apache.beam.sdk.io.snowflake
- 
Removes temporary staged files after reading. 
- SnowflakeIO.Read.MapCsvToStringArrayFn - Class in org.apache.beam.sdk.io.snowflake
- 
Parses  String from incoming data in  PCollection to have proper format for CSV
 files. 
- SnowflakeIO.UserDataMapper<T> - Interface in org.apache.beam.sdk.io.snowflake
- 
Interface for user-defined function mapping T into array of Objects. 
- SnowflakeIO.Write<T> - Class in org.apache.beam.sdk.io.snowflake
- 
- SnowflakeNumber - Class in org.apache.beam.sdk.io.snowflake.data.numeric
-  
- SnowflakeNumber() - Constructor for class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeNumber
-  
- SnowflakeNumber(int, int) - Constructor for class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeNumber
-  
- SnowflakeNumeric - Class in org.apache.beam.sdk.io.snowflake.data.numeric
-  
- SnowflakeNumeric() - Constructor for class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeNumeric
-  
- SnowflakeNumeric(int, int) - Constructor for class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeNumeric
-  
- SnowflakeObject - Class in org.apache.beam.sdk.io.snowflake.data.structured
-  
- SnowflakeObject() - Constructor for class org.apache.beam.sdk.io.snowflake.data.structured.SnowflakeObject
-  
- SnowflakePipelineOptions - Interface in org.apache.beam.sdk.io.snowflake
-  
- SnowflakeReal - Class in org.apache.beam.sdk.io.snowflake.data.numeric
-  
- SnowflakeReal() - Constructor for class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeReal
-  
- SnowflakeServices - Interface in org.apache.beam.sdk.io.snowflake.services
- 
Interface which defines common methods for interacting with Snowflake. 
- SnowflakeServices.BatchService - Interface in org.apache.beam.sdk.io.snowflake.services
-  
- SnowflakeServices.StreamingService - Interface in org.apache.beam.sdk.io.snowflake.services
-  
- SnowflakeServicesImpl - Class in org.apache.beam.sdk.io.snowflake.services
-  
- SnowflakeServicesImpl() - Constructor for class org.apache.beam.sdk.io.snowflake.services.SnowflakeServicesImpl
-  
- SnowflakeStreamingServiceConfig - Class in org.apache.beam.sdk.io.snowflake.services
- 
Class for preparing configuration for streaming write. 
- SnowflakeStreamingServiceConfig(List<String>, String, SimpleIngestManager) - Constructor for class org.apache.beam.sdk.io.snowflake.services.SnowflakeStreamingServiceConfig
- 
Constructor to create configuration for streaming write. 
- SnowflakeStreamingServiceImpl - Class in org.apache.beam.sdk.io.snowflake.services
- 
- SnowflakeStreamingServiceImpl() - Constructor for class org.apache.beam.sdk.io.snowflake.services.SnowflakeStreamingServiceImpl
-  
- SnowflakeString - Class in org.apache.beam.sdk.io.snowflake.data.text
-  
- SnowflakeString() - Constructor for class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeString
-  
- SnowflakeString(long) - Constructor for class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeString
-  
- SnowflakeTableSchema - Class in org.apache.beam.sdk.io.snowflake.data
- 
POJO representing schema of Table in Snowflake. 
- SnowflakeTableSchema() - Constructor for class org.apache.beam.sdk.io.snowflake.data.SnowflakeTableSchema
-  
- SnowflakeTableSchema(SnowflakeColumn...) - Constructor for class org.apache.beam.sdk.io.snowflake.data.SnowflakeTableSchema
-  
- SnowflakeText - Class in org.apache.beam.sdk.io.snowflake.data.text
-  
- SnowflakeText() - Constructor for class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeText
-  
- SnowflakeText(long) - Constructor for class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeText
-  
- SnowflakeTime - Class in org.apache.beam.sdk.io.snowflake.data.datetime
-  
- SnowflakeTime() - Constructor for class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeTime
-  
- SnowflakeTimestamp - Class in org.apache.beam.sdk.io.snowflake.data.datetime
-  
- SnowflakeTimestamp() - Constructor for class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeTimestamp
-  
- SnowflakeTimestampLTZ - Class in org.apache.beam.sdk.io.snowflake.data.datetime
-  
- SnowflakeTimestampLTZ() - Constructor for class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeTimestampLTZ
-  
- SnowflakeTimestampNTZ - Class in org.apache.beam.sdk.io.snowflake.data.datetime
-  
- SnowflakeTimestampNTZ() - Constructor for class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeTimestampNTZ
-  
- SnowflakeTimestampTZ - Class in org.apache.beam.sdk.io.snowflake.data.datetime
-  
- SnowflakeTimestampTZ() - Constructor for class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeTimestampTZ
-  
- SnowflakeTransformRegistrar - Class in org.apache.beam.sdk.io.snowflake.crosslanguage
- 
Exposes SnowflakeIO.ReadandSnowflakeIO.Writeas an external transform for
 cross-language usage.
 
- SnowflakeTransformRegistrar() - Constructor for class org.apache.beam.sdk.io.snowflake.crosslanguage.SnowflakeTransformRegistrar
-  
- SnowflakeVarBinary - Class in org.apache.beam.sdk.io.snowflake.data.text
-  
- SnowflakeVarBinary() - Constructor for class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeVarBinary
-  
- SnowflakeVarchar - Class in org.apache.beam.sdk.io.snowflake.data.text
-  
- SnowflakeVarchar() - Constructor for class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeVarchar
-  
- SnowflakeVarchar(long) - Constructor for class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeVarchar
-  
- SnowflakeVariant - Class in org.apache.beam.sdk.io.snowflake.data.structured
-  
- SnowflakeVariant() - Constructor for class org.apache.beam.sdk.io.snowflake.data.structured.SnowflakeVariant
-  
- SnsIO - Class in org.apache.beam.sdk.io.aws2.sns
- 
IO to send notifications via  SNS. 
- SnsIO() - Constructor for class org.apache.beam.sdk.io.aws2.sns.SnsIO
-  
- SnsIO.Write<T> - Class in org.apache.beam.sdk.io.aws2.sns
- 
- SocketAddressFactory - Class in org.apache.beam.sdk.fn.channel
- 
Creates a SocketAddressbased upon a supplied string.
 
- SocketAddressFactory() - Constructor for class org.apache.beam.sdk.fn.channel.SocketAddressFactory
-  
- socketTimeout(Integer) - Method in class org.apache.beam.sdk.io.aws2.common.HttpClientConfiguration.Builder
- 
Milliseconds to wait for data to be transferred over an established, open connection before
 the connection is timed out. 
- socketTimeout() - Method in class org.apache.beam.sdk.io.aws2.common.HttpClientConfiguration
- 
Milliseconds to wait for data to be transferred over an established, open connection before the
 connection is timed out. 
- Solace - Class in org.apache.beam.sdk.io.solace.data
- 
Provides core data models and utilities for working with Solace messages in the context of Apache
 Beam pipelines. 
- Solace() - Constructor for class org.apache.beam.sdk.io.solace.data.Solace
-  
- Solace.CorrelationKey - Class in org.apache.beam.sdk.io.solace.data
- 
The correlation key is an object that is passed back to the client during the event broker ack
 or nack. 
- Solace.CorrelationKey.Builder - Class in org.apache.beam.sdk.io.solace.data
-  
- Solace.Destination - Class in org.apache.beam.sdk.io.solace.data
- 
Represents a Solace message destination (either a Topic or a Queue). 
- Solace.Destination.Builder - Class in org.apache.beam.sdk.io.solace.data
-  
- Solace.DestinationType - Enum in org.apache.beam.sdk.io.solace.data
- 
Represents a Solace destination type. 
- Solace.PublishResult - Class in org.apache.beam.sdk.io.solace.data
- 
The result of writing a message to Solace. 
- Solace.PublishResult.Builder - Class in org.apache.beam.sdk.io.solace.data
-  
- Solace.Queue - Class in org.apache.beam.sdk.io.solace.data
- 
Represents a Solace queue. 
- Solace.Record - Class in org.apache.beam.sdk.io.solace.data
- 
Represents a Solace message record with its associated metadata. 
- Solace.Record.Builder - Class in org.apache.beam.sdk.io.solace.data
-  
- Solace.SolaceRecordMapper - Class in org.apache.beam.sdk.io.solace.data
- 
- Solace.Topic - Class in org.apache.beam.sdk.io.solace.data
- 
Represents a Solace topic. 
- SolaceCheckpointMark - Class in org.apache.beam.sdk.io.solace.read
- 
Checkpoint for an unbounded Solace source. 
- SolaceIO - Class in org.apache.beam.sdk.io.solace
- 
- SolaceIO() - Constructor for class org.apache.beam.sdk.io.solace.SolaceIO
-  
- SolaceIO.Read<T> - Class in org.apache.beam.sdk.io.solace
-  
- SolaceIO.SubmissionMode - Enum in org.apache.beam.sdk.io.solace
-  
- SolaceIO.Write<T> - Class in org.apache.beam.sdk.io.solace
-  
- SolaceIO.WriterType - Enum in org.apache.beam.sdk.io.solace
-  
- SolaceMessageProducer - Class in org.apache.beam.sdk.io.solace.broker
-  
- SolaceMessageProducer(XMLMessageProducer) - Constructor for class org.apache.beam.sdk.io.solace.broker.SolaceMessageProducer
-  
- SolaceMessageReceiver - Class in org.apache.beam.sdk.io.solace.broker
-  
- SolaceMessageReceiver(FlowReceiver) - Constructor for class org.apache.beam.sdk.io.solace.broker.SolaceMessageReceiver
-  
- SolaceOutput - Class in org.apache.beam.sdk.io.solace.write
- 
- SolaceRecordMapper() - Constructor for class org.apache.beam.sdk.io.solace.data.Solace.SolaceRecordMapper
-  
- solaceSessionServiceWithProducer() - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- SolrIO - Class in org.apache.beam.sdk.io.solr
- 
Transforms for reading and writing data from/to Solr. 
- SolrIO.ConnectionConfiguration - Class in org.apache.beam.sdk.io.solr
- 
A POJO describing a connection configuration to Solr. 
- SolrIO.Read - Class in org.apache.beam.sdk.io.solr
- 
- SolrIO.ReadAll - Class in org.apache.beam.sdk.io.solr
-  
- SolrIO.ReplicaInfo - Class in org.apache.beam.sdk.io.solr
- 
A POJO describing a replica of Solr. 
- SolrIO.RetryConfiguration - Class in org.apache.beam.sdk.io.solr
- 
A POJO encapsulating a configuration for retry behavior when issuing requests to Solr. 
- SolrIO.Write - Class in org.apache.beam.sdk.io.solr
- 
- sort() - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter
-  
- SORT_VALUES - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- sortBySchema(List<FieldValueTypeInformation>, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.StaticSchemaInference
-  
- sorted() - Method in class org.apache.beam.sdk.schemas.Schema
- 
Returns an identical Schema with lexicographically sorted fields. 
- sorted() - Method in class org.apache.beam.sdk.values.Row
- 
Returns an equivalent  Row with fields lexicographically sorted by their name. 
- SortedMapCoder<K extends java.lang.Comparable<? super K>,V> - Class in org.apache.beam.sdk.coders
- 
A  Coder for  Maps that encodes them according to provided coders for
 keys and values. 
- SortValues<PrimaryKeyT,SecondaryKeyT,ValueT> - Class in org.apache.beam.sdk.extensions.sorter
- 
SortValues<PrimaryKeyT, SecondaryKeyT, ValueT>takes aPCollection<KV<PrimaryKeyT,
 Iterable<KV<SecondaryKeyT, ValueT>>>>with elements consisting of a primary key and iterables
 over<secondary key, value>pairs, and returns aPCollection<KV<PrimaryKeyT,
 Iterable<KV<SecondaryKeyT, ValueT>>>of the same elements but with values sorted by a secondary
 key.
 
- SOURCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner.NonCumulativeCostImpl
-  
- SOURCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.planner.RelMdNodeStats
-  
- Source<T> - Class in org.apache.beam.sdk.io
- 
Base class for defining input formats and creating a Sourcefor reading the input.
 
- Source() - Constructor for class org.apache.beam.sdk.io.Source
-  
- source() - Method in class org.apache.beam.sdk.io.splunk.SplunkEvent
-  
- Source.Reader<T> - Class in org.apache.beam.sdk.io
- 
The interface that readers of custom input sources must implement. 
- SOURCE_DOES_NOT_NEED_SPLITTING - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- SOURCE_ESTIMATED_SIZE_BYTES - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- SOURCE_IS_INFINITE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- SOURCE_METADATA - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- SOURCE_SPEC - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- SOURCE_STEP_INPUT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- SourceMetrics - Class in org.apache.beam.sdk.metrics
- 
- SourceMetrics() - Constructor for class org.apache.beam.sdk.metrics.SourceMetrics
-  
- sourceName() - Method in class org.apache.beam.runners.spark.metrics.SparkBeamMetricSource
-  
- sourceName() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.SparkBeamMetricSource
-  
- SourceRDD - Class in org.apache.beam.runners.spark.io
- 
Classes implementing Beam  SourceRDDs. 
- SourceRDD() - Constructor for class org.apache.beam.runners.spark.io.SourceRDD
-  
- SourceRDD.Bounded<T> - Class in org.apache.beam.runners.spark.io
- 
- SourceRDD.Unbounded<T,CheckpointMarkT extends UnboundedSource.CheckpointMark> - Class in org.apache.beam.runners.spark.io
- 
- SourceRecordJson - Class in org.apache.beam.io.debezium
- 
This class can be used as a mapper for each SourceRecordretrieved.
 
- SourceRecordJson(SourceRecord) - Constructor for class org.apache.beam.io.debezium.SourceRecordJson
- 
Initializer. 
- SourceRecordJson.SourceRecordJsonMapper - Class in org.apache.beam.io.debezium
- 
- SourceRecordJsonMapper() - Constructor for class org.apache.beam.io.debezium.SourceRecordJson.SourceRecordJsonMapper
-  
- SourceRecordMapper<T> - Interface in org.apache.beam.io.debezium
- 
Interface used to map a Kafka source record. 
- SourceTestUtils - Class in org.apache.beam.sdk.testing
- 
Helper functions and test harnesses for checking correctness of  Source implementations. 
- SourceTestUtils() - Constructor for class org.apache.beam.sdk.testing.SourceTestUtils
-  
- SourceTestUtils.ExpectedSplitOutcome - Enum in org.apache.beam.sdk.testing
- 
- sourceType() - Method in class org.apache.beam.sdk.io.splunk.SplunkEvent
-  
- span(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
- 
Returns the minimal window that includes both this window and the given window. 
- SpannerAccessor - Class in org.apache.beam.sdk.io.gcp.spanner
- 
Manages lifecycle of DatabaseClientandSpannerinstances.
 
- SpannerChangestreamsReadConfiguration() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration
-  
- SpannerChangestreamsReadSchemaTransformProvider - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams
-  
- SpannerChangestreamsReadSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.SpannerChangestreamsReadSchemaTransformProvider
-  
- SpannerChangestreamsReadSchemaTransformProvider.DataChangeRecordToRow - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams
-  
- SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams
-  
- SpannerChangestreamsReadSchemaTransformProvider.SpannerChangestreamsReadConfiguration.Builder - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams
-  
- SpannerConfig - Class in org.apache.beam.sdk.io.gcp.spanner
- 
Configuration for a Cloud Spanner client. 
- SpannerConfig() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- SpannerConfig.Builder - Class in org.apache.beam.sdk.io.gcp.spanner
- 
- SpannerIO - Class in org.apache.beam.sdk.io.gcp.spanner
- 
Reading from Cloud Spanner 
- SpannerIO.CreateTransaction - Class in org.apache.beam.sdk.io.gcp.spanner
- 
- SpannerIO.CreateTransaction.Builder - Class in org.apache.beam.sdk.io.gcp.spanner
- 
- SpannerIO.FailureMode - Enum in org.apache.beam.sdk.io.gcp.spanner
- 
A failure handling strategy. 
- SpannerIO.Read - Class in org.apache.beam.sdk.io.gcp.spanner
- 
- SpannerIO.ReadAll - Class in org.apache.beam.sdk.io.gcp.spanner
- 
- SpannerIO.ReadChangeStream - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerIO.SpannerChangeStreamOptions - Interface in org.apache.beam.sdk.io.gcp.spanner
- 
Interface to display the name of the metadata table on Dataflow UI. 
- SpannerIO.Write - Class in org.apache.beam.sdk.io.gcp.spanner
- 
A  PTransform that writes  Mutation objects to Google Cloud Spanner. 
- SpannerIO.WriteGrouped - Class in org.apache.beam.sdk.io.gcp.spanner
- 
- SpannerReadSchemaTransformConfiguration() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration
-  
- SpannerReadSchemaTransformProvider - Class in org.apache.beam.sdk.io.gcp.spanner
- 
A provider for reading from Cloud Spanner using a Schema Transform Provider. 
- SpannerReadSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider
-  
- SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration.Builder - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerSchema - Class in org.apache.beam.sdk.io.gcp.spanner
- 
Encapsulates Cloud Spanner Schema. 
- SpannerSchema() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerSchema
-  
- SpannerSchema.Column - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerSchema.KeyPart - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerSchemaRetrievalException - Exception in org.apache.beam.sdk.io.gcp.spanner
- 
Exception to signal that Spanner schema retrieval failed. 
- SpannerTransformRegistrar - Class in org.apache.beam.sdk.io.gcp.spanner
- 
Exposes SpannerIO.WriteRowsandSpannerIO.ReadRowsas an external transform for
 cross-language usage.
 
- SpannerTransformRegistrar() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerTransformRegistrar
-  
- SpannerTransformRegistrar.CrossLanguageConfiguration - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerTransformRegistrar.DeleteBuilder - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerTransformRegistrar.InsertBuilder - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerTransformRegistrar.InsertOrUpdateBuilder - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerTransformRegistrar.ReadBuilder - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerTransformRegistrar.ReadBuilder.Configuration - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerTransformRegistrar.ReplaceBuilder - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerTransformRegistrar.UpdateBuilder - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerWriteResult - Class in org.apache.beam.sdk.io.gcp.spanner
- 
- SpannerWriteResult(Pipeline, PCollection<Void>, PCollection<MutationGroup>, TupleTag<MutationGroup>) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteResult
-  
- SpannerWriteSchemaTransformConfiguration() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration
-  
- SpannerWriteSchemaTransformProvider - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerWriteSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider
-  
- SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration.Builder - Class in org.apache.beam.sdk.io.gcp.spanner
-  
- SparkBeamMetricSource - Class in org.apache.beam.runners.spark.metrics
- 
A Spark  Source that is tailored to expose a  SparkBeamMetric, wrapping an
 underlying  MetricResults instance. 
- SparkBeamMetricSource(String) - Constructor for class org.apache.beam.runners.spark.metrics.SparkBeamMetricSource
-  
- SparkBeamMetricSource - Class in org.apache.beam.runners.spark.structuredstreaming.metrics
- 
A Spark  Source that is tailored to expose a  SparkBeamMetric, wrapping an
 underlying  MetricResults instance. 
- SparkBeamMetricSource(String, MetricsAccumulator) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.metrics.SparkBeamMetricSource
-  
- SparkCommonPipelineOptions - Interface in org.apache.beam.runners.spark
- 
Spark runner  PipelineOptions handles Spark execution-related configurations, such as the
 master address, and other user-related knobs. 
- SparkCommonPipelineOptions.StorageLevelFactory - Class in org.apache.beam.runners.spark
- 
Returns Spark's default storage level for the Dataset or RDD API based on the respective
 runner. 
- SparkCommonPipelineOptions.TmpCheckpointDirFactory - Class in org.apache.beam.runners.spark
- 
Returns the default checkpoint directory of /tmp/${job.name}. 
- SparkContextOptions - Interface in org.apache.beam.runners.spark
- 
A custom  PipelineOptions to work with properties related to  JavaSparkContext. 
- SparkContextOptions.EmptyListenersList - Class in org.apache.beam.runners.spark
- 
Returns an empty list, to avoid handling null. 
- SparkGroupAlsoByWindowViaWindowSet - Class in org.apache.beam.runners.spark.stateful
- 
An implementation of GroupByKeyViaGroupByKeyOnly.GroupAlsoByWindowlogic for grouping by windows and controlling
 trigger firings and pane accumulation.
 
- SparkGroupAlsoByWindowViaWindowSet() - Constructor for class org.apache.beam.runners.spark.stateful.SparkGroupAlsoByWindowViaWindowSet
-  
- SparkJobInvoker - Class in org.apache.beam.runners.spark
- 
Creates a job invocation to manage the Spark runner's execution of a portable pipeline. 
- SparkJobServerDriver - Class in org.apache.beam.runners.spark
- 
Driver program that starts a job server for the Spark runner. 
- SparkJobServerDriver.SparkServerConfiguration - Class in org.apache.beam.runners.spark
- 
Spark runner-specific Configuration for the jobServer. 
- SparkKryoRegistrator() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.SparkSessionFactory.SparkKryoRegistrator
-  
- SparkNativePipelineVisitor - Class in org.apache.beam.runners.spark
- 
Pipeline visitor for translating a Beam pipeline into equivalent Spark operations. 
- SparkPipelineOptions - Interface in org.apache.beam.runners.spark
- 
Spark runner  PipelineOptions handles Spark execution-related configurations, such as the
 master address, batch-interval, and other user-related knobs. 
- SparkPipelineResult - Class in org.apache.beam.runners.spark
- 
Represents a Spark pipeline execution result. 
- SparkPipelineRunner - Class in org.apache.beam.runners.spark
- 
Runs a portable pipeline on Apache Spark. 
- SparkPipelineRunner(SparkPipelineOptions) - Constructor for class org.apache.beam.runners.spark.SparkPipelineRunner
-  
- SparkPortableStreamingPipelineOptions - Interface in org.apache.beam.runners.spark
- 
Pipeline options specific to the Spark portable runner running a streaming job. 
- SparkReceiverIO - Class in org.apache.beam.sdk.io.sparkreceiver
- 
Streaming sources for Spark Receiver.
 
- SparkReceiverIO() - Constructor for class org.apache.beam.sdk.io.sparkreceiver.SparkReceiverIO
-  
- SparkReceiverIO.Read<V> - Class in org.apache.beam.sdk.io.sparkreceiver
- 
- SparkRunner - Class in org.apache.beam.runners.spark
- 
The SparkRunner translate operations defined on a pipeline to a representation executable by
 Spark, and then submitting the job to Spark to be executed. 
- SparkRunner.Evaluator - Class in org.apache.beam.runners.spark
- 
Evaluator on the pipeline. 
- SparkRunnerDebugger - Class in org.apache.beam.runners.spark
- 
Pipeline runner which translates a Beam pipeline into equivalent Spark operations, without
 running them. 
- SparkRunnerDebugger.DebugSparkPipelineResult - Class in org.apache.beam.runners.spark
- 
- SparkRunnerKryoRegistrator - Class in org.apache.beam.runners.spark.coders
- 
Custom KryoRegistrators for Beam's Spark runner needs and registering used class in spark
 translation for better serialization performance.
 
- SparkRunnerKryoRegistrator() - Constructor for class org.apache.beam.runners.spark.coders.SparkRunnerKryoRegistrator
-  
- SparkRunnerRegistrar - Class in org.apache.beam.runners.spark
- 
- SparkRunnerRegistrar.Options - Class in org.apache.beam.runners.spark
- 
- SparkRunnerRegistrar.Runner - Class in org.apache.beam.runners.spark
- 
- SparkServerConfiguration() - Constructor for class org.apache.beam.runners.spark.SparkJobServerDriver.SparkServerConfiguration
-  
- SparkSessionFactory - Class in org.apache.beam.runners.spark.structuredstreaming.translation
-  
- SparkSessionFactory() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.SparkSessionFactory
-  
- SparkSessionFactory.SparkKryoRegistrator - Class in org.apache.beam.runners.spark.structuredstreaming.translation
- 
KryoRegistratorfor Spark to serialize broadcast variables used for side-inputs.
 
- SparkSideInputReader - Class in org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions
- 
- SparkSideInputReader - Class in org.apache.beam.runners.spark.util
- 
A SideInputReaderfor the SparkRunner.
 
- SparkSideInputReader(Map<TupleTag<?>, KV<WindowingStrategy<?, ?>, SideInputBroadcast<?>>>) - Constructor for class org.apache.beam.runners.spark.util.SparkSideInputReader
-  
- SparkStateInternals<K> - Class in org.apache.beam.runners.spark.stateful
- 
An implementation of StateInternalsfor the SparkRunner.
 
- SparkStructuredStreamingPipelineOptions - Interface in org.apache.beam.runners.spark.structuredstreaming
- 
Spark runner  PipelineOptions handles Spark execution-related configurations, such as the
 master address, and other user-related knobs. 
- SparkStructuredStreamingPipelineResult - Class in org.apache.beam.runners.spark.structuredstreaming
-  
- SparkStructuredStreamingRunner - Class in org.apache.beam.runners.spark.structuredstreaming
- 
- SparkStructuredStreamingRunnerRegistrar - Class in org.apache.beam.runners.spark.structuredstreaming
- 
- SparkStructuredStreamingRunnerRegistrar.Options - Class in org.apache.beam.runners.spark.structuredstreaming
- 
- SparkStructuredStreamingRunnerRegistrar.Runner - Class in org.apache.beam.runners.spark.structuredstreaming
- 
- SparkTimerInternals - Class in org.apache.beam.runners.spark.stateful
- 
An implementation of TimerInternalsfor the SparkRunner.
 
- SparkTransformOverrides - Class in org.apache.beam.runners.spark
- 
- SparkTransformOverrides() - Constructor for class org.apache.beam.runners.spark.SparkTransformOverrides
-  
- SparkUnboundedSource - Class in org.apache.beam.runners.spark.io
- 
- SparkUnboundedSource() - Constructor for class org.apache.beam.runners.spark.io.SparkUnboundedSource
-  
- SparkUnboundedSource.Metadata - Class in org.apache.beam.runners.spark.io
- 
A metadata holder for an input stream partition. 
- SparkWatermarks(Instant, Instant, Instant) - Constructor for class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.SparkWatermarks
-  
- specific(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns an  AvroCoder instance for the provided element type respecting Avro's Specific*
 suite for encoding and decoding. 
- specific(Class<T>) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns an  AvroCoder instance for the provided element type respecting Avro's Specific*
 suite for encoding and decoding. 
- specific(Class<T>, Schema) - Static method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
- 
Returns an  AvroCoder instance for the provided element type respecting Avro's Specific*
 suite for encoding and decoding. 
- specific(Class<T>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory
- 
Returns an  AvroDatumFactory instance for the provided element type respecting Avro's
 Specific* suite for encoding and decoding. 
- SpecificDatumFactory(Class<T>) - Constructor for class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory.SpecificDatumFactory
-  
- split(BeamFnApi.ProcessBundleSplitResponse) - Method in interface org.apache.beam.runners.fnexecution.control.BundleSplitHandler
-  
- split(double) - Method in interface org.apache.beam.runners.fnexecution.control.RemoteBundle
- 
Ask the remote bundle to split its current processing based upon its knowledge of remaining
 work. 
- split(double) - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor.ActiveBundle
-  
- split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.azure.cosmos.CosmosIO.BoundedCosmosBDSource
-  
- split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.BoundedSource
- 
Splits the source into bundles of approximately desiredBundleSizeBytes.
 
- split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
-  
- split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
-  
- split(int, PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.UnboundedSourceImpl
-  
- split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
-  
- split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
-  
- split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-  
- split(int) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
- 
Returns a list of up to  numSplits + 1ByteKeys in ascending order,
 where the keys have been interpolated to form roughly equal sub-ranges of this  ByteKeyRange, assuming a uniform distribution of keys within this range. 
- split(long, long) - Method in class org.apache.beam.sdk.io.range.OffsetRange
-  
- split(int, PipelineOptions) - Method in class org.apache.beam.sdk.io.solace.read.UnboundedSolaceSource
-  
- split(int, PipelineOptions) - Method in class org.apache.beam.sdk.io.UnboundedSource
- 
Returns a list of UnboundedSourceobjects representing the instances of this source
 that should be used when executing the workflow.
 
- split(String) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
Returns a  Regex.SplitPTransform that splits a string on the regular expression
 and then outputs each item. 
- split(Pattern) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
Returns a  Regex.SplitPTransform that splits a string on the regular expression
 and then outputs each item. 
- split(String, boolean) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
Returns a  Regex.SplitPTransform that splits a string on the regular expression
 and then outputs each item. 
- split(Pattern, boolean) - Static method in class org.apache.beam.sdk.transforms.Regex
- 
Returns a  Regex.SplitPTransform that splits a string on the regular expression
 and then outputs each item. 
- Split(Pattern, boolean) - Constructor for class org.apache.beam.sdk.transforms.Regex.Split
-  
- SPLIT_POINTS_UNKNOWN - Static variable in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
- 
- splitAtFraction(double) - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
- 
Tells the reader to narrow the range of the input it's going to read and give up the
 remainder, so that the new range would contain approximately the given fraction of the amount
 of data in the current range. 
- splitAtFraction(double) - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-  
- SplitIntoRangesFn(long) - Constructor for class org.apache.beam.sdk.io.ReadAllViaFileBasedSourceTransform.SplitIntoRangesFn
-  
- splitReadStream(SplitReadStreamRequest) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StorageClient
-  
- splitReadStream(SplitReadStreamRequest, String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StorageClient
-  
- SplitResult<RestrictionT> - Class in org.apache.beam.sdk.transforms.splittabledofn
- 
A representation of a split result. 
- SplitResult() - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.SplitResult
-  
- SplunkEvent - Class in org.apache.beam.sdk.io.splunk
- 
A  SplunkEvent describes a single payload sent to Splunk's Http Event Collector (HEC)
 endpoint. 
- SplunkEvent() - Constructor for class org.apache.beam.sdk.io.splunk.SplunkEvent
-  
- SplunkEvent.Builder - Class in org.apache.beam.sdk.io.splunk
- 
- SplunkEventCoder - Class in org.apache.beam.sdk.io.splunk
- 
- SplunkEventCoder() - Constructor for class org.apache.beam.sdk.io.splunk.SplunkEventCoder
-  
- SplunkIO - Class in org.apache.beam.sdk.io.splunk
- 
An unbounded sink for Splunk's Http Event Collector (HEC). 
- SplunkIO.Write - Class in org.apache.beam.sdk.io.splunk
- 
- SplunkWriteError - Class in org.apache.beam.sdk.io.splunk
- 
A class for capturing errors that occur while writing  SplunkEvent to Splunk's Http Event
 Collector (HEC) end point. 
- SplunkWriteError() - Constructor for class org.apache.beam.sdk.io.splunk.SplunkWriteError
-  
- SplunkWriteError.Builder - Class in org.apache.beam.sdk.io.splunk
- 
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeDate
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeTime
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeTimestampLTZ
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeTimestampNTZ
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.datetime.SnowflakeTimestampTZ
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.geospatial.SnowflakeGeography
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.logical.SnowflakeBoolean
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeFloat
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.numeric.SnowflakeNumber
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.SnowflakeColumn
-  
- sql() - Method in interface org.apache.beam.sdk.io.snowflake.data.SnowflakeDataType
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.SnowflakeTableSchema
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.structured.SnowflakeArray
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.structured.SnowflakeObject
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.structured.SnowflakeVariant
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeBinary
-  
- sql() - Method in class org.apache.beam.sdk.io.snowflake.data.text.SnowflakeVarchar
-  
- SqlAnalyzer - Class in org.apache.beam.sdk.extensions.sql.zetasql
- 
Adapter for Analyzerto simplify the API for parsing the query and resolving the AST.
 
- SqlCheckConstraint - Class in org.apache.beam.sdk.extensions.sql.impl.parser
- 
Parse tree for UNIQUE,PRIMARY KEYconstraints.
 
- SqlColumnDeclaration - Class in org.apache.beam.sdk.extensions.sql.impl.parser
- 
Parse tree for column. 
- SqlConversionException - Exception in org.apache.beam.sdk.extensions.sql.impl
- 
Exception thrown when BeamSQL cannot convert sql to BeamRelNode. 
- SqlConversionException(Throwable) - Constructor for exception org.apache.beam.sdk.extensions.sql.impl.SqlConversionException
-  
- SqlConversionException(String) - Constructor for exception org.apache.beam.sdk.extensions.sql.impl.SqlConversionException
-  
- SqlConversionException(String, Throwable) - Constructor for exception org.apache.beam.sdk.extensions.sql.impl.SqlConversionException
-  
- SqlCreateExternalTable - Class in org.apache.beam.sdk.extensions.sql.impl.parser
- 
Parse tree for CREATE EXTERNAL TABLEstatement.
 
- SqlCreateExternalTable(SqlParserPos, boolean, boolean, SqlIdentifier, List<Schema.Field>, SqlNode, SqlNode, SqlNode, SqlNode) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.parser.SqlCreateExternalTable
- 
Creates a SqlCreateExternalTable. 
- SqlCreateFunction - Class in org.apache.beam.sdk.extensions.sql.impl.parser
- 
Parse tree for CREATE FUNCTIONstatement.
 
- SqlCreateFunction(SqlParserPos, boolean, SqlIdentifier, SqlNode, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.parser.SqlCreateFunction
- 
Creates a SqlCreateFunction. 
- SqlDdlNodes - Class in org.apache.beam.sdk.extensions.sql.impl.parser
- 
Utilities concerning SqlNodefor DDL.
 
- SqlDropTable - Class in org.apache.beam.sdk.extensions.sql.impl.parser
- 
Parse tree for DROP TABLEstatement.
 
- SqlOperators - Class in org.apache.beam.sdk.extensions.sql.zetasql.translation
- 
A separate SqlOperators table for those functions that do not exist or not compatible with
 Calcite. 
- SqlOperators() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- sqlScalarFunctions() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.UserFunctionDefinitions
-  
- SqlSetOptionBeam - Class in org.apache.beam.sdk.extensions.sql.impl.parser
- 
SQL parse tree node to represent SETandRESETstatements.
 
- SqlSetOptionBeam(SqlParserPos, String, SqlIdentifier, SqlNode) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.parser.SqlSetOptionBeam
-  
- sqlTableValuedFunctions() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.UserFunctionDefinitions
- 
SQL native user-defined table-valued function can be resolved by Analyzer. 
- SqlTransform - Class in org.apache.beam.sdk.extensions.sql
- 
- SqlTransform() - Constructor for class org.apache.beam.sdk.extensions.sql.SqlTransform
-  
- SqlTransformSchemaTransformProvider - Class in org.apache.beam.sdk.extensions.sql.expansion
-  
- SqlTransformSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.expansion.SqlTransformSchemaTransformProvider
-  
- SqlTypes - Class in org.apache.beam.sdk.schemas.logicaltypes
- 
- sqlTypeWithAutoCast(RelDataTypeFactory, Type) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
- 
SQL-Java type mapping, with specified Beam rules: 
 1.
 
- SqsIO - Class in org.apache.beam.sdk.io.aws2.sqs
- 
IO to read (unbounded) from and write to  SQS queues. 
- SqsIO.Read - Class in org.apache.beam.sdk.io.aws2.sqs
- 
- SqsIO.Write - Class in org.apache.beam.sdk.io.aws2.sqs
- 
- SqsIO.WriteBatches<T> - Class in org.apache.beam.sdk.io.aws2.sqs
- 
- SqsIO.WriteBatches.DynamicDestination<T> - Interface in org.apache.beam.sdk.io.aws2.sqs
-  
- SqsIO.WriteBatches.EntryMapperFn<T> - Interface in org.apache.beam.sdk.io.aws2.sqs
- 
- SqsIO.WriteBatches.EntryMapperFn.Builder<T> - Interface in org.apache.beam.sdk.io.aws2.sqs
- 
- SqsIO.WriteBatches.Result - Class in org.apache.beam.sdk.io.aws2.sqs
- 
- SqsMessage - Class in org.apache.beam.sdk.io.aws2.sqs
-  
- SqsMessage() - Constructor for class org.apache.beam.sdk.io.aws2.sqs.SqsMessage
-  
- src - Variable in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark.NewVsCopy.ArrayCopyState
-  
- src - Variable in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark.NewVsCopy.ArrayNewState
-  
- SSECustomerKey - Class in org.apache.beam.sdk.io.aws2.s3
- 
Customer provided key for use with Amazon S3 server-side encryption. 
- SSECustomerKey.Builder - Class in org.apache.beam.sdk.io.aws2.s3
-  
- SSECustomerKeyFactory() - Constructor for class org.apache.beam.sdk.io.aws2.options.S3Options.SSECustomerKeyFactory
-  
- stageArtifacts(RunnerApi.Pipeline) - Method in class org.apache.beam.runners.dataflow.DataflowRunner
-  
- StageBundleFactory - Interface in org.apache.beam.runners.fnexecution.control
- 
A bundle factory scoped to a particular  ExecutableStage, which has all of the resources it
 needs to provide new  RemoteBundles. 
- StagedFile() - Constructor for class org.apache.beam.runners.dataflow.util.PackageUtil.StagedFile
-  
- stageFiles(List<PackageUtil.StagedFile>) - Method in class org.apache.beam.runners.dataflow.util.GcsStager
- 
- stageFiles(List<PackageUtil.StagedFile>) - Method in interface org.apache.beam.runners.dataflow.util.Stager
- 
Stage files and return a list of packages  DataflowPackage objects describing th actual
 location at which each file was staged. 
- stagePackage(PackageUtil.PackageAttributes, Sleeper, CreateOptions) - Method in class org.apache.beam.runners.dataflow.util.PackageUtil
- 
Stages one file ("package") if necessary. 
- Stager - Interface in org.apache.beam.runners.dataflow.util
- 
Interface for staging files needed for running a Dataflow pipeline. 
- StagerFactory() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions.StagerFactory
-  
- stageToFile(byte[], String) - Method in class org.apache.beam.runners.dataflow.util.GcsStager
-  
- stageToFile(byte[], String, String, CreateOptions) - Method in class org.apache.beam.runners.dataflow.util.PackageUtil
-  
- stageToFile(byte[], String) - Method in interface org.apache.beam.runners.dataflow.util.Stager
- 
Stage bytes to a target file name wherever this stager stages things. 
- STAGING_TO_ARTIFACT_URN - Static variable in class org.apache.beam.runners.fnexecution.artifact.ArtifactRetrievalService
-  
- StagingLocationFactory() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowPipelineOptions.StagingLocationFactory
-  
- StandardCreateOptions() - Constructor for class org.apache.beam.sdk.io.fs.CreateOptions.StandardCreateOptions
-  
- start() - Method in class org.apache.beam.io.debezium.KafkaSourceConsumerFn.DebeziumSDFDatabaseHistory
-  
- start() - Method in class org.apache.beam.runners.jobsubmission.JobInvocation
- 
Start the job. 
- start() - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver
-  
- start() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
-  
- start() - Method in class org.apache.beam.runners.spark.metrics.sink.CsvSink
-  
- start() - Method in class org.apache.beam.runners.spark.metrics.sink.GraphiteSink
-  
- start() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.sink.CodahaleCsvSink
-  
- start() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.sink.CodahaleGraphiteSink
-  
- start() - Method in class org.apache.beam.sdk.extensions.python.PythonService
-  
- start() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataOutboundAggregator
- 
Starts the flushing daemon thread if data_buffer_time_limit_ms is set. 
- start() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.UnboundedReaderImpl
-  
- start() - Method in interface org.apache.beam.sdk.io.jms.AutoScaler
- 
The  AutoScaler is started when the  JmsIO.UnboundedJmsReader is started. 
- start() - Method in class org.apache.beam.sdk.io.jms.DefaultAutoscaler
-  
- start() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
-  
- start() - Method in interface org.apache.beam.sdk.io.solace.broker.MessageReceiver
- 
Starts the message receiver. 
- start() - Method in class org.apache.beam.sdk.io.solace.broker.SolaceMessageReceiver
-  
- start() - Method in class org.apache.beam.sdk.io.Source.Reader
- 
Initializes the reader and advances the reader to the first record. 
- start() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
- 
Initializes the reader and advances the reader to the first record. 
- start() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
- 
Returns the start of this window, inclusive. 
- start() - Method in class org.apache.beam.sdk.transformservice.launcher.TransformServiceLauncher
-  
- START_WITHS - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- startAt(Instant) - Method in class org.apache.beam.sdk.transforms.PeriodicImpulse
- 
Assign a timestamp when the pipeliene starts to produce data. 
- startBundle() - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
-  
- startBundle(DoFn<Row, Row>.StartBundleContext, PipelineOptions) - Method in interface org.apache.beam.sdk.extensions.sql.BeamSqlSeekableTable
-  
- startBundle(DoFn<Iterable<KV<DestinationT, WriteTables.Result>>, Iterable<KV<TableDestination, WriteTables.Result>>>.StartBundleContext) - Method in class org.apache.beam.sdk.io.gcp.bigquery.UpdateSchemaDestination
-  
- startBundle(DoFn<PubsubMessage, Void>.StartBundleContext) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write.PubsubBoundedWriter
-  
- startBundle() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.PubsubLiteSink
-  
- startBundle() - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- startBundle() - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- StartBundleContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.StartBundleContext
-  
- startCopyJob(JobReference, JobConfigurationTableCopy) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.JobService
- 
Start a BigQuery copy job. 
- startCopyJob(JobReference, JobConfigurationTableCopy) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
-  
- startExtractJob(JobReference, JobConfigurationExtract) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.JobService
- 
Start a BigQuery extract job. 
- startExtractJob(JobReference, JobConfigurationExtract) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
-  
- startImpl() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
-  
- startImpl() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
- 
- startLoadJob(JobReference, JobConfigurationLoad) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.JobService
- 
Start a BigQuery load job. 
- startLoadJob(JobReference, JobConfigurationLoad, AbstractInputStreamContent) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.JobService
- 
Start a BigQuery load job with stream content. 
- startLoadJob(JobReference, JobConfigurationLoad) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
-  
- startLoadJob(JobReference, JobConfigurationLoad, AbstractInputStreamContent) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
-  
- startProcess(String, String, List<String>, Map<String, String>) - Method in class org.apache.beam.runners.fnexecution.environment.ProcessManager
- 
Forks a process with the given command, arguments, and additional environment variables. 
- startProcess(String, String, List<String>, Map<String, String>, File) - Method in class org.apache.beam.runners.fnexecution.environment.ProcessManager
-  
- startQueryJob(JobReference, JobConfigurationQuery) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.JobService
- 
Start a BigQuery query job. 
- startQueryJob(JobReference, JobConfigurationQuery) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
-  
- startReading(ReadableByteChannel) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource.AvroReader
-  
- startReading(ReadableByteChannel) - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
- 
Creates a decompressing channel from the input channel and passes it to its delegate reader's
 FileBasedReader#startReading(ReadableByteChannel).
 
- startReading(ReadableByteChannel) - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
- 
Performs any initialization of the subclass of FileBasedReaderthat involves IO
 operations.
 
- startRunnerBundle(DoFnRunner<KV<?, ?>, OutputT>) - Method in class org.apache.beam.runners.jet.processors.StatefulParDoP
-  
- STARTS_WITH_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.BeamBuiltinMethods
-  
- startsWith(Path) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- startsWith(String) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- startsWith(String, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
-  
- startsWith(String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- startsWith(String) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
- 
- state(StreamObserver<BeamFnApi.StateResponse>) - Method in class org.apache.beam.runners.fnexecution.state.GrpcStateService
-  
- state(StateNamespace, StateTag<T>, StateContext<?>) - Method in class org.apache.beam.runners.spark.stateful.SparkStateInternals
-  
- State - Interface in org.apache.beam.sdk.state
- 
- STATE_CACHE_SIZE - Static variable in interface org.apache.beam.sdk.options.ExperimentalOptions
-  
- STATE_SAMPLING_PERIOD_MILLIS - Static variable in interface org.apache.beam.sdk.options.ExperimentalOptions
-  
- StateAndTimerBundleCheckpointHandler(TimerInternalsFactory<T>, StateInternalsFactory<T>, Coder<WindowedValue<T>>, Coder) - Constructor for class org.apache.beam.runners.fnexecution.control.BundleCheckpointHandlers.StateAndTimerBundleCheckpointHandler
-  
- StateAndTimers - Class in org.apache.beam.runners.spark.stateful
- 
State and Timers wrapper. 
- StateAndTimers() - Constructor for class org.apache.beam.runners.spark.stateful.StateAndTimers
-  
- StateBinder - Interface in org.apache.beam.sdk.state
- 
For internal use only; no backwards-compatibility guarantees. 
- StateContext<W extends BoundedWindow> - Interface in org.apache.beam.sdk.state
- 
For internal use only; no backwards-compatibility guarantees. 
- StateContexts - Class in org.apache.beam.sdk.state
- 
For internal use only; no backwards-compatibility guarantees. 
- StateContexts() - Constructor for class org.apache.beam.sdk.state.StateContexts
-  
- StateDelegator - Interface in org.apache.beam.runners.fnexecution.state
- 
The  StateDelegator is able to delegate  BeamFnApi.StateRequests to a set of registered
 handlers. 
- StateDelegator.Registration - Interface in org.apache.beam.runners.fnexecution.state
- 
Allows callers to deregister from receiving further state requests. 
- StatefulParDoP<OutputT> - Class in org.apache.beam.runners.jet.processors
- 
Jet Processorimplementation for Beam's stateful ParDo primitive.
 
- StatefulParDoP.Supplier<OutputT> - Class in org.apache.beam.runners.jet.processors
- 
- stateInternals() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.NoOpStepContext
-  
- stateInternals() - Method in class org.apache.beam.runners.twister2.utils.NoOpStepContext
-  
- StateKeySpec - Class in org.apache.beam.sdk.state
-  
- StateRequestHandler - Interface in org.apache.beam.runners.fnexecution.state
- 
Handler for StateRequests.
 
- StateRequestHandlers - Class in org.apache.beam.runners.fnexecution.state
- 
- StateRequestHandlers() - Constructor for class org.apache.beam.runners.fnexecution.state.StateRequestHandlers
-  
- StateRequestHandlers.BagUserStateHandler<K,V,W extends BoundedWindow> - Interface in org.apache.beam.runners.fnexecution.state
- 
A handler for bag user state. 
- StateRequestHandlers.BagUserStateHandlerFactory<K,V,W extends BoundedWindow> - Interface in org.apache.beam.runners.fnexecution.state
- 
- StateRequestHandlers.IterableSideInputHandler<V,W extends BoundedWindow> - Interface in org.apache.beam.runners.fnexecution.state
- 
A handler for iterable side inputs. 
- StateRequestHandlers.MultimapSideInputHandler<K,V,W extends BoundedWindow> - Interface in org.apache.beam.runners.fnexecution.state
- 
A handler for multimap side inputs. 
- StateRequestHandlers.SideInputHandler - Interface in org.apache.beam.runners.fnexecution.state
- 
Marker interface that denotes some type of side input handler. 
- StateRequestHandlers.SideInputHandlerFactory - Interface in org.apache.beam.runners.fnexecution.state
- 
- StateSpec<StateT extends State> - Interface in org.apache.beam.sdk.state
- 
A specification of a persistent state cell. 
- StateSpec.Cases<ResultT> - Interface in org.apache.beam.sdk.state
- 
Cases for doing a "switch" on the type of  StateSpec. 
- StateSpec.Cases.WithDefault<ResultT> - Class in org.apache.beam.sdk.state
- 
A base class for a visitor with a default method for cases it is not interested in. 
- StateSpecFunctions - Class in org.apache.beam.runners.spark.stateful
- 
A class containing StateSpecmappingFunctions.
 
- StateSpecFunctions() - Constructor for class org.apache.beam.runners.spark.stateful.StateSpecFunctions
-  
- StateSpecs - Class in org.apache.beam.sdk.state
- 
- StaticGrpcProvisionService - Class in org.apache.beam.runners.fnexecution.provisioning
- 
A provision servicethat returns a static response to all calls.
 
- StaticRemoteEnvironment - Class in org.apache.beam.runners.fnexecution.environment
- 
- StaticRemoteEnvironmentFactory - Class in org.apache.beam.runners.fnexecution.environment
- 
An  EnvironmentFactory that creates StaticRemoteEnvironment used by a runner harness that
 would like to use an existing InstructionRequestHandler. 
- StaticRemoteEnvironmentFactory.Provider - Class in org.apache.beam.runners.fnexecution.environment
- 
Provider for StaticRemoteEnvironmentFactory. 
- StaticSchemaInference - Class in org.apache.beam.sdk.schemas.utils
- 
A set of utilities for inferring a Beam  Schema from static Java types. 
- StaticSchemaInference() - Constructor for class org.apache.beam.sdk.schemas.utils.StaticSchemaInference
-  
- status() - Method in class org.apache.beam.sdk.io.fs.MatchResult
- 
- status() - Method in class org.apache.beam.sdk.transformservice.launcher.TransformServiceLauncher
-  
- STATUS_BACKOFF_FACTORY - Static variable in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-  
- statusCode() - Method in class org.apache.beam.sdk.io.splunk.SplunkWriteError
-  
- statusMessage() - Method in class org.apache.beam.sdk.io.splunk.SplunkWriteError
-  
- stepName() - Method in class org.apache.beam.sdk.metrics.MetricKey
- 
The step name that is associated with this metric or Null if none is associated. 
- steps() - Method in class org.apache.beam.sdk.metrics.MetricsFilter
-  
- stop() - Method in class org.apache.beam.runners.jobsubmission.JobServerDriver
-  
- stop() - Method in class org.apache.beam.runners.spark.metrics.sink.CsvSink
-  
- stop() - Method in class org.apache.beam.runners.spark.metrics.sink.GraphiteSink
-  
- stop() - Method in class org.apache.beam.runners.spark.SparkPipelineResult
-  
- stop() - Method in class org.apache.beam.runners.spark.SparkRunnerDebugger.DebugSparkPipelineResult
-  
- stop() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.sink.CodahaleCsvSink
-  
- stop() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.sink.CodahaleGraphiteSink
-  
- stop() - Method in interface org.apache.beam.sdk.io.jms.AutoScaler
- 
The  AutoScaler is stopped when the  JmsIO.UnboundedJmsReader is closed. 
- stop() - Method in class org.apache.beam.sdk.io.jms.DefaultAutoscaler
-  
- stop() - Static method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
- 
Indicates that there is no more work to be done for the current element. 
- stopAfter(Duration) - Method in class org.apache.beam.sdk.transforms.PeriodicImpulse
- 
For internal use only; no backwards-compatibility guarantees. 
- stopAt(Instant) - Method in class org.apache.beam.sdk.transforms.PeriodicImpulse
- 
Assign a timestamp when the pipeliene stops producing data. 
- stopProcess(String) - Method in class org.apache.beam.runners.fnexecution.environment.ProcessManager
- 
Stops a previously started process identified by its unique id. 
- stopSampling(int, long) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.LimitNumberOfFiles
-  
- stopSampling(int, long) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.LimitNumberOfTotalBytes
-  
- stopSampling(int, long) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.SampleAllFiles
-  
- stopSampling(int, long) - Method in interface org.apache.beam.sdk.io.TextRowCountEstimator.SamplingStrategy
-  
- StorageApiCDC - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
Constants and variables for CDC support. 
- StorageApiCDC() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StorageApiCDC
-  
- StorageApiConvertMessages<DestinationT,ElementT> - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
A transform that converts messages to protocol buffers in preparation for writing to BigQuery. 
- StorageApiConvertMessages(StorageApiDynamicDestinations<ElementT, DestinationT>, BigQueryServices, TupleTag<BigQueryStorageApiInsertError>, TupleTag<KV<DestinationT, StorageApiWritePayload>>, Coder<BigQueryStorageApiInsertError>, Coder<KV<DestinationT, StorageApiWritePayload>>, SerializableFunction<ElementT, RowMutationInformation>, BadRecordRouter) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StorageApiConvertMessages
-  
- StorageApiConvertMessages.ConvertMessagesDoFn<DestinationT,ElementT> - Class in org.apache.beam.sdk.io.gcp.bigquery
-  
- StorageApiDynamicDestinationsTableRow<T,DestinationT> - Class in org.apache.beam.sdk.io.gcp.bigquery
-  
- StorageApiFlushAndFinalizeDoFn - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
This DoFn flushes and optionally (if requested) finalizes Storage API streams. 
- StorageApiFlushAndFinalizeDoFn(BigQueryServices) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StorageApiFlushAndFinalizeDoFn
-  
- StorageApiLoads<DestinationT,ElementT> - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
This  PTransform manages loads into BigQuery using the Storage API. 
- StorageApiLoads(Coder<DestinationT>, StorageApiDynamicDestinations<ElementT, DestinationT>, SerializableFunction<ElementT, RowMutationInformation>, BigQueryIO.Write.CreateDisposition, String, Duration, BigQueryServices, int, boolean, boolean, boolean, boolean, boolean, Predicate<String>, boolean, AppendRowsRequest.MissingValueInterpretation, Map<String, String>, BadRecordRouter, ErrorHandler<BadRecord, ?>) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StorageApiLoads
-  
- StorageApiWritePayload - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
Class used to wrap elements being sent to the Storage API sinks. 
- StorageApiWritePayload() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWritePayload
-  
- StorageApiWritePayload.Builder - Class in org.apache.beam.sdk.io.gcp.bigquery
-  
- StorageApiWriteRecordsInconsistent<DestinationT,ElementT> - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
A transform to write sharded records to BigQuery using the Storage API. 
- StorageApiWriteRecordsInconsistent(StorageApiDynamicDestinations<ElementT, DestinationT>, BigQueryServices, TupleTag<BigQueryStorageApiInsertError>, TupleTag<TableRow>, Predicate<String>, Coder<BigQueryStorageApiInsertError>, Coder<TableRow>, boolean, boolean, BigQueryIO.Write.CreateDisposition, String, boolean, AppendRowsRequest.MissingValueInterpretation, Map<String, String>) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteRecordsInconsistent
-  
- StorageApiWritesShardedRecords<DestinationT,ElementT> - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
A transform to write sharded records to BigQuery using the Storage API (Streaming). 
- StorageApiWritesShardedRecords(StorageApiDynamicDestinations<ElementT, DestinationT>, BigQueryIO.Write.CreateDisposition, String, BigQueryServices, Coder<DestinationT>, Coder<BigQueryStorageApiInsertError>, Coder<TableRow>, TupleTag<BigQueryStorageApiInsertError>, TupleTag<TableRow>, Predicate<String>, boolean, boolean, AppendRowsRequest.MissingValueInterpretation, Map<String, String>) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWritesShardedRecords
-  
- StorageApiWriteUnshardedRecords<DestinationT,ElementT> - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
Write records to the Storage API using a standard batch approach. 
- StorageApiWriteUnshardedRecords(StorageApiDynamicDestinations<ElementT, DestinationT>, BigQueryServices, TupleTag<BigQueryStorageApiInsertError>, TupleTag<TableRow>, Predicate<String>, Coder<BigQueryStorageApiInsertError>, Coder<TableRow>, boolean, boolean, BigQueryIO.Write.CreateDisposition, String, boolean, AppendRowsRequest.MissingValueInterpretation, Map<String, String>) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords
-  
- storageExists() - Method in class org.apache.beam.io.debezium.KafkaSourceConsumerFn.DebeziumSDFDatabaseHistory
-  
- StorageLevelFactory() - Constructor for class org.apache.beam.runners.spark.SparkCommonPipelineOptions.StorageLevelFactory
-  
- storageObject() - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.StorageObjectOrIOException
- 
- StorageObjectOrIOException() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.StorageObjectOrIOException
-  
- storeId - Variable in class org.apache.beam.sdk.io.gcp.healthcare.WebPathParser.DicomWebPath
-  
- storeRecord(HistoryRecord) - Method in class org.apache.beam.io.debezium.KafkaSourceConsumerFn.DebeziumSDFDatabaseHistory
-  
- STREAM_PARTITION_PREFIX - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableAdminDao
-  
- STREAMING_ENGINE_EXPERIMENT - Static variable in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
- 
Experiment to turn on the Streaming Engine experiment. 
- StreamingInserts<DestinationT,ElementT> - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
PTransform that performs streaming BigQuery write. 
- StreamingInserts(BigQueryIO.Write.CreateDisposition, DynamicDestinations<?, DestinationT>, Coder<ElementT>, SerializableFunction<ElementT, TableRow>, SerializableFunction<ElementT, TableRow>) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StreamingInserts
- 
Constructor. 
- StreamingInsertsMetrics - Interface in org.apache.beam.sdk.io.gcp.bigquery
- 
Stores and exports metrics for a batch of Streaming Inserts RPCs. 
- StreamingInsertsMetrics.NoOpStreamingInsertsMetrics - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
No-op implementation of StreamingInsertsResults.
 
- StreamingInsertsMetrics.StreamingInsertsMetricsImpl - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
Metrics of a batch of InsertAll RPCs. 
- StreamingInsertsMetricsImpl() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StreamingInsertsMetrics.StreamingInsertsMetricsImpl
-  
- StreamingIT - Interface in org.apache.beam.sdk.testing
- 
- StreamingLogLevel - Enum in org.apache.beam.sdk.io.snowflake.enums
-  
- StreamingOptions - Interface in org.apache.beam.sdk.options
- 
Options used to configure streaming. 
- StreamingSideInputHandlerFactory - Class in org.apache.beam.runners.fnexecution.translation
- 
StateRequestHandler that uses  SideInputHandler to
 access the broadcast state that represents side inputs. 
- StreamingSourceContextImpl - Class in org.apache.beam.sdk.io.cdap.context
- 
Class for creating context object of different CDAP classes with stream source type. 
- StreamingSourceContextImpl() - Constructor for class org.apache.beam.sdk.io.cdap.context.StreamingSourceContextImpl
-  
- StreamingWriteTables<ElementT> - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
- StreamingWriteTables() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StreamingWriteTables
-  
- StreamPartitionWithWatermark - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.model
-  
- StreamPartitionWithWatermark(Range.ByteStringRange, Instant) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.StreamPartitionWithWatermark
-  
- StreamProgress - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction
- 
- StreamProgress() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.StreamProgress
-  
- StreamProgress(ChangeStreamContinuationToken, Instant, BigDecimal, Instant, boolean) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.StreamProgress
-  
- StreamProgress(CloseStream) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.StreamProgress
-  
- streamToTable(SnowflakeServices, ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
-  
- StreamTransformTranslator<TransformT extends PTransform> - Interface in org.apache.beam.runners.twister2.translators
- 
Stream TransformTranslator interface. 
- STRING - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- STRING - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
The type of string fields. 
- STRING_TYPES - Static variable in enum org.apache.beam.sdk.schemas.Schema.TypeName
-  
- StringAgg - Class in org.apache.beam.sdk.extensions.sql.impl.udaf
- 
Combine.CombineFns for aggregating strings or bytes with an optional delimiter (default comma). 
- StringAgg() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.udaf.StringAgg
-  
- StringAgg.StringAggByte - Class in org.apache.beam.sdk.extensions.sql.impl.udaf
- 
- StringAgg.StringAggString - Class in org.apache.beam.sdk.extensions.sql.impl.udaf
- 
- StringAggByte(byte[]) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.udaf.StringAgg.StringAggByte
-  
- StringAggString(String) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.udaf.StringAgg.StringAggString
-  
- StringBuilderBundle() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.StringBuilderBundle
-  
- StringBundle() - Constructor for class org.apache.beam.sdk.jmh.schemas.RowBundles.StringBundle
-  
- StringCompiler - Class in org.apache.beam.sdk.schemas.transforms.providers
-  
- StringCompiler() - Constructor for class org.apache.beam.sdk.schemas.transforms.providers.StringCompiler
-  
- StringCompiler.CompileException - Exception in org.apache.beam.sdk.schemas.transforms.providers
-  
- StringDelegateCoder<T> - Class in org.apache.beam.sdk.coders
- 
A  Coder that wraps a  Coder<String> and encodes/decodes values via string
 representations. 
- StringDelegateCoder(Class<T>, TypeDescriptor<T>) - Constructor for class org.apache.beam.sdk.coders.StringDelegateCoder
-  
- StringFunctions - Class in org.apache.beam.sdk.extensions.sql.zetasql.translation.impl
- 
StringFunctions. 
- StringFunctions() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- strings() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
- 
- stringSet(String, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
- 
Create a metric that accumulates and reports set of unique string values. 
- stringSet(Class<?>, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
- 
Create a metric that accumulates and reports set of unique string values. 
- StringSet - Interface in org.apache.beam.sdk.metrics
- 
A metric that reports set of unique string values. 
- StringSetImpl - Class in org.apache.beam.runners.jet.metrics
- 
- StringSetImpl(MetricName) - Constructor for class org.apache.beam.runners.jet.metrics.StringSetImpl
-  
- StringSetResult - Class in org.apache.beam.sdk.metrics
- 
- StringSetResult() - Constructor for class org.apache.beam.sdk.metrics.StringSetResult
-  
- StringSetResult.EmptyStringSetResult - Class in org.apache.beam.sdk.metrics
- 
- StringUtf8Coder - Class in org.apache.beam.sdk.coders
- 
A  Coder that encodes  Strings in UTF-8 encoding. 
- stripGetterPrefix(String) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
-  
- StripIdsDoFn() - Constructor for class org.apache.beam.sdk.values.ValueWithRecordId.StripIdsDoFn
-  
- stripPartitionDecorator(String) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
- 
Strip off any partition decorator information from a tablespec. 
- stripPrefix(String, String) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
-  
- stripSetterPrefix(String) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
-  
- Structs - Class in org.apache.beam.runners.dataflow.util
- 
A collection of static methods for manipulating datastructure representations transferred via the
 Dataflow API. 
- StructuralByteArray - Class in org.apache.beam.sdk.coders
- 
A wrapper around a byte[] that uses structural, value-based equality rather than byte[]'s normal
 object identity. 
- StructuralByteArray(byte[]) - Constructor for class org.apache.beam.sdk.coders.StructuralByteArray
-  
- StructuralKey<K> - Class in org.apache.beam.runners.local
- 
- structuralValue(byte[]) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
- 
Returns an object with an Object.equals()method that represents structural equality on
 the argument.
 
- structuralValue(T) - Method in class org.apache.beam.sdk.coders.Coder
- 
Returns an object with an Object.equals()method that represents structural equality on
 the argument.
 
- structuralValue(T) - Method in class org.apache.beam.sdk.coders.DelegateCoder
- 
Returns an object with an Object.equals()method that represents structural equality on
 the argument.
 
- structuralValue(Deque<T>) - Method in class org.apache.beam.sdk.coders.DequeCoder
-  
- structuralValue(Iterable<T>) - Method in class org.apache.beam.sdk.coders.IterableCoder
-  
- structuralValue(KV<K, V>) - Method in class org.apache.beam.sdk.coders.KvCoder
-  
- structuralValue(List<T>) - Method in class org.apache.beam.sdk.coders.ListCoder
-  
- structuralValue(Map<K, V>) - Method in class org.apache.beam.sdk.coders.MapCoder
-  
- structuralValue(T) - Method in class org.apache.beam.sdk.coders.NullableCoder
-  
- structuralValue(T) - Method in class org.apache.beam.sdk.coders.SerializableCoder
- 
The structural value of the object is the object itself. 
- structuralValue(SortedMap<K, V>) - Method in class org.apache.beam.sdk.coders.SortedMapCoder
-  
- structuralValue(T) - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
-  
- structuralValue(Void) - Method in class org.apache.beam.sdk.coders.VoidCoder
-  
- structuralValue(T) - Method in class org.apache.beam.sdk.coders.ZstdCoder
- 
Returns an object with an Object.equals()method that represents structural equality on
 the argument.
 
- structuralValue(KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
-  
- structuralValue(ProducerRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
-  
- structuralValue(TimestampedValue<T>) - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
-  
- structuralValueConsistentWithEquals(Coder<T>, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<T>and values of typeT, the structural
 values are equal if and only if the encoded bytes are equal.
 
- structuralValueConsistentWithEqualsInContext(Coder<T>, Coder.Context, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<T>,Coder.Context, and values of typeT, the structural values are equal if and only if the encoded bytes are equal, in anyCoder.Context.
 
- structuralValueDecodeEncodeEqual(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<T>and value of typeT, the structural value
 is equal to the structural value yield by encoding and decoding the original value.
 
- structuralValueDecodeEncodeEqualInContext(Coder<T>, Coder.Context, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<T>,Coder.Context, and value of typeT, the structural value is equal to the structural value yield by encoding and decoding the
 original value, in anyCoder.Context.
 
- structuralValueDecodeEncodeEqualIterable(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<T>and value of typeT, the structural value
 of the content of the Iterable is equal to the structural value yield by encoding and decoding
 the original value.
 
- structuralValueDecodeEncodeEqualIterableInContext(Coder<T>, Coder.Context, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
Verifies that for the given Coder<T>,Coder.Context, and value of typeT, the structural content of the Iterable of the value is equal to the structural value yield
 by encoding and decoding the original value, in anyCoder.Context.
 
- StructuredCoder<T> - Class in org.apache.beam.sdk.coders
- 
- StructuredCoder() - Constructor for class org.apache.beam.sdk.coders.StructuredCoder
-  
- studyId - Variable in class org.apache.beam.sdk.io.gcp.healthcare.WebPathParser.DicomWebPath
-  
- subclassGetterInterface(ByteBuddy, Type, Type) - Static method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils
-  
- subclassSetterInterface(ByteBuddy, Type, Type) - Static method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils
-  
- submitFn - Variable in class org.apache.beam.sdk.io.aws2.common.AsyncBatchWriteHandler
-  
- subpath(int, int) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- subPathMatches(String, String) - Static method in class org.apache.beam.sdk.metrics.MetricFiltering
- 
subPathMatches(haystack, needle)returns true ifneedlerepresents a path
 withinhaystack.
 
- SubscriberOptions - Class in org.apache.beam.sdk.io.gcp.pubsublite
-  
- SubscriberOptions() - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.SubscriberOptions
-  
- SubscriberOptions.Builder - Class in org.apache.beam.sdk.io.gcp.pubsublite
-  
- SubscribeTransform - Class in org.apache.beam.sdk.io.gcp.pubsublite.internal
-  
- SubscribeTransform(SubscriberOptions) - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.internal.SubscribeTransform
-  
- SubscriptionPartition - Class in org.apache.beam.sdk.io.gcp.pubsublite.internal
-  
- SubscriptionPartition() - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.internal.SubscriptionPartition
-  
- SubscriptionPartitionCoder - Class in org.apache.beam.sdk.io.gcp.pubsublite.internal
-  
- SubscriptionPartitionCoder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.internal.SubscriptionPartitionCoder
-  
- subscriptionPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
- 
- subscriptionPath() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.SubscriberOptions
-  
- subscriptionPathFromName(String, String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
-  
- subscriptionPathFromPath(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
-  
- substr(String, long, long) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- SUBSTR - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- SUBSTR_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.BeamBuiltinMethods
-  
- SUBSTR_PARAMETER_EXCEED_INTEGER - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- subTriggers - Variable in class org.apache.beam.sdk.transforms.windowing.Trigger
-  
- subTriggers() - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
-  
- SUCCESS - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.Write
- 
The tag for the successful writes to HL7v2 store`. 
- success(String, String, Metadata) - Static method in class org.apache.beam.sdk.io.tika.ParseResult
-  
- success(String, String) - Static method in class org.apache.beam.sdk.io.tika.ParseResult
-  
- success() - Static method in class org.apache.beam.sdk.testing.SuccessOrFailure
-  
- SUCCESS_COUNTER - Static variable in class org.apache.beam.sdk.testing.PAssert
-  
- SUCCESS_TAG - Static variable in class org.apache.beam.sdk.extensions.ml.RecommendationAICreateCatalogItem
-  
- SUCCESS_TAG - Static variable in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportCatalogItems
-  
- SUCCESS_TAG - Static variable in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportUserEvents
-  
- SUCCESS_TAG - Static variable in class org.apache.beam.sdk.extensions.ml.RecommendationAIPredict
-  
- SUCCESS_TAG - Static variable in class org.apache.beam.sdk.extensions.ml.RecommendationAIWriteUserEvent
-  
- SUCCESSFUL_BODY - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write
- 
The tag for successful writes to FHIR store. 
- SUCCESSFUL_BUNDLES - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.ExecuteBundles
- 
The TupleTag used for bundles that were executed successfully. 
- SUCCESSFUL_PUBLISH_TAG - Static variable in class org.apache.beam.sdk.io.solace.SolaceIO.Write
-  
- SUCCESSFUL_WRITES - Static variable in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
-  
- SuccessOrFailure - Class in org.apache.beam.sdk.testing
- 
- Sum() - Constructor for class org.apache.beam.sdk.extensions.sql.provider.UdfTestProvider.Sum
-  
- Sum - Class in org.apache.beam.sdk.transforms
- 
PTransforms for computing the sum of the elements in aPCollection, or the sum of
 the values associated with each key in aPCollectionofKVs.
 
- supplier(Coder, Coder, WindowingStrategy<T, BoundedWindow>, String) - Static method in class org.apache.beam.runners.jet.processors.AssignWindowP
-  
- supplier(BoundedSource<T>, SerializablePipelineOptions, Coder, String) - Static method in class org.apache.beam.runners.jet.processors.BoundedSourceP
-  
- Supplier(Map<String, Coder>, Coder, String) - Constructor for class org.apache.beam.runners.jet.processors.FlattenP.Supplier
-  
- supplier(Coder, String) - Static method in class org.apache.beam.runners.jet.processors.ImpulseP
-  
- Supplier(String, String, DoFn<InputT, OutputT>, WindowingStrategy<?, ?>, DoFnSchemaInformation, SerializablePipelineOptions, TupleTag<OutputT>, Set<TupleTag<OutputT>>, Coder<InputT>, Map<PCollectionView<?>, Coder<?>>, Map<TupleTag<?>, Coder<?>>, Coder<InputT>, Map<TupleTag<?>, Coder<?>>, Collection<PCollectionView<?>>, Map<String, PCollectionView<?>>) - Constructor for class org.apache.beam.runners.jet.processors.ParDoP.Supplier
-  
- Supplier(String, String, DoFn<KV<?, ?>, OutputT>, WindowingStrategy<?, ?>, DoFnSchemaInformation, SerializablePipelineOptions, TupleTag<OutputT>, Set<TupleTag<OutputT>>, Coder<KV<?, ?>>, Map<PCollectionView<?>, Coder<?>>, Map<TupleTag<?>, Coder<?>>, Coder<KV<?, ?>>, Map<TupleTag<?>, Coder<?>>, Collection<PCollectionView<?>>, Map<String, PCollectionView<?>>) - Constructor for class org.apache.beam.runners.jet.processors.StatefulParDoP.Supplier
-  
- supplier(UnboundedSource<T, CmT>, SerializablePipelineOptions, Coder, String) - Static method in class org.apache.beam.runners.jet.processors.UnboundedSourceP
-  
- supplier(Coder, Coder, WindowingStrategy<?, ?>, String) - Static method in class org.apache.beam.runners.jet.processors.ViewP
-  
- supplier(SerializablePipelineOptions, WindowedValue.WindowedValueCoder<KV<K, V>>, Coder, WindowingStrategy, String) - Static method in class org.apache.beam.runners.jet.processors.WindowGroupP
-  
- SUPPORTED_FORMATS - Static variable in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider
-  
- SUPPORTED_FORMATS - Static variable in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider
-  
- SUPPORTED_FORMATS_STR - Static variable in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider
-  
- SUPPORTED_FORMATS_STR - Static variable in class org.apache.beam.sdk.io.kafka.KafkaWriteSchemaTransformProvider
-  
- supportsCondition() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.CalcRelSplitter.RelType
-  
- supportsProjectionPushdown() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-  
- supportsProjectionPushdown() - Method in interface org.apache.beam.sdk.schemas.ProjectionProducer
- 
Whether thissupports projection pushdown.
 
- supportsProjects() - Method in class org.apache.beam.sdk.extensions.sql.meta.BaseBeamTable
-  
- supportsProjects() - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTable
- 
Whether project push-down is supported by the IO API. 
- supportsProjects() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable
-  
- SynchronizedStreamObserver<V> - Class in org.apache.beam.sdk.fn.stream
- 
A StreamObserverwhich provides synchronous access access to an underlyingStreamObserver.
 
- SystemReduceFnBuffering<K,T,W extends BoundedWindow> - Class in org.apache.beam.runners.twister2.translators.functions.internal
-  
- SystemReduceFnBuffering() - Constructor for class org.apache.beam.runners.twister2.translators.functions.internal.SystemReduceFnBuffering
-  
- SystemReduceFnBuffering(Coder<T>) - Constructor for class org.apache.beam.runners.twister2.translators.functions.internal.SystemReduceFnBuffering
-  
- T__0 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- T__0 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- T__1 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- T__1 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- T__2 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- T__2 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- T__3 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- T__3 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- T__4 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- T__4 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- table() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
-  
- Table - Class in org.apache.beam.sdk.extensions.sql.meta
- 
Represents the metadata of a BeamSqlTable.
 
- Table() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.Table
-  
- table() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
-  
- Table.Builder - Class in org.apache.beam.sdk.extensions.sql.meta
- 
- TABLE_FIELD_SCHEMAS - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.HealthcareIOErrorToTableRow
-  
- TABLE_ROW_ERROR_CONTAINER - Static variable in interface org.apache.beam.sdk.io.gcp.bigquery.ErrorContainer
-  
- TableAndQuery() - Constructor for class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher.TableAndQuery
-  
- TableAndRecord<T> - Class in org.apache.beam.sdk.io.kudu
- 
A wrapper for a KuduTableand theTrepresenting a typed record.
 
- TableAndRecord(KuduTable, T) - Constructor for class org.apache.beam.sdk.io.kudu.TableAndRecord
-  
- TableDestination - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
Encapsulates a BigQuery table destination. 
- TableDestination(String, String) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- TableDestination(TableReference, String) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- TableDestination(TableReference, String, TimePartitioning) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- TableDestination(String, String, TimePartitioning) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- TableDestination(String, String, TimePartitioning, Clustering) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- TableDestination(String, String, String) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- TableDestination(TableReference, String, String) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- TableDestination(TableReference, String, String, String) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- TableDestination(String, String, String, String) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- TableDestinationCoder - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
- TableDestinationCoderV2 - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
- TableDestinationCoderV2() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV2
-  
- TableDestinationCoderV3 - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
- TableDestinationCoderV3() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV3
-  
- tableExists() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao
- 
Checks whether the metadata table already exists in the database. 
- tableFieldToProtoTableField(TableFieldSchema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProto
-  
- TableName - Class in org.apache.beam.sdk.extensions.sql.impl
- 
Represents a parsed table name that is specified in a FROM clause (and other places). 
- TableName() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.TableName
-  
- TableNameExtractionUtils - Class in org.apache.beam.sdk.extensions.sql
- 
Helper class to extract table identifiers from the query. 
- TableNameExtractionUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.TableNameExtractionUtils
-  
- TableProvider - Interface in org.apache.beam.sdk.extensions.sql.meta.provider
- 
A TableProviderhandles the metadata CRUD of a specified kind of tables.
 
- tableReference() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery
-  
- TableResolution - Class in org.apache.beam.sdk.extensions.sql.zetasql
- 
Utility methods to resolve a table, given a top-level Calcite schema and a table path. 
- TableResolution() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.TableResolution
-  
- tableRowFromBeamRow() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
-  
- tableRowFromMessage(Message, boolean, Predicate<String>) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProto
-  
- tableRowFromMessage(Message, boolean, Predicate<String>, String) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProto
-  
- TableRowJsonCoder - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
A  Coder that encodes BigQuery  TableRow objects in their native JSON format. 
- tableRows(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
-  
- tableRowToBeamRow() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
-  
- TableRowToStorageApiProto - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
Utility methods for converting JSON  TableRow objects to dynamic protocol message, for use
 with the Storage write API. 
- TableRowToStorageApiProto() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProto
-  
- TableRowToStorageApiProto.SchemaDoesntMatchException - Exception in org.apache.beam.sdk.io.gcp.bigquery
-  
- TableRowToStorageApiProto.SchemaTooNarrowException - Exception in org.apache.beam.sdk.io.gcp.bigquery
-  
- TableRowToStorageApiProto.SingleValueConversionException - Exception in org.apache.beam.sdk.io.gcp.bigquery
-  
- tables() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
-  
- tableSchema() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
-  
- TableSchema - Class in org.apache.beam.sdk.io.clickhouse
- 
A descriptor for ClickHouse table schema. 
- TableSchema() - Constructor for class org.apache.beam.sdk.io.clickhouse.TableSchema
-  
- TableSchema.Column - Class in org.apache.beam.sdk.io.clickhouse
- 
A column in ClickHouse table. 
- TableSchema.ColumnType - Class in org.apache.beam.sdk.io.clickhouse
- 
A descriptor for a column type. 
- TableSchema.DefaultType - Enum in org.apache.beam.sdk.io.clickhouse
- 
An enumeration of possible kinds of default values in ClickHouse. 
- TableSchema.TypeName - Enum in org.apache.beam.sdk.io.clickhouse
- 
An enumeration of possible types in ClickHouse. 
- TableSchemaCache - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
An updatable cache for table schemas. 
- TableSchemaUpdateUtils - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
Helper utilities for handling schema-update responses. 
- TableSchemaUpdateUtils() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableSchemaUpdateUtils
-  
- tableSpec() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery
-  
- TableUtils - Class in org.apache.beam.sdk.extensions.sql
-  
- TableWithRows(long, Table) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider.TableWithRows
-  
- TaggedKeyedPCollection(TupleTag<V>, PCollection<KV<K, V>>) - Constructor for class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple.TaggedKeyedPCollection
-  
- TaggedPValue - Class in org.apache.beam.sdk.values
- 
For internal use only; no backwards-compatibility guarantees. 
- TaggedPValue() - Constructor for class org.apache.beam.sdk.values.TaggedPValue
-  
- take(String, Duration) - Method in interface org.apache.beam.runners.fnexecution.control.ControlClientPool.Source
- 
- take() - Method in class org.apache.beam.sdk.fn.CancellableQueue
- 
Takes an element from this queue. 
- takeOutputElements() - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- takeOutputElements(TupleTag<T>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- takeOutputElementsWithTimestamp() - Method in class org.apache.beam.sdk.transforms.DoFnTester
- 
- tanh(Double) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinTrigonometricFunctions
- 
TANH(X) 
- targetForRootUrl(String) - Static method in interface org.apache.beam.sdk.io.gcp.pubsub.PubsubOptions
- 
- TDigestQuantiles - Class in org.apache.beam.sdk.extensions.sketching
- 
PTransforms for getting information about quantiles in a stream.
 
- TDigestQuantiles() - Constructor for class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles
-  
- TDigestQuantiles.GlobalDigest - Class in org.apache.beam.sdk.extensions.sketching
- 
- TDigestQuantiles.PerKeyDigest<K> - Class in org.apache.beam.sdk.extensions.sketching
- 
- TDigestQuantiles.TDigestQuantilesFn - Class in org.apache.beam.sdk.extensions.sketching
- 
- teardown() - Method in interface org.apache.beam.io.requestresponse.SetupTeardown
- 
Called during the  DoFn's teardown lifecycle method. 
- teardown() - Method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides.BatchStatefulDoFn
-  
- tearDown() - Method in interface org.apache.beam.sdk.extensions.sql.BeamSqlSeekableTable
- 
cleanup resources of the instance. 
- teardown() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadSpannerSchema
-  
- teardown() - Method in interface org.apache.beam.sdk.io.kafka.CheckStopReadingFn
-  
- teardown() - Method in class org.apache.beam.sdk.io.pulsar.ReadFromPulsarDoFn
-  
- teardown() - Method in class org.apache.beam.sdk.io.pulsar.WriteToPulsarDoFn
-  
- teardown() - Method in class org.apache.beam.sdk.io.solace.write.UnboundedSolaceWriter
-  
- tearDown() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark.ProtobufByteStringOutputStream
-  
- tearDown() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark.SdkCoreByteStringOutputStream
-  
- tearDown(Blackhole) - Method in class org.apache.beam.sdk.jmh.util.VarIntBenchmark.ByteStringOutput
-  
- Tee<T> - Class in org.apache.beam.sdk.transforms
- 
A PTransform that returns its input, but also applies its input to an auxiliary PTransform, akin
 to the shell teecommand, which is named after the T-splitter used in plumbing.
 
- TEMP_FILES - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write
- 
The tag for temp files for import to FHIR store. 
- test(RunnerApi.PTransform) - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.IsFlinkNativeTransform
-  
- test(RunnerApi.PTransform) - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.IsFlinkNativeTransform
-  
- TestBigQuery - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
Test rule which creates a new table with specified schema, with randomized name and exposes few
 APIs to work with it. 
- TestBigQuery.PollingAssertion - Interface in org.apache.beam.sdk.io.gcp.bigquery
- 
Interface to implement a polling assertion. 
- TestBigQuery.RowsAssertion - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
Interface for creating a polling eventual assertion. 
- TestBigQueryOptions - Interface in org.apache.beam.sdk.io.gcp.bigquery
- 
- TestBoundedTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.test
- 
Mocked table for bounded data sources. 
- TestBoundedTable(Schema) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestBoundedTable
-  
- testByteCount(Coder<T>, Coder.Context, T[]) - Static method in class org.apache.beam.sdk.testing.CoderProperties
- 
A utility method that passes the given (unencoded) elements through coder's
 registerByteSizeObserver() and encode() methods, and confirms they are mutually consistent. 
- testCombineFn(Combine.CombineFn<InputT, AccumT, OutputT>, List<InputT>, OutputT) - Static method in class org.apache.beam.sdk.testing.CombineFnTester
- 
Tests that the  Combine.CombineFn, when applied to the provided input, produces the provided
 output. 
- testCombineFn(Combine.CombineFn<InputT, AccumT, OutputT>, List<InputT>, Matcher<? super OutputT>) - Static method in class org.apache.beam.sdk.testing.CombineFnTester
-  
- testCopyArray(ByteStringOutputStreamBenchmark.NewVsCopy.ArrayCopyState, Blackhole) - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark.NewVsCopy
-  
- TestDataflowPipelineOptions - Interface in org.apache.beam.runners.dataflow
- 
- TestDataflowRunner - Class in org.apache.beam.runners.dataflow
- 
- TestElementByteSizeObserver() - Constructor for class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
-  
- TestExecutors - Class in org.apache.beam.sdk.fn.test
- 
A  TestRule that validates that all submitted tasks finished and were completed. 
- TestExecutors() - Constructor for class org.apache.beam.sdk.fn.test.TestExecutors
-  
- TestExecutors.TestExecutorService - Interface in org.apache.beam.sdk.fn.test
- 
A union of the  ExecutorService and  TestRule interfaces. 
- TestFlinkRunner - Class in org.apache.beam.runners.flink
- 
Test Flink runner. 
- testingPipelineOptions() - Static method in class org.apache.beam.sdk.testing.TestPipeline
- 
- TestJobService - Class in org.apache.beam.runners.portability.testing
- 
A JobService for tests. 
- TestJobService(Endpoints.ApiServiceDescriptor, String, String, JobApi.JobState.Enum, JobApi.MetricResults) - Constructor for class org.apache.beam.runners.portability.testing.TestJobService
-  
- testNewArray(ByteStringOutputStreamBenchmark.NewVsCopy.ArrayNewState, Blackhole) - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark.NewVsCopy
-  
- TestPipeline - Class in org.apache.beam.sdk.testing
- 
A creator of test pipelines that can be used inside of tests that can be configured to run
 locally or against a remote pipeline runner. 
- TestPipeline.AbandonedNodeException - Exception in org.apache.beam.sdk.testing
- 
An exception thrown in case an abandoned  PTransform is
 detected, that is, a  PTransform that has not been run. 
- TestPipeline.PipelineRunMissingException - Exception in org.apache.beam.sdk.testing
- 
An exception thrown in case a test finishes without invoking  Pipeline.run(). 
- TestPipeline.TestValueProviderOptions - Interface in org.apache.beam.sdk.testing
- 
- TestPipelineOptions - Interface in org.apache.beam.sdk.testing
- 
- TestPipelineOptions.AlwaysPassMatcher - Class in org.apache.beam.sdk.testing
- 
Matcher which will always pass. 
- TestPipelineOptions.AlwaysPassMatcherFactory - Class in org.apache.beam.sdk.testing
- 
- TestPortablePipelineOptions - Interface in org.apache.beam.runners.portability.testing
- 
- TestPortablePipelineOptions.DefaultJobServerConfigFactory - Class in org.apache.beam.runners.portability.testing
- 
Factory for default config. 
- TestPortablePipelineOptions.TestPortablePipelineOptionsRegistrar - Class in org.apache.beam.runners.portability.testing
- 
- TestPortablePipelineOptionsRegistrar() - Constructor for class org.apache.beam.runners.portability.testing.TestPortablePipelineOptions.TestPortablePipelineOptionsRegistrar
-  
- TestPortableRunner - Class in org.apache.beam.runners.portability.testing
- 
- TestPrismPipelineOptions - Interface in org.apache.beam.runners.prism
- 
- TestPrismRunner - Class in org.apache.beam.runners.prism
- 
- testProtobufByteStringOutputStreamFewLargeWrites() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testProtobufByteStringOutputStreamFewMixedWritesWithoutReuse() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testProtobufByteStringOutputStreamFewMixedWritesWithReuse(ByteStringOutputStreamBenchmark.ProtobufByteStringOutputStream) - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testProtobufByteStringOutputStreamFewSmallWrites() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testProtobufByteStringOutputStreamFewTinyWrites() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testProtobufByteStringOutputStreamManyLargeWrites() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testProtobufByteStringOutputStreamManyMixedWritesWithoutReuse() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testProtobufByteStringOutputStreamManyMixedWritesWithReuse(ByteStringOutputStreamBenchmark.ProtobufByteStringOutputStream) - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testProtobufByteStringOutputStreamManySmallWrites() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testProtobufByteStringOutputStreamManyTinyWrites() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- TestPubsub - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Test rule which creates a new topic and subscription with randomized names and exposes the APIs
 to work with them. 
- TestPubsub.PollingAssertion - Interface in org.apache.beam.sdk.io.gcp.pubsub
-  
- TestPubsubOptions - Interface in org.apache.beam.sdk.io.gcp.pubsub
- 
- TestPubsubSignal - Class in org.apache.beam.sdk.io.gcp.pubsub
- 
Test rule which observes elements of the  PCollection and checks whether they match the
 success criteria. 
- TestSchemaTransformProvider - Class in org.apache.beam.sdk.managed.testing
-  
- TestSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.managed.testing.TestSchemaTransformProvider
-  
- TestSchemaTransformProvider.Config - Class in org.apache.beam.sdk.managed.testing
-  
- TestSchemaTransformProvider.Config.Builder - Class in org.apache.beam.sdk.managed.testing
-  
- testSdkCoreByteStringOutputStreamFewLargeWrites() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testSdkCoreByteStringOutputStreamFewMixedWritesWithoutReuse() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testSdkCoreByteStringOutputStreamFewMixedWritesWithReuse(ByteStringOutputStreamBenchmark.SdkCoreByteStringOutputStream) - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testSdkCoreByteStringOutputStreamFewSmallWrites() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testSdkCoreByteStringOutputStreamFewTinyWrites() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testSdkCoreByteStringOutputStreamManyLargeWrites() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testSdkCoreByteStringOutputStreamManyMixedWritesWithoutReuse() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testSdkCoreByteStringOutputStreamManyMixedWritesWithReuse(ByteStringOutputStreamBenchmark.SdkCoreByteStringOutputStream) - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testSdkCoreByteStringOutputStreamManySmallWrites() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- testSdkCoreByteStringOutputStreamManyTinyWrites() - Method in class org.apache.beam.sdk.jmh.util.ByteStringOutputStreamBenchmark
-  
- TestSparkPipelineOptions - Interface in org.apache.beam.runners.spark
- 
- TestSparkPipelineOptions.DefaultStopPipelineWatermarkFactory - Class in org.apache.beam.runners.spark
- 
A factory to provide the default watermark to stop a pipeline that reads from an unbounded
 source. 
- TestSparkRunner - Class in org.apache.beam.runners.spark
- 
The SparkRunner translate operations defined on a pipeline to a representation executable by
 Spark, and then submitting the job to Spark to be executed. 
- TestStream<T> - Class in org.apache.beam.sdk.testing
- 
A testing input that generates an unbounded  PCollection of elements, advancing the
 watermark and processing time as elements are emitted. 
- TestStream.Builder<T> - Class in org.apache.beam.sdk.testing
- 
- TestStream.ElementEvent<T> - Class in org.apache.beam.sdk.testing
- 
- TestStream.Event<T> - Interface in org.apache.beam.sdk.testing
- 
- TestStream.EventType - Enum in org.apache.beam.sdk.testing
- 
- TestStream.ProcessingTimeEvent<T> - Class in org.apache.beam.sdk.testing
- 
- TestStream.TestStreamCoder<T> - Class in org.apache.beam.sdk.testing
- 
- TestStream.WatermarkEvent<T> - Class in org.apache.beam.sdk.testing
- 
- TestStreams - Class in org.apache.beam.sdk.fn.test
- 
Utility methods which enable testing of StreamObservers.
 
- TestStreams() - Constructor for class org.apache.beam.sdk.fn.test.TestStreams
-  
- TestStreams.Builder<T> - Class in org.apache.beam.sdk.fn.test
- 
A builder for a test CallStreamObserverthat performs various callbacks.
 
- TestTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.test
- 
Base class for mocked table. 
- TestTable(Schema) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTable
-  
- TestTableFilter - Class in org.apache.beam.sdk.extensions.sql.meta.provider.test
-  
- TestTableFilter(List<RexNode>) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableFilter
-  
- TestTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.test
- 
Test in-memory table provider for use in tests. 
- TestTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
-  
- TestTableProvider.PushDownOptions - Enum in org.apache.beam.sdk.extensions.sql.meta.provider.test
-  
- TestTableProvider.TableWithRows - Class in org.apache.beam.sdk.extensions.sql.meta.provider.test
- 
TableWitRows. 
- TestTableUtils - Class in org.apache.beam.sdk.extensions.sql.meta.provider.test
- 
Utility functions for mock classes. 
- TestTableUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableUtils
-  
- TestUnboundedTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.test
- 
A mocked unbounded table. 
- TestUniversalRunner - Class in org.apache.beam.runners.portability.testing
- 
- TestUniversalRunner.Options - Interface in org.apache.beam.runners.portability.testing
-  
- TestUniversalRunner.OptionsRegistrar - Class in org.apache.beam.runners.portability.testing
- 
- TestUniversalRunner.RunnerRegistrar - Class in org.apache.beam.runners.portability.testing
- 
Registrar for the portable runner. 
- TextIO - Class in org.apache.beam.sdk.io
- 
- TextIO.CompressionType - Enum in org.apache.beam.sdk.io
- 
- TextIO.Read - Class in org.apache.beam.sdk.io
- 
- TextIO.ReadAll - Class in org.apache.beam.sdk.io
- 
- TextIO.ReadFiles - Class in org.apache.beam.sdk.io
- 
- TextIO.Sink - Class in org.apache.beam.sdk.io
- 
- TextIO.TypedWrite<UserT,DestinationT> - Class in org.apache.beam.sdk.io
- 
- TextIO.Write - Class in org.apache.beam.sdk.io
- 
- TextJsonTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.text
- 
- TextJsonTable(Schema, String, TextTableProvider.JsonToRow, TextTableProvider.RowToJson) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextJsonTable
-  
- TextMessageMapper - Class in org.apache.beam.sdk.io.jms
- 
The TextMessageMapper takes a Stringvalue, aSessionand returns aTextMessage.
 
- TextMessageMapper() - Constructor for class org.apache.beam.sdk.io.jms.TextMessageMapper
-  
- TextRowCountEstimator - Class in org.apache.beam.sdk.io
- 
This returns a row count estimation for files associated with a file pattern. 
- TextRowCountEstimator() - Constructor for class org.apache.beam.sdk.io.TextRowCountEstimator
-  
- TextRowCountEstimator.Builder - Class in org.apache.beam.sdk.io
- 
- TextRowCountEstimator.LimitNumberOfFiles - Class in org.apache.beam.sdk.io
- 
This strategy stops sampling if we sample enough number of bytes. 
- TextRowCountEstimator.LimitNumberOfTotalBytes - Class in org.apache.beam.sdk.io
- 
This strategy stops sampling when total number of sampled bytes are more than some threshold. 
- TextRowCountEstimator.NoEstimationException - Exception in org.apache.beam.sdk.io
- 
An exception that will be thrown if the estimator cannot get an estimation of the number of
 lines. 
- TextRowCountEstimator.SampleAllFiles - Class in org.apache.beam.sdk.io
- 
This strategy samples all the files. 
- TextRowCountEstimator.SamplingStrategy - Interface in org.apache.beam.sdk.io
- 
Sampling Strategy shows us when should we stop reading further files. 
- TextSource - Class in org.apache.beam.sdk.io
- 
- TextSource(ValueProvider<String>, EmptyMatchTreatment, byte[], int) - Constructor for class org.apache.beam.sdk.io.TextSource
-  
- TextSource(ValueProvider<String>, EmptyMatchTreatment, byte[]) - Constructor for class org.apache.beam.sdk.io.TextSource
-  
- TextSource(MatchResult.Metadata, long, long, byte[], int) - Constructor for class org.apache.beam.sdk.io.TextSource
-  
- TextSource(MatchResult.Metadata, long, long, byte[]) - Constructor for class org.apache.beam.sdk.io.TextSource
-  
- TextSourceBenchmark - Class in org.apache.beam.sdk.jmh.io
-  
- TextSourceBenchmark() - Constructor for class org.apache.beam.sdk.jmh.io.TextSourceBenchmark
-  
- TextSourceBenchmark.Data - Class in org.apache.beam.sdk.jmh.io
-  
- TextTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.text
- 
- TextTable(Schema, String, PTransform<PCollection<String>, PCollection<Row>>, PTransform<PCollection<Row>, PCollection<String>>) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTable
- 
Text table with the specified read and write transforms. 
- TextTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.text
- 
Text table provider. 
- TextTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTableProvider
-  
- TextTableProvider.CsvToRow - Class in org.apache.beam.sdk.extensions.sql.meta.provider.text
- 
Read-side converter for  TextTable with format  'csv'. 
- TextTableProvider.LinesReadConverter - Class in org.apache.beam.sdk.extensions.sql.meta.provider.text
- 
Read-side converter for  TextTable with format  'lines'. 
- TextTableProvider.LinesWriteConverter - Class in org.apache.beam.sdk.extensions.sql.meta.provider.text
- 
Write-side converter for for  TextTable with format  'lines'. 
- TextualIntegerCoder - Class in org.apache.beam.sdk.coders
- 
A  Coder that encodes  Integer Integers as the ASCII bytes of their textual,
 decimal, representation. 
- TextualIntegerCoder() - Constructor for class org.apache.beam.sdk.coders.TextualIntegerCoder
-  
- TFRecordIO - Class in org.apache.beam.sdk.io
- 
PTransforms for reading and writing TensorFlow TFRecord files. 
- TFRecordIO.CompressionType - Enum in org.apache.beam.sdk.io
- 
- TFRecordIO.Read - Class in org.apache.beam.sdk.io
- 
- TFRecordIO.ReadFiles - Class in org.apache.beam.sdk.io
- 
- TFRecordIO.Sink - Class in org.apache.beam.sdk.io
- 
- TFRecordIO.Write - Class in org.apache.beam.sdk.io
- 
- that(PCollection<T>) - Static method in class org.apache.beam.sdk.testing.PAssert
- 
- that(String, PCollection<T>) - Static method in class org.apache.beam.sdk.testing.PAssert
- 
- thatFlattened(PCollectionList<T>) - Static method in class org.apache.beam.sdk.testing.PAssert
- 
- thatFlattened(String, PCollectionList<T>) - Static method in class org.apache.beam.sdk.testing.PAssert
- 
- thatList(PCollectionList<T>) - Static method in class org.apache.beam.sdk.testing.PAssert
- 
- thatMap(PCollection<KV<K, V>>) - Static method in class org.apache.beam.sdk.testing.PAssert
- 
- thatMap(String, PCollection<KV<K, V>>) - Static method in class org.apache.beam.sdk.testing.PAssert
- 
- thatMultimap(PCollection<KV<K, V>>) - Static method in class org.apache.beam.sdk.testing.PAssert
- 
- thatMultimap(String, PCollection<KV<K, V>>) - Static method in class org.apache.beam.sdk.testing.PAssert
- 
- thatSingleton(PCollection<T>) - Static method in class org.apache.beam.sdk.testing.PAssert
- 
Constructs a  PAssert.SingletonAssert for the value of the provided  PCollection
 PCollection<T>, which must be a singleton. 
- thatSingleton(String, PCollection<T>) - Static method in class org.apache.beam.sdk.testing.PAssert
- 
Constructs a  PAssert.SingletonAssert for the value of the provided  PCollection
 PCollection<T> with the specified reason. 
- thatSingletonIterable(PCollection<? extends Iterable<T>>) - Static method in class org.apache.beam.sdk.testing.PAssert
- 
- thatSingletonIterable(String, PCollection<? extends Iterable<T>>) - Static method in class org.apache.beam.sdk.testing.PAssert
- 
- threadSafe(WatermarkEstimator<WatermarkEstimatorStateT>) - Static method in class org.apache.beam.sdk.fn.splittabledofn.WatermarkEstimators
- 
Returns a thread safe  WatermarkEstimator which allows getting a snapshot of the current
 watermark and watermark estimator state. 
- ThriftCoder<T> - Class in org.apache.beam.sdk.io.thrift
- 
A  Coder using a Thrift  TProtocol to
 serialize/deserialize elements. 
- ThriftCoder(Class<T>, TProtocolFactory) - Constructor for class org.apache.beam.sdk.io.thrift.ThriftCoder
-  
- ThriftIO - Class in org.apache.beam.sdk.io.thrift
- 
PTransforms for reading and writing files containing Thrift encoded data. 
- ThriftIO.ReadFiles<T> - Class in org.apache.beam.sdk.io.thrift
- 
- ThriftIO.Sink<T extends org.apache.thrift.TBase<?,?>> - Class in org.apache.beam.sdk.io.thrift
- 
- ThriftIO.ThriftWriter<T extends org.apache.thrift.TBase<?,?>> - Class in org.apache.beam.sdk.io.thrift
- 
Writer to write Thrift object to OutputStream.
 
- ThriftPayloadSerializerProvider - Class in org.apache.beam.sdk.io.thrift
-  
- ThriftPayloadSerializerProvider() - Constructor for class org.apache.beam.sdk.io.thrift.ThriftPayloadSerializerProvider
-  
- ThriftSchema - Class in org.apache.beam.sdk.io.thrift
- 
Schema provider for generated thrift types. 
- ThriftSchema.Customizer - Class in org.apache.beam.sdk.io.thrift
-  
- THROTTLE_TIME_COUNTER_NAME - Static variable in class org.apache.beam.sdk.metrics.Metrics
-  
- THROTTLE_TIME_NAMESPACE - Static variable in class org.apache.beam.sdk.metrics.Metrics
-  
- THROTTLED_TIME - Static variable in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySinkMetrics
-  
- throttledBaseBackoff(Duration) - Method in class org.apache.beam.sdk.io.aws2.common.RetryConfiguration.Builder
-  
- throttledBaseBackoff() - Method in class org.apache.beam.sdk.io.aws2.common.RetryConfiguration
-  
- throttledTimeCounter(BigQuerySinkMetrics.RpcMethod) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySinkMetrics
-  
- THROUGHPUT_WINDOW_SECONDS - Static variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.ChangeStreamsConstants
- 
The sliding window size in seconds for throughput reporting. 
- ThroughputEstimator<T> - Interface in org.apache.beam.sdk.io.gcp.spanner.changestreams.estimator
- 
An estimator to calculate the throughput of the outputted elements from a DoFn. 
- throwable() - Method in class org.apache.beam.sdk.values.EncodableThrowable
- 
Returns the underlying Throwable.
 
- ThrowableHandler() - Constructor for class org.apache.beam.sdk.transforms.WithFailures.ThrowableHandler
-  
- throwableToGRPCCodeString(Throwable) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQuerySinkMetrics
- 
Converts a Throwable to a gRPC Status code. 
- THROWING_ROUTER - Static variable in interface org.apache.beam.sdk.transforms.errorhandling.BadRecordRouter
-  
- ThrowingBadRecordRouter() - Constructor for class org.apache.beam.sdk.transforms.errorhandling.BadRecordRouter.ThrowingBadRecordRouter
-  
- ThrowingBiConsumer<T1,T2> - Interface in org.apache.beam.sdk.function
- 
A BiConsumerwhich can throwExceptions.
 
- ThrowingBiFunction<T1,T2,T3> - Interface in org.apache.beam.sdk.function
- 
A BiFunctionwhich can throwExceptions.
 
- ThrowingConsumer<ExceptionT extends java.lang.Exception,T> - Interface in org.apache.beam.sdk.function
- 
A Consumerwhich can throwExceptions.
 
- ThrowingFunction<T1,T2> - Interface in org.apache.beam.sdk.function
- 
A Functionwhich can throwExceptions.
 
- ThrowingRunnable - Interface in org.apache.beam.sdk.function
- 
A Runnablewhich can throwExceptions.
 
- throwNullCredentialException() - Static method in class org.apache.beam.sdk.extensions.gcp.auth.NullCredentialInitializer
-  
- TikaIO - Class in org.apache.beam.sdk.io.tika
- 
Transforms for parsing arbitrary files using  Apache Tika. 
- TikaIO() - Constructor for class org.apache.beam.sdk.io.tika.TikaIO
-  
- TikaIO.Parse - Class in org.apache.beam.sdk.io.tika
- 
- TikaIO.ParseFiles - Class in org.apache.beam.sdk.io.tika
- 
- TIME - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- time() - Method in class org.apache.beam.sdk.io.splunk.SplunkEvent
-  
- TIME - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.SqlTypes
- 
Beam LogicalType corresponding to ZetaSQL/CalciteSQL TIME type. 
- Time - Class in org.apache.beam.sdk.schemas.logicaltypes
- 
A time without a time-zone. 
- Time() - Constructor for class org.apache.beam.sdk.schemas.logicaltypes.Time
-  
- TIME_FIELD_NAME - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.DateTime
-  
- TIME_WITH_LOCAL_TZ - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- TimeConversion() - Constructor for class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimeConversion
-  
- TimeDomain - Enum in org.apache.beam.sdk.state
- 
TimeDomain specifies whether an operation is based on timestamps of elements or current
 "real-world" time as reported while processing. 
- timeDomain() - Method in class org.apache.beam.sdk.transforms.DoFn.OnTimerContext
- 
Returns the time domain of the current timer. 
- TimeMicrosConversion() - Constructor for class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimeMicrosConversion
-  
- TimeMicrosConversion() - Constructor for class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimeMicrosConversion
-  
- TimeMillisConversion() - Constructor for class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimeMillisConversion
-  
- timer(String, String, String) - Static method in class org.apache.beam.sdk.fn.data.LogicalEndpoint
-  
- Timer - Interface in org.apache.beam.sdk.state
- 
A timer for a specified time domain that can be set to register the desire for further processing
 at particular time in its specified time domain. 
- timer(TimeDomain) - Static method in class org.apache.beam.sdk.state.TimerSpecs
-  
- TimerEndpoint<T> - Class in org.apache.beam.sdk.fn.data
-  
- TimerEndpoint() - Constructor for class org.apache.beam.sdk.fn.data.TimerEndpoint
-  
- timerId() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.TimerSpec
-  
- timerInternals() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.NoOpStepContext
-  
- timerInternals() - Method in class org.apache.beam.runners.twister2.utils.NoOpStepContext
-  
- TimerMap - Interface in org.apache.beam.sdk.state
-  
- timerMap(TimeDomain) - Static method in class org.apache.beam.sdk.state.TimerSpecs
-  
- TimerReceiverFactory - Class in org.apache.beam.runners.fnexecution.control
- 
- TimerReceiverFactory(StageBundleFactory, BiConsumer<Timer<?>, TimerInternals.TimerData>, Coder) - Constructor for class org.apache.beam.runners.fnexecution.control.TimerReceiverFactory
-  
- Timers - Interface in org.apache.beam.sdk.state
- 
Interface for interacting with time. 
- TimerSpec() - Constructor for class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.TimerSpec
-  
- TimerSpec - Interface in org.apache.beam.sdk.state
- 
A specification for a  Timer. 
- TimerSpecs - Class in org.apache.beam.sdk.state
- 
- TimerSpecs() - Constructor for class org.apache.beam.sdk.state.TimerSpecs
-  
- TimerUtils - Class in org.apache.beam.runners.spark.util
-  
- TimerUtils() - Constructor for class org.apache.beam.runners.spark.util.TimerUtils
-  
- timestamp() - Method in class org.apache.beam.runners.twister2.utils.Twister2AssignContext
-  
- TIMESTAMP - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- timestamp(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.TimestampFunctions
-  
- timestamp(String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.TimestampFunctions
-  
- timestamp(Integer) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.TimestampFunctions
-  
- timestamp(Integer, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.TimestampFunctions
-  
- TIMESTAMP - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.SqlTypes
- 
Beam LogicalType corresponding to ZetaSQL TIMESTAMP type. 
- timestamp() - Method in class org.apache.beam.sdk.transforms.DoFn.OnTimerContext
- 
Returns the output timestamp of the current timer. 
- timestamp() - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContext
- 
Returns the timestamp of the input element. 
- timestamp() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn.AssignContext
- 
Returns the timestamp of the current element. 
- TIMESTAMP_FIELD_NAME - Static variable in class org.apache.beam.sdk.io.gcp.healthcare.HealthcareIOErrorToTableRow
-  
- TIMESTAMP_MAX_VALUE - Static variable in class org.apache.beam.sdk.transforms.windowing.BoundedWindow
- 
The maximum value for any Beam timestamp. 
- TIMESTAMP_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.BeamBuiltinMethods
-  
- TIMESTAMP_MICROS - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.RowUtils
-  
- TIMESTAMP_MIN_VALUE - Static variable in class org.apache.beam.sdk.transforms.windowing.BoundedWindow
- 
The minimum value for any Beam timestamp. 
- TIMESTAMP_OP - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- TIMESTAMP_WITH_LOCAL_TZ - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- timestampColumnIndex(int) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestUnboundedTable
-  
- TimestampCombiner - Enum in org.apache.beam.sdk.transforms.windowing
- 
Policies for combining timestamps that occur within a window. 
- TimeStampComparator() - Constructor for class org.apache.beam.runners.dataflow.util.MonitoringUtil.TimeStampComparator
-  
- TimestampConversion() - Constructor for class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimestampConversion
-  
- TimestampConvert() - Constructor for class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.TimestampConvert
-  
- TimestampConverter - Class in org.apache.beam.sdk.io.gcp.bigtable.changestreams
- 
Convert between different Timestamp and Instant classes. 
- TimestampConverter() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.changestreams.TimestampConverter
-  
- timestamped(Iterable<TimestampedValue<T>>) - Static method in class org.apache.beam.sdk.transforms.Create
- 
- timestamped(TimestampedValue<T>, TimestampedValue<T>...) - Static method in class org.apache.beam.sdk.transforms.Create
- 
- timestamped(Iterable<T>, Iterable<Long>) - Static method in class org.apache.beam.sdk.transforms.Create
- 
Returns a new root transform that produces a  PCollection containing the specified
 elements with the specified timestamps. 
- TimestampedValue<V> - Class in org.apache.beam.sdk.values
- 
An immutable pair of a value and a timestamp. 
- TimestampedValue(V, Instant) - Constructor for class org.apache.beam.sdk.values.TimestampedValue
-  
- TimestampedValue.TimestampedValueCoder<T> - Class in org.apache.beam.sdk.values
- 
- TimestampEncoding - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.encoder
- 
This encoder/decoder writes a com.google.cloud.Timestamp object as a pair of long and int to avro
 and reads a Timestamp object from the same pair. 
- TimestampEncoding() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.encoder.TimestampEncoding
-  
- TimestampFunctions - Class in org.apache.beam.sdk.extensions.sql.zetasql.translation.impl
- 
TimestampFunctions. 
- TimestampFunctions() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.TimestampFunctions
-  
- TimestampMicrosConversion() - Constructor for class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimestampMicrosConversion
-  
- TimestampMicrosConversion() - Constructor for class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimestampMicrosConversion
-  
- TimestampMillisConversion() - Constructor for class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimestampMillisConversion
-  
- timestampMsSinceEpoch() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.IncomingMessage
- 
Timestamp for element (ms since epoch). 
- TimestampObservingWatermarkEstimator<WatermarkEstimatorStateT> - Interface in org.apache.beam.sdk.transforms.splittabledofn
- 
- TimestampPolicy<K,V> - Class in org.apache.beam.sdk.io.kafka
- 
A timestamp policy to assign event time for messages in a Kafka partition and watermark for it. 
- TimestampPolicy() - Constructor for class org.apache.beam.sdk.io.kafka.TimestampPolicy
-  
- TimestampPolicy.PartitionContext - Class in org.apache.beam.sdk.io.kafka
- 
The context contains state maintained in the reader for the partition. 
- TimestampPolicyFactory<KeyT,ValueT> - Interface in org.apache.beam.sdk.io.kafka
- 
An extendable factory to create a  TimestampPolicy for each partition at runtime by
 KafkaIO reader. 
- TimestampPolicyFactory.LogAppendTimePolicy<K,V> - Class in org.apache.beam.sdk.io.kafka
- 
Assigns Kafka's log append time (server side ingestion time) to each record. 
- TimestampPolicyFactory.ProcessingTimePolicy<K,V> - Class in org.apache.beam.sdk.io.kafka
- 
A simple policy that uses current time for event time and watermark. 
- TimestampPolicyFactory.TimestampFnPolicy<K,V> - Class in org.apache.beam.sdk.io.kafka
- 
Internal policy to support deprecated withTimestampFn API. 
- TimestampPrefixingWindowCoder<T extends BoundedWindow> - Class in org.apache.beam.sdk.coders
- 
- TimestampRange - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction
- 
A restriction represented by a range of timestamps [from, to). 
- TimestampRangeTracker - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction
- 
- TimestampRangeTracker(TimestampRange) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRangeTracker
-  
- timestamps() - Static method in class org.apache.beam.sdk.transforms.Reify
- 
- timestampsInValue() - Static method in class org.apache.beam.sdk.transforms.Reify
- 
Create a  PTransform that will output all input  KVs with the timestamp inside
 the value. 
- TimestampTransform - Class in org.apache.beam.sdk.transforms.windowing
- 
For internal use only; no backwards-compatibility guarantees. 
- TimestampTransform.AlignTo - Class in org.apache.beam.sdk.transforms.windowing
- 
For internal use only; no backwards-compatibility guarantees. 
- TimestampTransform.Delay - Class in org.apache.beam.sdk.transforms.windowing
- 
For internal use only; no backwards-compatibility guarantees. 
- TimestampUtils - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction
- 
Provides methods in order to convert timestamp to nanoseconds representation and back. 
- TimestampUtils() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampUtils
-  
- timeSupplier - Variable in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRangeTracker
-  
- TimeUtil - Class in org.apache.beam.runners.dataflow.util
- 
A helper class for converting between Dataflow API and SDK time representations. 
- TimeUtil - Class in org.apache.beam.sdk.io.aws2.kinesis
- 
Time conversion utilities. 
- TimeWithLocalTzType() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.TimeWithLocalTzType
-  
- TINY_INT - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- TmpCheckpointDirFactory() - Constructor for class org.apache.beam.runners.spark.SparkCommonPipelineOptions.TmpCheckpointDirFactory
-  
- to(String) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
Writes to file(s) with the given output prefix. 
- to(ResourceId) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
Writes to file(s) with the given output prefix. 
- to(ValueProvider<String>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
- to(FileBasedSink.FilenamePolicy) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
- to(DynamicAvroDestinations<UserT, NewDestinationT, OutputT>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
- to(String) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- to(ResourceId) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- to(ValueProvider<String>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- to(FileBasedSink.FilenamePolicy) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
See TypedWrite#to(FilenamePolicy).
 
- to(DynamicAvroDestinations<T, ?, T>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- to(SqsIO.WriteBatches.DynamicDestination<T>) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.WriteBatches
- 
Dynamic record based destination to write to. 
- to(String) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.WriteBatches
- 
Queue url to write to. 
- to(String) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
Specifies a common directory for all generated files. 
- to(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- to(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- to(TableReference) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- to(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- to(SerializableFunction<ValueInSingleWindow<T>, TableDestination>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Writes to table specified by the specified table function. 
- to(DynamicDestinations<T, ?>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- to(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
- 
Publishes to the specified topic. 
- to(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
- 
- to(SerializableFunction<ValueInSingleWindow<T>, String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
- 
Provides a function to dynamically specify the target topic per message. 
- to(long) - Method in class org.apache.beam.sdk.io.GenerateSequence
- 
Specifies the maximum number to generate (exclusive). 
- to(TableIdentifier) - Method in class org.apache.beam.sdk.io.iceberg.IcebergIO.WriteRows
-  
- to(DynamicDestinations) - Method in class org.apache.beam.sdk.io.iceberg.IcebergIO.WriteRows
-  
- to(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
A table name to be written in Snowflake. 
- to(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
-  
- to(Solace.Topic) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Write
- 
Write to a Solace topic. 
- to(Solace.Queue) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Write
- 
Write to a Solace queue. 
- to(String) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Write
- 
Provide name of collection while reading from Solr. 
- to(String) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Writes to text files with the given prefix. 
- to(ResourceId) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
- to(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
- to(FileBasedSink.FilenamePolicy) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
- to(FileBasedSink.DynamicDestinations<UserT, NewDestinationT, String>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
- to(SerializableFunction<UserT, DefaultFilenamePolicy.Params>, DefaultFilenamePolicy.Params) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
- to(String) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- to(ResourceId) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- to(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- to(FileBasedSink.FilenamePolicy) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
See TypedWrite#to(FilenamePolicy).
 
- to(FileBasedSink.DynamicDestinations<String, ?, String>) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- to(SerializableFunction<String, DefaultFilenamePolicy.Params>, DefaultFilenamePolicy.Params) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- to(String) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
- 
Writes TFRecord file(s) with the given output prefix. 
- to(ResourceId) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
- 
Writes TFRecord file(s) with a prefix given by the specified resource. 
- to(FileBasedSink<UserT, DestinationT, OutputT>) - Static method in class org.apache.beam.sdk.io.WriteFiles
- 
Creates a  WriteFiles transform that writes to the given  FileBasedSink, letting
 the runner control how many different shards are produced. 
- to(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
- 
Writes to files with the given path prefix. 
- to(Class<OutputT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Convert
- 
- to(TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Convert
- 
- toAbsolutePath() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- toAdditionalInputs(Iterable<PCollectionView<?>>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
- 
- toAvroField(Schema.Field, String) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Get Avro Field from Beam Field. 
- toAvroSchema(Schema, String, String) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Converts a Beam Schema into an AVRO schema. 
- toAvroSchema(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
-  
- toAvroType(String, String) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.FixedBytesField
- 
Convert to an AVRO type. 
- toBaseType(OffsetTime) - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.TZTimeOnly
-  
- toBaseType(OffsetDateTime) - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.TZTimestamp
-  
- toBaseType(LocalTime) - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.UTCTimeOnly
-  
- toBaseType(Instant) - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.UTCTimestamp
-  
- toBaseType(LocalDate) - Method in class org.apache.beam.sdk.schemas.logicaltypes.Date
-  
- toBaseType(LocalDateTime) - Method in class org.apache.beam.sdk.schemas.logicaltypes.DateTime
-  
- toBaseType(EnumerationType.Value) - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
-  
- toBaseType(byte[]) - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
-  
- toBaseType(Instant) - Method in class org.apache.beam.sdk.schemas.logicaltypes.MicrosInstant
-  
- toBaseType(Duration) - Method in class org.apache.beam.sdk.schemas.logicaltypes.NanosDuration
-  
- toBaseType(Instant) - Method in class org.apache.beam.sdk.schemas.logicaltypes.NanosInstant
-  
- toBaseType(OneOfType.Value) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
-  
- toBaseType(T) - Method in class org.apache.beam.sdk.schemas.logicaltypes.PassThroughLogicalType
-  
- toBaseType(PythonCallableSource) - Method in class org.apache.beam.sdk.schemas.logicaltypes.PythonCallable
-  
- toBaseType(Schema) - Method in class org.apache.beam.sdk.schemas.logicaltypes.SchemaLogicalType
-  
- toBaseType(LocalTime) - Method in class org.apache.beam.sdk.schemas.logicaltypes.Time
-  
- toBaseType(UUID) - Method in class org.apache.beam.sdk.schemas.logicaltypes.UuidLogicalType
-  
- toBaseType(InputT) - Method in interface org.apache.beam.sdk.schemas.Schema.LogicalType
- 
- toBeamField(Schema.Field) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Get Beam Field from avro Field. 
- toBeamObject(Value, boolean) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlBeamTranslationUtils
-  
- toBeamObject(Value, Schema.FieldType, boolean) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlBeamTranslationUtils
-  
- toBeamRow(Value, Schema, boolean) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlBeamTranslationUtils
-  
- toBeamRow(GenericRecord, Schema, BigQueryUtils.ConversionOptions) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
-  
- toBeamRow(Schema, TableRow) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
Tries to convert a JSON  TableRow from BigQuery into a Beam  Row. 
- toBeamRow(Schema, TableSchema, TableRow) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
Tries to parse the JSON  TableRow from BigQuery. 
- toBeamRow() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformReadConfiguration
- 
Serializes configuration to a  Row. 
- toBeamRow() - Method in class org.apache.beam.sdk.io.singlestore.schematransform.SingleStoreSchemaTransformWriteConfiguration
- 
Serializes configuration to a  Row. 
- toBeamRow(String, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.YamlUtils
-  
- toBeamRow(String, Schema, boolean) - Static method in class org.apache.beam.sdk.schemas.utils.YamlUtils
-  
- toBeamRow(Map<String, Object>, Schema, boolean) - Static method in class org.apache.beam.sdk.schemas.utils.YamlUtils
-  
- toBeamRowStrict(GenericRecord, Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Strict conversion from AVRO to Beam, strict because it doesn't do widening or narrowing during
 conversion. 
- toBeamSchema(Schema) - Static method in class org.apache.beam.sdk.extensions.arrow.ArrowConversion.ArrowSchemaTranslator
-  
- toBeamSchema(List<Field>) - Static method in class org.apache.beam.sdk.extensions.arrow.ArrowConversion.ArrowSchemaTranslator
-  
- toBeamSchema(Class<?>) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Converts AVRO schema to Beam row schema. 
- toBeamSchema(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Converts AVRO schema to Beam row schema. 
- toBeamSchema(ResultSetMetaData) - Static method in class org.apache.beam.sdk.io.jdbc.SchemaUtil
- 
Infers the Beam  Schema from  ResultSetMetaData. 
- toBeamType() - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.FixedBytesField
- 
Convert to a Beam type. 
- toBeamType(Type) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlBeamTranslationUtils
-  
- ToBigtableRowFn(Map<String, String>) - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.BeamRowToBigtableMutation.ToBigtableRowFn
-  
- toBuilder() - Method in class org.apache.beam.io.requestresponse.Monitoring
-  
- toBuilder() - Method in class org.apache.beam.sdk.extensions.sbe.SbeSchema.IrOptions
-  
- toBuilder() - Method in class org.apache.beam.sdk.extensions.sbe.UnsignedOptions
-  
- toBuilder() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table
-  
- toBuilder() - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.Globally
-  
- toBuilder() - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.PerKey
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.aws2.common.ClientConfiguration
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.aws2.common.RetryConfiguration
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.aws2.kinesis.WatermarkParameters
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.aws2.s3.S3FileSystemConfiguration
- 
- toBuilder() - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWritePayload
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.BatchGetDocuments
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.BatchWriteWithDeadLetterQueue
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.BatchWriteWithSummary
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.ListCollectionIds
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.ListDocuments
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.PartitionQuery
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.RunQuery
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
- 
- toBuilder() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.SubscriberOptions
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata
- 
Transforms the instance into a builder, so field values can be modified. 
- toBuilder() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.solace.data.Semp.Queue
-  
- toBuilder() - Method in class org.apache.beam.sdk.io.solace.data.Semp.QueueData
-  
- toBuilder() - Method in class org.apache.beam.sdk.schemas.Schema.Field
-  
- toBuilder() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
-  
- toBuilder() - Method in class org.apache.beam.sdk.transforms.DoFnSchemaInformation
-  
- toBuilder() - Method in class org.apache.beam.sdk.transforms.JsonToRow.JsonToRowWithErrFn.ParseWithError
-  
- toByteArray(T, Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
- 
Utility method for serializing an object using the specified coder. 
- toByteArray(T, Coder<T>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.CoderHelpers
- 
Utility method for serializing an object using the specified coder. 
- toByteArrays(Iterable<T>, Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
- 
Utility method for serializing a Iterable of values using the specified coder. 
- toByteArrayWithTs(T, Coder<T>, Instant) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
- 
Utility method for serializing an object using the specified coder, appending timestamp
 representation. 
- toByteFunction(Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
- 
A function wrapper for converting an object to a bytearray. 
- toByteFunction(Coder<K>, Coder<V>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
- 
A function wrapper for converting a key-value pair to a byte array pair. 
- toByteFunctionWithTs(Coder<K>, Coder<V>, Function<Tuple2<K, V>, Instant>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
- 
A function wrapper for converting a key-value pair to a byte array pair, where the key in
 resulting ByteArray contains (key, timestamp). 
- toBytes() - Method in class org.apache.beam.sdk.transforms.resourcehints.ResourceHint
- 
Defines how to represent the as bytestring. 
- toCalciteRowType(Schema, RelDataTypeFactory) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
- 
Create an instance of RelDataTypeso it can be used to create a table.
 
- toCalciteType(Type, boolean, RexBuilder) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlCalciteTranslationUtils
-  
- toCamelCase() - Method in class org.apache.beam.sdk.schemas.Schema
- 
Recursively converts all field names to `lowerCamelCase`. 
- toCamelCase() - Method in class org.apache.beam.sdk.values.Row
- 
Returns an equivalent  Row with `lowerCamelCase` field names. 
- toChangeStreamRecords(PartitionMetadata, ChangeStreamResultSet, ChangeStreamResultSetMetadata) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.mapper.ChangeStreamRecordMapper
- 
In GoogleSQL, change stream records are returned as an array of Struct.
 
- toCloudDuration(ReadableDuration) - Static method in class org.apache.beam.runners.dataflow.util.TimeUtil
- 
- toCloudObject(T, SdkComponents) - Method in interface org.apache.beam.runners.dataflow.util.CloudObjectTranslator
- 
Converts the provided object into an equivalent  CloudObject. 
- toCloudObject(RowCoder, SdkComponents) - Method in class org.apache.beam.runners.dataflow.util.RowCoderCloudObjectTranslator
- 
Convert to a cloud object. 
- toCloudObject(SchemaCoder, SdkComponents) - Method in class org.apache.beam.runners.dataflow.util.SchemaCoderCloudObjectTranslator
- 
Convert to a cloud object. 
- toCloudPubsubMessages() - Static method in class org.apache.beam.sdk.io.gcp.pubsublite.CloudPubsubTransforms
- 
Transform messages read from Pub/Sub Lite to their equivalent Cloud Pub/Sub Message that would
 have been read from PubsubIO. 
- toCloudTime(ReadableInstant) - Static method in class org.apache.beam.runners.dataflow.util.TimeUtil
- 
- toConfigRow(BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransform) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQuerySchemaTransformTranslation.BigQueryStorageReadSchemaTransformTranslator
-  
- toConfigRow(BigQueryWriteSchemaTransformProvider.BigQueryWriteSchemaTransform) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQuerySchemaTransformTranslation.BigQueryWriteSchemaTransformTranslator
-  
- toDefaultPolicies(SerializableFunction<UserT, DefaultFilenamePolicy.Params>, DefaultFilenamePolicy.Params, SerializableFunction<UserT, OutputT>) - Static method in class org.apache.beam.sdk.io.DynamicFileDestinations
- 
- toDuration(Row) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.DurationConvert
- 
ByteBuddy conversion for NanosDuration base type to Duration. 
- toEnumerable(BeamRelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter
-  
- toFeedRange() - Method in class org.apache.beam.sdk.io.azure.cosmos.NormalizedRange
-  
- toField(RelDataTypeField) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- toField(String, RelDataType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- toFieldType(SqlTypeNameSpec) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- toFieldType(SqlTypeName) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- toFieldType(RelDataType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- toFile() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- toGcpBackOff(BackOff) - Static method in class org.apache.beam.sdk.extensions.gcp.util.BackOffAdapter
- 
Returns an adapter to convert from BackOfftoBackOff.
 
- toGenericAvroSchema(TableSchema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
- toGenericAvroSchema(TableSchema, Boolean) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
- toGenericAvroSchema(String, List<TableFieldSchema>) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
- toGenericAvroSchema(String, List<TableFieldSchema>, Boolean) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
- toGenericRecord(Row) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Convert from a Beam Row to an AVRO GenericRecord. 
- toGenericRecord(Row, Schema) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils
- 
Convert from a Beam Row to an AVRO GenericRecord. 
- toHex(byte[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
-  
- toInputType(String) - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.TZTimeOnly
-  
- toInputType(String) - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.TZTimestamp
-  
- toInputType(String) - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.UTCTimeOnly
-  
- toInputType(String) - Method in class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.UTCTimestamp
-  
- toInputType(Long) - Method in class org.apache.beam.sdk.schemas.logicaltypes.Date
-  
- toInputType(Row) - Method in class org.apache.beam.sdk.schemas.logicaltypes.DateTime
-  
- toInputType(Integer) - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
-  
- toInputType(byte[]) - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
-  
- toInputType(BigDecimal) - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedPrecisionNumeric
-  
- toInputType(String) - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedString
-  
- toInputType(Row) - Method in class org.apache.beam.sdk.schemas.logicaltypes.MicrosInstant
-  
- toInputType(Row) - Method in class org.apache.beam.sdk.schemas.logicaltypes.NanosDuration
-  
- toInputType(Row) - Method in class org.apache.beam.sdk.schemas.logicaltypes.NanosInstant
-  
- toInputType(Row) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
-  
- toInputType(T) - Method in class org.apache.beam.sdk.schemas.logicaltypes.PassThroughLogicalType
-  
- toInputType(String) - Method in class org.apache.beam.sdk.schemas.logicaltypes.PythonCallable
-  
- toInputType(byte[]) - Method in class org.apache.beam.sdk.schemas.logicaltypes.SchemaLogicalType
-  
- toInputType(Long) - Method in class org.apache.beam.sdk.schemas.logicaltypes.Time
-  
- toInputType(Row) - Method in class org.apache.beam.sdk.schemas.logicaltypes.UuidLogicalType
-  
- toInputType(byte[]) - Method in class org.apache.beam.sdk.schemas.logicaltypes.VariableBytes
-  
- toInputType(String) - Method in class org.apache.beam.sdk.schemas.logicaltypes.VariableString
-  
- toInputType(BaseT) - Method in interface org.apache.beam.sdk.schemas.Schema.LogicalType
- 
- toInt(LocalDate, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.DateConversion
-  
- toInt(LocalTime, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimeMillisConversion
-  
- toInt(LocalDate, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.DateConversion
-  
- toInt(LocalTime, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimeConversion
-  
- toJava(Instant) - Static method in class org.apache.beam.sdk.io.aws2.kinesis.TimeUtil
-  
- toJoda(Instant) - Static method in class org.apache.beam.sdk.io.aws2.kinesis.TimeUtil
-  
- toJodaTime(Instant) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.TimestampConverter
-  
- toJson() - Method in class org.apache.beam.io.debezium.SourceRecordJson
- 
Transforms the extracted data to a JSON string. 
- ToJson<T> - Class in org.apache.beam.sdk.transforms
- 
Creates a  PTransform that serializes UTF-8 JSON objects from a  Schema-aware
 PCollection (i.e. 
- toJsonString(Object) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
-  
- tokenNames - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
- 
- tokenNames - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
- 
- ToListViewDoFn() - Constructor for class org.apache.beam.sdk.transforms.View.ToListViewDoFn
-  
- toLogicalBaseType(Schema.LogicalType<InputT, BaseT>, InputT) - Static method in class org.apache.beam.sdk.schemas.SchemaUtils
- 
Returns the base type given a logical type and the input type. 
- toLogicalInputType(Schema.LogicalType<InputT, BaseT>, BaseT) - Static method in class org.apache.beam.sdk.schemas.SchemaUtils
- 
Returns the input type given a logical type and the base type. 
- toLong(LocalDateTime, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.LocalTimestampMicrosConversion
-  
- toLong(LocalDateTime, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.LocalTimestampMillisConversion
-  
- toLong(LocalTime, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimeMicrosConversion
-  
- toLong(Instant, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimestampMicrosConversion
-  
- toLong(Instant, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.TimestampMillisConversion
-  
- toLong(LocalTime, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.LossyTimeMicrosConversion
-  
- toLong(DateTime, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.LossyTimestampMicrosConversion
-  
- toLong(DateTime, Schema, LogicalType) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJodaTimeConversions.TimestampConversion
-  
- toMap(ArrayData, ArrayData, DataType, DataType) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.Utils
-  
- toModel() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2Message
- 
To model message. 
- toModificationRel(RelOptCluster, RelOptTable, Prepare.CatalogReader, RelNode, TableModify.Operation, List<String>, List<RexNode>, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteTable
-  
- toNanos(Timestamp) - Static method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampUtils
- 
Converts the given timestamp to respective nanoseconds representation. 
- toNullableRecordField(Object[], int) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableUtils
-  
- Top - Class in org.apache.beam.sdk.transforms
- 
PTransforms for finding the largest (or smallest) set of elements in aPCollection, or the largest (or smallest) set of values associated with each key in aPCollectionofKVs.
 
- Top.Largest<T extends java.lang.Comparable<? super T>> - Class in org.apache.beam.sdk.transforms
- 
- Top.Natural<T extends java.lang.Comparable<? super T>> - Class in org.apache.beam.sdk.transforms
- 
A SerializableComparatorthat that uses the compared elements' natural
 ordering.
 
- Top.Reversed<T extends java.lang.Comparable<? super T>> - Class in org.apache.beam.sdk.transforms
- 
SerializableComparatorthat that uses the reverse of the compared elements'
 natural ordering.
 
- Top.Smallest<T extends java.lang.Comparable<? super T>> - Class in org.apache.beam.sdk.transforms
- 
- Top.TopCombineFn<T,ComparatorT extends java.util.Comparator<T> & java.io.Serializable> - Class in org.apache.beam.sdk.transforms
- 
CombineFnforToptransforms that combines a bunch ofTs into a singlecount-longList<T>, usingcompareFnto choose the largestTs.
 
- toPCollection(Pipeline, BeamRelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils
-  
- toPCollection(Pipeline, BeamRelNode, PTransform<PCollection<Row>, ? extends POutput>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils
-  
- TopCombineFn(int, ComparatorT) - Constructor for class org.apache.beam.sdk.transforms.Top.TopCombineFn
-  
- topic() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.OutgoingMessage
-  
- TopicPartitionCoder - Class in org.apache.beam.sdk.io.kafka
- 
- TopicPartitionCoder() - Constructor for class org.apache.beam.sdk.io.kafka.TopicPartitionCoder
-  
- topicPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
- 
Topic path where events will be published to. 
- topicPath() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PublisherOptions
-  
- topicPathFromName(String, String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
-  
- topicPathFromPath(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
-  
- toProto() - Method in class org.apache.beam.runners.jobsubmission.JobInvocation
- 
Convert to JobApi.JobInfo.
 
- toProto(PubsubMessage) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessages
-  
- toProvisionInfo() - Method in class org.apache.beam.runners.fnexecution.provisioning.JobInfo
-  
- toPTransform() - Method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortRead
-  
- toPTransform() - Method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortWrite
-  
- toRealPath(LinkOption...) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- toRecordField(Object[], int) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableUtils
-  
- toRel(RelOptTable.ToRelContext, RelOptTable) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteTable
-  
- toRelDataType(RelDataTypeFactory, Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- toResource(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
- toResource(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- toResource(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
- toResource(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- toResource(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
- 
- toResourceName() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- toRexNode(Value, RexBuilder) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlCalciteTranslationUtils
-  
- toRow(Duration) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.DurationConvert
- 
ByteBuddy conversion for Duration to NanosDuration base type. 
- toRow(Timestamp) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.TimestampConvert
- 
ByteBuddy conversion for Timestamp to NanosInstant base type. 
- toRow() - Method in class org.apache.beam.sdk.io.iceberg.SnapshotInfo
-  
- toRow(Schema) - Static method in class org.apache.beam.sdk.values.Row
- 
- toRowFunction(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.annotations.DefaultSchema.DefaultSchemaProvider
- 
Given a type, return a function that converts that type to a  Row object If no schema
 exists, returns null. 
- toRowFunction(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
- 
Deprecated. 
- toRowFunction(TypeDescriptor<T>) - Method in interface org.apache.beam.sdk.schemas.SchemaProvider
- 
Given a type, return a function that converts that type to a  Row object If no schema
 exists, returns null. 
- toRowList(BeamRelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter
-  
- toRowList(BeamRelNode, Map<String, String>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter
-  
- toRows() - Static method in class org.apache.beam.sdk.schemas.transforms.Convert
- 
- toSchema(RelDataType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
- 
Generate  Schema from  RelDataType which is used to create table. 
- toSchema() - Static method in class org.apache.beam.sdk.schemas.Schema
- 
- toSeconds(Instant) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.TimestampConverter
-  
- toSeq(ArrayData) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.Utils
-  
- toSeq(Collection<Object>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.Utils
-  
- toSnakeCase() - Method in class org.apache.beam.sdk.schemas.Schema
- 
Recursively converts all field names to `snake_case`. 
- toSnakeCase() - Method in class org.apache.beam.sdk.values.Row
- 
Returns an equivalent  Row with `snake_case` field names. 
- toSql(RexProgram, RexNode) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamSqlUnparseContext
-  
- toSqlTypeName(Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- toState(String) - Static method in class org.apache.beam.runners.dataflow.util.MonitoringUtil
-  
- toString() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator
-  
- toString() - Method in class org.apache.beam.runners.dataflow.DataflowRunner
-  
- toString() - Method in class org.apache.beam.runners.dataflow.TestDataflowRunner
-  
- toString() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
-  
- toString() - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
-  
- toString() - Method in class org.apache.beam.runners.flink.FlinkRunner
-  
- toString() - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
-  
- toString(Metric) - Static method in class org.apache.beam.runners.flink.metrics.Metrics
-  
- toString() - Method in class org.apache.beam.runners.portability.PortableRunner
-  
- toString() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
-  
- toString() - Method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.SparkWatermarks
-  
- toString() - Method in class org.apache.beam.sdk.coders.Coder.Context
- 
Deprecated. 
- toString() - Method in class org.apache.beam.sdk.coders.DelegateCoder
-  
- toString() - Method in class org.apache.beam.sdk.coders.SerializableCoder
-  
- toString() - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
-  
- toString() - Method in class org.apache.beam.sdk.coders.StructuralByteArray
-  
- toString() - Method in class org.apache.beam.sdk.coders.StructuredCoder
-  
- toString() - Method in class org.apache.beam.sdk.coders.ZstdCoder
-  
- toString() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
-  
- toString() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- toString() - Method in class org.apache.beam.sdk.extensions.ordered.combiner.SequenceRangeAccumulator
-  
- toString() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
-  
- toString() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Order
-  
- toString() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPCall
-  
- toString() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPFieldRef
-  
- toString() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPOperator
-  
- toString() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.CEPPattern
-  
- toString() - Method in class org.apache.beam.sdk.extensions.sql.impl.cep.Quantifier
-  
- toString() - Method in class org.apache.beam.sdk.extensions.sql.impl.LazyAggregateCombineFn
-  
- toString() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
-  
- toString() - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
-  
- toString() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.CalcRelSplitter.RelType
-  
- toString() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryFilter
-  
- toString() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableFilter
-  
- toString() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer
-  
- toString() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisRecord
-  
- toString() - Method in class org.apache.beam.sdk.io.cassandra.RingRange
-  
- toString() - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
-  
- toString() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-  
- toString() - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
-  
- toString() - Method in class org.apache.beam.sdk.io.FileBasedSource
-  
- toString() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
-  
- toString() - Method in interface org.apache.beam.sdk.io.fs.ResourceId
- 
Returns the string representation of this  ResourceId. 
- toString() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageApiInsertError
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableConfig
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.WriteWithResults
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.DetectNewPartitionsState
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.InitialPipelineState
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.NewPartition
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.PartitionRecord
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.model.StreamPartitionWithWatermark
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.ReadChangeStreamPartitionProgressTracker
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.StreamProgress
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.WriteSuccessSummary
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.WriteSuccessSummary
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.healthcare.FhirSearchParameter
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2Message
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2ReadResponse
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.ProjectPath
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.SubscriptionPath
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.TopicPath
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao.TransactionResult
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataTableNames
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChildPartition
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChildPartitionsRecord
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ColumnType
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.DataChangeRecord
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.HeartbeatRecord
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.Mod
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.TypeCode
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRange
-  
- toString() - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
-  
- toString() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
-  
- toString() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark
-  
- toString() - Method in enum org.apache.beam.sdk.io.kafka.KafkaTimestampType
-  
- toString() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-  
- toString() - Method in class org.apache.beam.sdk.io.range.ByteKey
-  
- toString() - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
-  
- toString() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-  
- toString() - Method in class org.apache.beam.sdk.io.range.OffsetRange
-  
- toString() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-  
- toString() - Method in class org.apache.beam.sdk.io.solace.broker.BrokerResponse
-  
- toString() - Method in class org.apache.beam.sdk.io.tika.ParseResult
-  
- toString() - Method in enum org.apache.beam.sdk.metrics.Lineage.Type
-  
- toString() - Method in class org.apache.beam.sdk.metrics.MetricKey
-  
- toString() - Method in class org.apache.beam.sdk.metrics.MetricName
-  
- toString() - Method in class org.apache.beam.sdk.metrics.MetricQueryResults
-  
- toString() - Method in class org.apache.beam.sdk.metrics.MetricResults
-  
- toString() - Method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
-  
- toString() - Method in class org.apache.beam.sdk.options.ValueProvider.RuntimeValueProvider
-  
- toString() - Method in class org.apache.beam.sdk.options.ValueProvider.StaticValueProvider
-  
- toString() - Method in class org.apache.beam.sdk.Pipeline
-  
- toString() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
-  
- toString(EnumerationType.Value) - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
-  
- toString() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
-  
- toString() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType.Value
-  
- toString() - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
-  
- toString() - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedString
-  
- toString() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
-  
- toString() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType.Value
-  
- toString() - Method in class org.apache.beam.sdk.schemas.logicaltypes.VariableBytes
-  
- toString() - Method in class org.apache.beam.sdk.schemas.logicaltypes.VariableString
-  
- toString() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
-  
- toString() - Method in class org.apache.beam.sdk.schemas.Schema.Options
-  
- toString() - Method in class org.apache.beam.sdk.schemas.Schema
-  
- toString() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
-  
- toString() - Method in class org.apache.beam.sdk.schemas.transforms.Cast.Narrowing
-  
- toString() - Method in class org.apache.beam.sdk.schemas.transforms.Cast.Widening
-  
- toString() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
-  
- toString() - Method in class org.apache.beam.sdk.testing.SuccessOrFailure
-  
- toString() - Method in class org.apache.beam.sdk.testing.TestPipeline
-  
- toString() - Method in class org.apache.beam.sdk.transforms.Combine.Holder
-  
- toString() - Method in class org.apache.beam.sdk.transforms.Contextful
-  
- toString() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
-  
- toString() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
-  
- toString() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
-  
- toString() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Path
-  
- toString() - Method in class org.apache.beam.sdk.transforms.display.DisplayData
-  
- toString() - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches
-  
- toString() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
-  
- toString() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
-  
- toString() - Method in class org.apache.beam.sdk.transforms.join.RawUnionValue
-  
- toString() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-  
- toString() - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
-  
- toString() - Method in class org.apache.beam.sdk.transforms.PeriodicSequence.OutputRangeTracker
-  
- toString() - Method in class org.apache.beam.sdk.transforms.PTransform
-  
- toString() - Method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
-  
- toString() - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
-  
- ToString - Class in org.apache.beam.sdk.transforms
- 
- toString() - Method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
-  
- toString(StateT) - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
- 
Creates a human-readable representation of the given state of this condition. 
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterAll
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterEach
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterFirst
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterSynchronizedProcessingTime
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.ReshuffleTrigger
-  
- toString() - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
-  
- toString() - Method in class org.apache.beam.sdk.values.KV
-  
- toString() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
-  
- toString() - Method in class org.apache.beam.sdk.values.PValueBase
-  
- toString() - Method in class org.apache.beam.sdk.values.Row
-  
- toString(boolean) - Method in class org.apache.beam.sdk.values.Row
- 
Convert Row to String. 
- toString() - Method in class org.apache.beam.sdk.values.ShardedKey
-  
- toString() - Method in class org.apache.beam.sdk.values.TimestampedValue
-  
- toString() - Method in class org.apache.beam.sdk.values.TupleTag
-  
- toString() - Method in class org.apache.beam.sdk.values.TupleTagList
-  
- toString() - Method in class org.apache.beam.sdk.values.TypeDescriptor
-  
- toString() - Method in class org.apache.beam.sdk.values.TypeParameter
-  
- toString() - Method in class org.apache.beam.sdk.values.ValueWithRecordId
-  
- toString() - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- toStringTimestamp(long) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.BeamCodegenUtils
-  
- toStringUTF8(byte[]) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.BeamCodegenUtils
-  
- toTableReference(String) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
-  
- toTableRow() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
- toTableRow(SerializableFunction<T, Row>) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
Convert a Beam schema type to a BigQuery  TableRow. 
- toTableRow(Row) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
Convert a Beam Row to a BigQuery TableRow. 
- toTableSchema(Schema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
- 
- toTableSpec(TableReference) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
- 
- toTableSpec(TableReference) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
-  
- toThreetenInstant(Instant) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.TimestampConverter
-  
- toTimestamp(Row) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.TimestampConvert
- 
ByteBuddy conversion for NanosInstant base type to Timestamp. 
- toTimestamp(BigDecimal) - Static method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampUtils
- 
Converts nanoseconds to their respective timestamp. 
- toTreeMap(ArrayData, ArrayData, DataType, DataType) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.Utils
-  
- toUnsplittableSource(BoundedSource<T>) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
- 
Returns an equivalent unsplittable BoundedSource<T>.
 
- toUri() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
-  
- toWrite(Schema, PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryFileLoadsSchemaTransformProvider.BigQueryFileLoadsSchemaTransform
-  
- toZetaSqlStructType(Schema) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlBeamTranslationUtils
-  
- toZetaSqlStructValue(Row, Schema) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlBeamTranslationUtils
-  
- toZetaSqlType(Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlBeamTranslationUtils
-  
- toZetaSqlType(RelDataType) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlCalciteTranslationUtils
-  
- toZetaSqlValue(Object, Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlBeamTranslationUtils
-  
- TrackerWithProgress - Class in org.apache.beam.sdk.io.gcp.pubsublite.internal
-  
- TrackerWithProgress() - Constructor for class org.apache.beam.sdk.io.gcp.pubsublite.internal.TrackerWithProgress
-  
- Transaction - Class in org.apache.beam.sdk.io.gcp.spanner
- 
A transaction object. 
- Transaction() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.Transaction
-  
- transactionId() - Method in class org.apache.beam.sdk.io.gcp.spanner.Transaction
-  
- TransactionResult(T, Timestamp) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.dao.PartitionMetadataDao.TransactionResult
-  
- transfer() - Method in class org.apache.beam.runners.portability.CloseableResource
- 
- transform(Function<T, V>) - Method in class org.apache.beam.sdk.metrics.MetricResult
-  
- TRANSFORM_URN - Static variable in class org.apache.beam.runners.spark.io.ConsoleIO.Write.Unbound
-  
- TRANSFORM_URN - Static variable in class org.apache.beam.runners.spark.io.CreateStream
-  
- transformContainer(Iterable<FromT>, Function<FromT, DestT>) - Static method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils
-  
- TransformExecutor - Interface in org.apache.beam.runners.direct
- 
A Runnablethat will execute aPTransformon some bundle of input.
 
- transformId() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.BagUserStateSpec
-  
- transformId() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.SideInputSpec
-  
- transformId() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.TimerSpec
-  
- TransformingMap(Map<K1, V1>, Function<K1, K2>, Function<V1, V2>) - Constructor for class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
-  
- TransformProvider<InputT extends PInput,OutputT extends POutput> - Interface in org.apache.beam.sdk.expansion.service
- 
Provides a mapping of  RunnerApi.FunctionSpec to a  PTransform, together with
 mappings of its inputs and outputs to maps of PCollections. 
- TransformServiceLauncher - Class in org.apache.beam.sdk.transformservice.launcher
- 
A utility that can be used to manage a Beam Transform Service. 
- transformTo(RelNode, Map<RelNode, RelNode>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
-  
- transformTo(RelNode, Map<RelNode, RelNode>, RelHintsPropagator) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
-  
- TransformTranslator<TransformT extends PTransform> - Interface in org.apache.beam.runners.dataflow
- 
- TransformTranslator<InT extends PInput,OutT extends POutput,TransformT extends PTransform<InT,OutT>> - Class in org.apache.beam.runners.spark.structuredstreaming.translation
- 
- TransformTranslator(float) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator
-  
- TransformTranslator.Context - Class in org.apache.beam.runners.spark.structuredstreaming.translation
- 
Available mutable context to translate a  PTransform. 
- TransformTranslator.StepTranslationContext - Interface in org.apache.beam.runners.dataflow
- 
- TransformTranslator.TranslationContext - Interface in org.apache.beam.runners.dataflow
- 
The interface provided to registered callbacks for interacting with the  DataflowRunner,
 including reading and writing the values of  PCollections and side inputs. 
- translate(Pipeline, RunnerApi.Pipeline, SdkComponents, DataflowRunner, List<DataflowPackage>) - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator
- 
Translates a  Pipeline into a  JobSpecification. 
- translate(AppliedPTransform<?, ?, PrimitiveParDoSingleFactory.ParDoSingle<?, ?>>, SdkComponents) - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.PayloadTranslator
-  
- translate(TransformT, TransformTranslator.TranslationContext) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator
-  
- translate(PipelineNode.PTransformNode, RunnerApi.Pipeline, FlinkBatchPortablePipelineTranslator.BatchTranslationContext) - Method in interface org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.PTransformTranslator
- 
Translate a PTransform into the given translation context. 
- translate(FlinkBatchPortablePipelineTranslator.BatchTranslationContext, RunnerApi.Pipeline) - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator
-  
- translate(T, RunnerApi.Pipeline) - Method in interface org.apache.beam.runners.flink.FlinkPortablePipelineTranslator
- 
Translates the given pipeline. 
- translate(String, RunnerApi.Pipeline, T) - Method in interface org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.PTransformTranslator
-  
- translate(FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext, RunnerApi.Pipeline) - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator
-  
- translate(TransformHierarchy.Node, TransformT) - Method in class org.apache.beam.runners.spark.SparkRunner.Evaluator
- 
Determine if this Node belongs to a Bounded branch of the pipeline, or Unbounded, and
 translate with the proper translator. 
- translate(Pipeline, SparkSession, SparkCommonPipelineOptions) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
- 
Translates a Beam pipeline into its Spark correspondence using the Spark SQL / Dataset API. 
- translate(TransformT, TransformTranslator<InT, OutT, TransformT>.Context) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator
-  
- translate(Pipeline) - Method in class org.apache.beam.runners.twister2.translators.Twister2PipelineTranslator
- 
Translates the pipeline by passing this class as a visitor. 
- translate(Pipeline) - Method in class org.apache.beam.runners.twister2.Twister2PipelineExecutionEnvironment
- 
translate the pipline into Twister2 TSet graph. 
- translate(AppliedPTransform<?, ?, T>, SdkComponents) - Method in class org.apache.beam.sdk.schemas.transforms.SchemaTransformTranslation.SchemaTransformPayloadTranslator
-  
- translateNode(Window.Assign<T>, Twister2BatchTranslationContext) - Method in class org.apache.beam.runners.twister2.translators.batch.AssignWindowTranslatorBatch
-  
- translateNode(Flatten.PCollections<T>, Twister2BatchTranslationContext) - Method in class org.apache.beam.runners.twister2.translators.batch.FlattenTranslatorBatch
-  
- translateNode(GroupByKey<K, V>, Twister2BatchTranslationContext) - Method in class org.apache.beam.runners.twister2.translators.batch.GroupByKeyTranslatorBatch
-  
- translateNode(Impulse, Twister2BatchTranslationContext) - Method in class org.apache.beam.runners.twister2.translators.batch.ImpulseTranslatorBatch
-  
- translateNode(ParDo.MultiOutput<InputT, OutputT>, Twister2BatchTranslationContext) - Method in class org.apache.beam.runners.twister2.translators.batch.ParDoMultiOutputTranslatorBatch
-  
- translateNode(View.CreatePCollectionView<ElemT, ViewT>, Twister2BatchTranslationContext) - Method in class org.apache.beam.runners.twister2.translators.batch.PCollectionViewTranslatorBatch
-  
- translateNode(SplittableParDo.PrimitiveBoundedRead<T>, Twister2BatchTranslationContext) - Method in class org.apache.beam.runners.twister2.translators.batch.ReadSourceTranslatorBatch
-  
- translateNode(TransformT, Twister2BatchTranslationContext) - Method in interface org.apache.beam.runners.twister2.translators.BatchTransformTranslator
-  
- translateNode(PTransform<PBegin, PCollection<T>>, Twister2StreamTranslationContext) - Method in class org.apache.beam.runners.twister2.translators.streaming.ReadSourceTranslatorStream
-  
- translateNode(TransformT, Twister2StreamTranslationContext) - Method in interface org.apache.beam.runners.twister2.translators.StreamTransformTranslator
-  
- TranslationUtils - Class in org.apache.beam.runners.twister2.utils
- 
doc. 
- translator - Variable in class org.apache.beam.runners.spark.SparkRunner.Evaluator
-  
- Transport - Class in org.apache.beam.sdk.extensions.gcp.util
- 
Helpers for cloud communication. 
- Transport() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.Transport
-  
- traverseTopologically(Pipeline.PipelineVisitor) - Method in class org.apache.beam.sdk.Pipeline
- 
For internal use only; no backwards-compatibility guarantees. 
- Trigger - Class in org.apache.beam.sdk.transforms.windowing
- 
Triggers control when the elements for a specific key and window are output. 
- Trigger(List<Trigger>) - Constructor for class org.apache.beam.sdk.transforms.windowing.Trigger
-  
- Trigger() - Constructor for class org.apache.beam.sdk.transforms.windowing.Trigger
-  
- Trigger.OnceTrigger - Class in org.apache.beam.sdk.transforms.windowing
- 
For internal use only; no backwards-compatibility guarantees. 
- triggering(Trigger) - Method in class org.apache.beam.sdk.transforms.windowing.Window
- 
Sets a non-default trigger for this WindowPTransform.
 
- trim(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- trim(String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.StringFunctions
-  
- TRIM - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- TRIM_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.impl.BeamBuiltinMethods
-  
- trivial() - Static method in class org.apache.beam.sdk.fn.stream.OutboundObserverFactory
- 
Creates an  OutboundObserverFactory that simply delegates to the base factory, with no
 flow control or synchronization. 
- trueLiteral() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
-  
- truncate(long) - Method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingSeekableByteChannel
-  
- TruncateResult() - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker.TruncateResult
-  
- trustManagers() - Method in class org.apache.beam.sdk.io.aws2.common.ClientBuilderFactory.SkipCertificateVerificationTrustManagerProvider
-  
- tryAcquireJobLock(Configuration) - Method in interface org.apache.beam.sdk.io.hadoop.format.ExternalSynchronization
- 
Tries to acquire lock for given job. 
- tryAcquireJobLock(Configuration) - Method in class org.apache.beam.sdk.io.hadoop.format.HDFSSynchronization
-  
- tryClaim(StreamProgress) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.ReadChangeStreamPartitionProgressTracker
- 
Claims a new StreamProgress to be processed. 
- tryClaim(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.DetectNewPartitionsRangeTracker
- 
Attempts to claim the given position. 
- tryClaim(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.ReadChangeStreamPartitionRangeTracker
- 
Attempts to claim the given position. 
- tryClaim(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRangeTracker
- 
Attempts to claim the given position. 
- tryClaim(Timestamp, PartitionMetadata) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRangeTracker
-  
- tryClaim(Long) - Method in class org.apache.beam.sdk.transforms.PeriodicSequence.OutputRangeTracker
-  
- tryClaim(ByteKey) - Method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
- 
Attempts to claim the given key. 
- tryClaim(Long) - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
- 
Attempts to claim the given offset. 
- tryClaim(PositionT) - Method in class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker
- 
Attempts to claim the block of work in the current restriction identified by the given
 position. 
- tryInterrupt(T) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.RestrictionInterrupter
- 
Returns true if the restriction tracker should be interrupted in claiming new positions. 
- tryProcess(int, Object) - Method in class org.apache.beam.runners.jet.processors.AssignWindowP
-  
- tryProcess(int, Object) - Method in class org.apache.beam.runners.jet.processors.FlattenP
-  
- tryProcess(int, Object) - Method in class org.apache.beam.runners.jet.processors.ViewP
-  
- tryProcess() - Method in class org.apache.beam.runners.jet.processors.WindowGroupP
-  
- tryProcess(int, Object) - Method in class org.apache.beam.runners.jet.processors.WindowGroupP
-  
- tryProcessWatermark(Watermark) - Method in class org.apache.beam.runners.jet.processors.StatefulParDoP
-  
- tryProcessWatermark(Watermark) - Method in class org.apache.beam.runners.jet.processors.WindowGroupP
-  
- tryReturnRecordAt(boolean, ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-  
- tryReturnRecordAt(boolean, Long) - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-  
- tryReturnRecordAt(boolean, long) - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-  
- tryReturnRecordAt(boolean, PositionT) - Method in interface org.apache.beam.sdk.io.range.RangeTracker
- 
Atomically determines whether a record at the given position can be returned and updates
 internal state. 
- trySplit(double) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.DetectNewPartitionsTracker
-  
- trySplit(double) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.restriction.ReadChangeStreamPartitionProgressTracker
- 
Splits the work that's left. 
- trySplit(double) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.ReadChangeStreamPartitionRangeTracker
- 
- trySplit(double) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.TimestampRangeTracker
- 
Splits the restriction through the following algorithm: 
- trySplit(double) - Method in class org.apache.beam.sdk.transforms.PeriodicSequence.OutputRangeTracker
-  
- trySplit(double) - Method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
-  
- trySplit(double) - Method in class org.apache.beam.sdk.transforms.splittabledofn.GrowableOffsetRangeTracker
-  
- trySplit(double) - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
-  
- trySplit(double) - Method in class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker
- 
Splits current restriction based on fractionOfRemainder.
 
- trySplitAtPosition(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
-  
- trySplitAtPosition(Long) - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-  
- trySplitAtPosition(long) - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
-  
- trySplitAtPosition(PositionT) - Method in interface org.apache.beam.sdk.io.range.RangeTracker
- 
- tuple(T1, T2) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.utils.ScalaInterop
-  
- TUPLE - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- tuple(Map<String, TableSchema.ColumnType>) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- TUPLE_TAGS - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- tupleEncoder(Encoder<T1>, Encoder<T2>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- TupleTag<V> - Class in org.apache.beam.sdk.values
- 
- TupleTag() - Constructor for class org.apache.beam.sdk.values.TupleTag
- 
Constructs a new TupleTag, with a fresh unique id.
 
- TupleTag(String) - Constructor for class org.apache.beam.sdk.values.TupleTag
- 
Constructs a new TupleTagwith the given id.
 
- TupleTagList - Class in org.apache.beam.sdk.values
- 
- tupleTypes() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- TVFSlidingWindowFn - Class in org.apache.beam.sdk.extensions.sql.impl
- 
TVFSlidingWindowFn assigns window based on input row's "window_start" and "window_end"
 timestamps. 
- TVFSlidingWindowFn() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.TVFSlidingWindowFn
-  
- TVFStreamingUtils - Class in org.apache.beam.sdk.extensions.sql.impl.utils
- 
Provides static constants or utils for TVF streaming. 
- TVFStreamingUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.TVFStreamingUtils
-  
- Twister2AssignContext<T,W extends BoundedWindow> - Class in org.apache.beam.runners.twister2.utils
- 
doc. 
- Twister2AssignContext(WindowFn<T, W>, WindowedValue<T>) - Constructor for class org.apache.beam.runners.twister2.utils.Twister2AssignContext
-  
- Twister2BatchPipelineTranslator - Class in org.apache.beam.runners.twister2.translators
- 
Twister pipeline translator for batch pipelines. 
- Twister2BatchPipelineTranslator(Twister2PipelineOptions, Twister2BatchTranslationContext) - Constructor for class org.apache.beam.runners.twister2.translators.Twister2BatchPipelineTranslator
-  
- Twister2BatchTranslationContext - Class in org.apache.beam.runners.twister2
- 
Twister2BatchTranslationContext. 
- Twister2BatchTranslationContext(Twister2PipelineOptions) - Constructor for class org.apache.beam.runners.twister2.Twister2BatchTranslationContext
-  
- Twister2BoundedSource<T> - Class in org.apache.beam.runners.twister2.translation.wrappers
- 
Twister2 wrapper for Bounded Source. 
- Twister2BoundedSource(BoundedSource<T>, Twister2TranslationContext, PipelineOptions) - Constructor for class org.apache.beam.runners.twister2.translation.wrappers.Twister2BoundedSource
-  
- Twister2EmptySource<T> - Class in org.apache.beam.runners.twister2.translation.wrappers
- 
Empty Source wrapper. 
- Twister2EmptySource() - Constructor for class org.apache.beam.runners.twister2.translation.wrappers.Twister2EmptySource
-  
- Twister2PipelineExecutionEnvironment - Class in org.apache.beam.runners.twister2
- 
Twister2PipelineExecutionEnvironment. 
- Twister2PipelineExecutionEnvironment(Twister2PipelineOptions) - Constructor for class org.apache.beam.runners.twister2.Twister2PipelineExecutionEnvironment
-  
- Twister2PipelineOptions - Interface in org.apache.beam.runners.twister2
- 
Twister2PipelineOptions. 
- Twister2PipelineResult - Class in org.apache.beam.runners.twister2
- 
Represents a Twister2 pipeline execution result. 
- Twister2PipelineResult(Twister2JobState) - Constructor for class org.apache.beam.runners.twister2.Twister2PipelineResult
-  
- Twister2PipelineTranslator - Class in org.apache.beam.runners.twister2.translators
- 
Twister2PipelineTranslator, both batch and streaming translators need to extend from this. 
- Twister2PipelineTranslator() - Constructor for class org.apache.beam.runners.twister2.translators.Twister2PipelineTranslator
-  
- Twister2Runner - Class in org.apache.beam.runners.twister2
- 
A  PipelineRunner that executes the operations in the pipeline by first translating them
 to a Twister2 Plan and then executing them either locally or on a Twister2 cluster, depending on
 the configuration. 
- Twister2Runner(Twister2PipelineOptions) - Constructor for class org.apache.beam.runners.twister2.Twister2Runner
-  
- Twister2RunnerRegistrar - Class in org.apache.beam.runners.twister2
- 
AutoService registrar - will register Twister2Runner and Twister2Options as possible pipeline
 runner services. 
- Twister2RunnerRegistrar.Options - Class in org.apache.beam.runners.twister2
- 
Pipeline options registrar. 
- Twister2RunnerRegistrar.Runner - Class in org.apache.beam.runners.twister2
- 
Pipeline runner registrar. 
- Twister2SideInputReader - Class in org.apache.beam.runners.twister2.utils
-  
- Twister2SideInputReader(Map<TupleTag<?>, WindowingStrategy<?, ?>>, TSetContext) - Constructor for class org.apache.beam.runners.twister2.utils.Twister2SideInputReader
-  
- Twister2SinkFunction<T> - Class in org.apache.beam.runners.twister2.translators.functions
- 
Sink Function that collects results. 
- Twister2SinkFunction() - Constructor for class org.apache.beam.runners.twister2.translators.functions.Twister2SinkFunction
-  
- Twister2StreamPipelineTranslator - Class in org.apache.beam.runners.twister2.translators
- 
Twister pipeline translator for stream pipelines. 
- Twister2StreamPipelineTranslator() - Constructor for class org.apache.beam.runners.twister2.translators.Twister2StreamPipelineTranslator
-  
- Twister2StreamTranslationContext - Class in org.apache.beam.runners.twister2
- 
Twister2StreamingTranslationContext. 
- Twister2StreamTranslationContext(Twister2PipelineOptions) - Constructor for class org.apache.beam.runners.twister2.Twister2StreamTranslationContext
-  
- Twister2TestRunner - Class in org.apache.beam.runners.twister2
- 
A  PipelineRunner that executes the operations in the pipeline by first translating them
 to a Twister2 Plan and then executing them either locally or on a Twister2 cluster, depending on
 the configuration. 
- Twister2TestRunner(Twister2PipelineOptions) - Constructor for class org.apache.beam.runners.twister2.Twister2TestRunner
-  
- Twister2TranslationContext - Class in org.apache.beam.runners.twister2
- 
Twister2TranslationContext. 
- Twister2TranslationContext(Twister2PipelineOptions) - Constructor for class org.apache.beam.runners.twister2.Twister2TranslationContext
-  
- type - Variable in class org.apache.beam.runners.dataflow.util.OutputReference
-  
- type - Variable in class org.apache.beam.sdk.extensions.avro.io.AvroDatumFactory
-  
- type - Variable in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.TypeWithNullability
-  
- type(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.Table.Builder
-  
- TypeCode - Class in org.apache.beam.sdk.io.gcp.spanner.changestreams.model
- 
Represents a type of a column within Cloud Spanner. 
- TypeCode(String) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.TypeCode
- 
Constructs a type code from the given String code. 
- TypeConversion() - Constructor for class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
-  
- TypedCombineFnDelegate<InputT,AccumT,OutputT> - Class in org.apache.beam.sdk.extensions.sql
- 
A Combine.CombineFndelegating all relevant calls to given delegate.
 
- TypedCombineFnDelegate(Combine.CombineFn<InputT, AccumT, OutputT>) - Constructor for class org.apache.beam.sdk.extensions.sql.TypedCombineFnDelegate
-  
- typedef(String, Schema.FieldType) - Method in class org.apache.beam.sdk.io.thrift.ThriftSchema.Customizer
-  
- TypeDescriptor<T> - Class in org.apache.beam.sdk.values
- 
A description of a Java type, including actual generic parameters where possible. 
- TypeDescriptor() - Constructor for class org.apache.beam.sdk.values.TypeDescriptor
- 
- TypeDescriptor(Object) - Constructor for class org.apache.beam.sdk.values.TypeDescriptor
- 
Creates a  TypeDescriptor representing the type parameter  T, which should
 resolve to a concrete type in the context of the class  clazz. 
- TypeDescriptor(Class<?>) - Constructor for class org.apache.beam.sdk.values.TypeDescriptor
- 
Creates a  TypeDescriptor representing the type parameter  T, which should
 resolve to a concrete type in the context of the class  clazz. 
- TypeDescriptors - Class in org.apache.beam.sdk.values
- 
A utility class for creating  TypeDescriptor objects for different types, such as Java
 primitive types, containers and  KVs of other  TypeDescriptor objects, and
 extracting type variables of parameterized types (e.g. 
- TypeDescriptors() - Constructor for class org.apache.beam.sdk.values.TypeDescriptors
-  
- TypeDescriptors.TypeVariableExtractor<InputT,OutputT> - Interface in org.apache.beam.sdk.values
- 
- TypeDescriptorWithSchema() - Constructor for class org.apache.beam.sdk.schemas.utils.ReflectUtils.TypeDescriptorWithSchema
-  
- TypedRead() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-  
- TypedSchemaTransformProvider<ConfigT> - Class in org.apache.beam.sdk.schemas.transforms
- 
- TypedSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.schemas.transforms.TypedSchemaTransformProvider
-  
- TypedWrite() - Constructor for class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
-  
- TypedWrite() - Constructor for class org.apache.beam.sdk.io.TextIO.TypedWrite
-  
- typeName() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- TypeParameter<T> - Class in org.apache.beam.sdk.values
- 
- TypeParameter() - Constructor for class org.apache.beam.sdk.values.TypeParameter
-  
- typesEqual(Schema.Field) - Method in class org.apache.beam.sdk.schemas.Schema.Field
- 
Returns true if two fields are equal, ignoring name and description. 
- typesEqual(Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
Returns true if two FieldTypes are equal. 
- typesEqual(Schema) - Method in class org.apache.beam.sdk.schemas.Schema
- 
Returns true if two schemas are equal ignoring field names and descriptions. 
- typeToProtoType(String) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProto
-  
- TZTimeOnly() - Constructor for class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.TZTimeOnly
-  
- TZTimestamp() - Constructor for class org.apache.beam.sdk.extensions.sbe.SbeLogicalTypes.TZTimestamp
-  
- v1() - Static method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreIO
- 
Returns a  DatastoreV1 that provides an API for accessing Cloud Datastore through v1
 version of Datastore Client library. 
- v1() - Static method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreIO
-  
- v17() - Static method in class org.apache.beam.sdk.io.googleads.GoogleAdsIO
-  
- V1_READ_OVERRIDE - Static variable in class org.apache.beam.sdk.io.gcp.pubsublite.internal.SubscribeTransform
-  
- VALID_DATA_FORMATS - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformProvider
-  
- VALID_DATA_FORMATS - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformProvider
-  
- VALID_DATA_FORMATS - Static variable in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider
-  
- VALID_DATA_FORMATS - Static variable in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
-  
- VALID_FIELD_TYPE_SET - Static variable in class org.apache.beam.sdk.io.csv.CsvIO
- 
The valid  Schema.FieldType from which  CsvIO converts CSV records to the fields. 
- VALID_FORMATS_STR - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.PubsubReadSchemaTransformProvider
-  
- VALID_FORMATS_STR - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteSchemaTransformProvider
-  
- VALID_FORMATS_STR - Static variable in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider
-  
- VALID_FORMATS_STR - Static variable in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
-  
- VALID_PROVIDERS - Static variable in class org.apache.beam.sdk.io.fileschematransform.FileReadSchemaTransformConfiguration
-  
- VALID_START_OFFSET_VALUES - Static variable in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
-  
- validate() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
-  
- validate() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
-  
- validate(Schema) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.LocalTimestampMicros
-  
- validate(Schema) - Method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroJavaTimeConversions.LocalTimestampMillis
-  
- validate(AwsOptions, ClientConfiguration) - Static method in interface org.apache.beam.sdk.io.aws2.common.ClientBuilderFactory
- 
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
-  
- validate() - Method in class org.apache.beam.sdk.io.CompressedSource
- 
Validates that the delegate source is a valid source and that the channel factory is not null. 
- validate() - Method in class org.apache.beam.sdk.io.csv.providers.CsvWriteTransformProvider.CsvWriteConfiguration
-  
- validate() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSink
-  
- validate() - Method in class org.apache.beam.sdk.io.FileBasedSource
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-  
- validate() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProvider.BigQueryDirectReadSchemaTransformConfiguration
-  
- validate() - Method in class org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryWriteConfiguration
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.WriteWithResults
-  
- validate() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProvider.BigtableReadSchemaTransformConfiguration
- 
- validate() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProvider.BigtableWriteSchemaTransformConfiguration
- 
Validates the configuration object. 
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
-  
- validate() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteReadSchemaTransformProvider.PubsubLiteReadSchemaTransformConfiguration
-  
- validate() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteWriteSchemaTransformProvider.PubsubLiteWriteSchemaTransformConfiguration
-  
- validate() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- validate() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerReadSchemaTransformProvider.SpannerReadSchemaTransformConfiguration
-  
- validate() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteSchemaTransformProvider.SpannerWriteSchemaTransformConfiguration
-  
- validate() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write
-  
- validate() - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- validate(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcReadSchemaTransformProvider.JdbcReadSchemaTransformConfiguration
-  
- validate() - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration
-  
- validate(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransformConfiguration
-  
- validate() - Method in class org.apache.beam.sdk.io.json.providers.JsonWriteTransformProvider.JsonWriteConfiguration
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- validate() - Method in class org.apache.beam.sdk.io.kafka.KafkaReadSchemaTransformConfiguration
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Read
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Write
-  
- validate(T) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-  
- validate() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Read
-  
- validate() - Method in class org.apache.beam.sdk.io.Source
- 
Checks that this source is valid, before it can be used in a pipeline. 
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.WriteFiles
-  
- validate(Class<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.options.PipelineOptionsValidator
- 
Validates that the passed  PipelineOptions conforms to all the validation criteria from
 the passed in interface. 
- validate() - Method in class org.apache.beam.sdk.providers.GenerateSequenceSchemaTransformProvider.GenerateSequenceConfiguration
-  
- validate() - Method in class org.apache.beam.sdk.schemas.transforms.providers.JavaRowUdf.Configuration
-  
- validate(PipelineOptions, Map<TupleTag<?>, PCollection<?>>, Map<TupleTag<?>, PCollection<?>>) - Method in class org.apache.beam.sdk.transforms.GroupByKey
-  
- validate(PipelineOptions) - Method in class org.apache.beam.sdk.transforms.PTransform
- 
Called before running the Pipeline to verify this transform is fully and correctly specified. 
- validate(PipelineOptions, Map<TupleTag<?>, PCollection<?>>, Map<TupleTag<?>, PCollection<?>>) - Method in class org.apache.beam.sdk.transforms.PTransform
- 
Called before running the Pipeline to verify this transform, its inputs, and outputs are fully
 and correctly specified. 
- VALIDATE_TIME_INTERVAL - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- VALIDATE_TIMESTAMP - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.translation.SqlOperators
-  
- validateCli(Class<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.options.PipelineOptionsValidator
- 
Validates that the passed  PipelineOptions from command line interface (CLI) conforms to
 all the validation criteria from the passed in interface. 
- validateGetOutputTimestamps(WindowFn<T, W>, TimestampCombiner, List<List<Long>>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
- 
Verifies that later-ending merged windows from any of the timestamps hold up output of
 earlier-ending windows, using the provided  WindowFn and  TimestampCombiner. 
- validateGetOutputTimestampsWithValue(WindowFn<T, W>, TimestampCombiner, List<List<TimestampedValue<T>>>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
- 
Verifies that later-ending merged windows from any of the timestampValues hold up output of
 earlier-ending windows, using the provided  WindowFn and  TimestampCombiner. 
- validateInputFilePatternSupported(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
- 
Validates the input GCS path is accessible and that the path is well formed. 
- validateInputFilePatternSupported(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
-  
- validateInputFilePatternSupported(String) - Method in interface org.apache.beam.sdk.extensions.gcp.storage.PathValidator
- 
Validate that a file pattern is conforming. 
- validateJavaBean(List<FieldValueTypeInformation>, List<FieldValueTypeInformation>, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
-  
- validateMaterializations(Iterable<PCollectionView<?>>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.SparkSideInputReader
-  
- validateMethod(Method) - Static method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
-  
- validateOutputFilePrefixSupported(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
- 
Validates the output GCS path is accessible and that the path is well formed. 
- validateOutputFilePrefixSupported(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
-  
- validateOutputFilePrefixSupported(String) - Method in interface org.apache.beam.sdk.extensions.gcp.storage.PathValidator
- 
Validate that an output file prefix is conforming. 
- validateOutputResourceSupported(ResourceId) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
-  
- validateOutputResourceSupported(ResourceId) - Method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
-  
- validateOutputResourceSupported(ResourceId) - Method in interface org.apache.beam.sdk.extensions.gcp.storage.PathValidator
- 
Validates that an output path is conforming. 
- ValidatesRunner - Interface in org.apache.beam.sdk.testing
- 
Category tag for tests which validate that a Beam runner is correctly implemented. 
- validateTimeInterval(Long, TimeUnit) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
- 
This function validates that interval is compatible with ZetaSQL timestamp values range. 
- validateTimestamp(Long) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
- 
This function validates that Long representation of timestamp is compatible with ZetaSQL
 timestamp values range. 
- validateTimestampBounds(Instant) - Static method in class org.apache.beam.sdk.transforms.windowing.BoundedWindow
- 
Validates that a given timestamp is within min and max bounds. 
- validateTransform() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
- 
Validates construction of this transform. 
- validateTransform() - Method in class org.apache.beam.sdk.io.sparkreceiver.SparkReceiverIO.Read
-  
- Validation - Annotation Type in org.apache.beam.sdk.options
- 
- Validation.Required - Annotation Type in org.apache.beam.sdk.options
- 
This criteria specifies that the value must be not null. 
- validator() - Method in class org.apache.beam.sdk.schemas.transforms.Cast
-  
- VALUE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- value() - Method in class org.apache.beam.runners.spark.metrics.MetricsContainerStepMapAccumulator
-  
- value() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsAccumulator
-  
- value(Object) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedList
-  
- value(Object) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedMap
-  
- VALUE - Static variable in class org.apache.beam.sdk.io.contextualtextio.RecordWithMetadata
-  
- VALUE - Static variable in class org.apache.beam.sdk.io.gcp.bigtable.RowUtils
-  
- value() - Method in class org.apache.beam.sdk.io.gcp.pubsublite.internal.Uuid
-  
- Value(int) - Constructor for class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType.Value
-  
- Value(EnumerationType.Value, Object) - Constructor for class org.apache.beam.sdk.schemas.logicaltypes.OneOfType.Value
-  
- value() - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
Create a  StateSpec for a single value of type  T. 
- value(Coder<T>) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
- ValueCaptureType - Enum in org.apache.beam.sdk.io.gcp.spanner.changestreams.model
- 
Represents the capture type of a change stream. 
- valueCoder() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.BagUserStateSpec
-  
- valueCoder() - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.MultimapSideInputHandler
- 
Returns the  Coder to use for the elements of the resulting values iterable. 
- valueEncoderOf(KvCoder<K, V>) - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderProvider
-  
- ValueInSingleWindow<T> - Class in org.apache.beam.sdk.values
- 
An immutable tuple of value, timestamp, window, and pane. 
- ValueInSingleWindow() - Constructor for class org.apache.beam.sdk.values.ValueInSingleWindow
-  
- ValueInSingleWindow.Coder<T> - Class in org.apache.beam.sdk.values
- 
- valueOf(String) - Static method in enum org.apache.beam.io.debezium.Connectors
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.runners.dataflow.options.DataflowPipelineOptions.FlexResourceSchedulingGoal
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.Level
- 
Deprecated. Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.runners.local.ExecutionDriver.DriverState
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.coders.CannotProvideCoderException.ReasonCode
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.ordered.UnprocessedEvent.Reason
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sbe.UnsignedOptions.Behavior
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options.SorterType
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sql.impl.cep.CEPKind
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters.Kind
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamSetOperatorRelBase.OpType
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sql.meta.ProjectSupport
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider.PushDownOptions
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.cassandra.CassandraIO.MutationType
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.cdap.PluginConstants.Format
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.cdap.PluginConstants.FormatProvider
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.cdap.PluginConstants.Hadoop
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.cdap.PluginConstants.PluginType
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.clickhouse.TableSchema.DefaultType
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.clickhouse.TableSchema.TypeName
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.CompressedSource.CompressionMode
- 
Deprecated. Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.Compression
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
- 
Deprecated. Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.FileBasedSource.Mode
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.FileIO.ReadMatches.DirectoryTreatment
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.FileSystem.LineageLevel
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.fs.EmptyMatchTreatment
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.fs.MatchResult.Status
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.fs.MoveOptions.StandardMoveOptions
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.fs.ResolveOptions.StandardResolveOptions
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead.Method
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead.QueryPriority
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.CreateDisposition
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.Method
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.SchemaUpdateOption
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.WriteDisposition
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService.TableMetadataView
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQuerySinkMetrics.RpcMethod
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.ConversionOptions.TruncateTimestamps
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.RowMutationInformation.MutationType
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ExistingPipelineOptions
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Import.ContentStructure
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write.WriteMethod
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.Write.WriteMethod
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient.FhirResourcePagesIterator.FhirMethod
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ModType
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata.State
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ValueCaptureType
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.spanner.SpannerIO.FailureMode
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.iceberg.IcebergScanConfig.ScanType
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.kafka.KafkaTimestampType
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.redis.RedisIO.Write.Method
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.snowflake.enums.CreateDisposition
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.snowflake.enums.StreamingLogLevel
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.snowflake.enums.WriteDisposition
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.solace.data.Solace.DestinationType
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.solace.SolaceIO.SubmissionMode
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.solace.SolaceIO.WriterType
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.TextIO.CompressionType
- 
Deprecated. Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.TFRecordIO.CompressionType
- 
Deprecated. Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.io.xml.XmlIO.Read.CompressionType
- 
Deprecated. Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.jmh.schemas.RowBundle.Action
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.metrics.Lineage.Type
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.options.PipelineOptions.CheckEnabled
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.options.SdkHarnessOptions.LogLevel
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.Pipeline.PipelineVisitor.CompositeBehavior
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.PipelineResult.State
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.ListQualifier
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.MapQualifier
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier.Kind
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
- 
- valueOf(int) - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
- 
- valueOf(String) - Static method in enum org.apache.beam.sdk.schemas.Schema.EquivalenceNullablePolicy
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.schemas.Schema.TypeName
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields.Fanout.Kind
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.state.TimeDomain
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.testing.SourceTestUtils.ExpectedSplitOutcome
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.testing.TestStream.EventType
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.display.DisplayData.Type
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.DoFnTester.CloningBehavior
- 
Deprecated. Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker.IsBounded
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.windowing.PaneInfo.Timing
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.windowing.Window.ClosingBehavior
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.windowing.Window.OnTimeBehavior
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.values.PCollection.IsBounded
- 
Returns the enum constant of this type with the specified name. 
- valueOf(String) - Static method in enum org.apache.beam.sdk.values.WindowingStrategy.AccumulationMode
- 
Returns the enum constant of this type with the specified name. 
- ValueOrMetadata(boolean, T, MetaT) - Constructor for class org.apache.beam.sdk.values.PCollectionViews.ValueOrMetadata
-  
- ValueProvider<T> - Interface in org.apache.beam.sdk.options
- 
A  ValueProvider abstracts the notion of fetching a value that may or may not be currently
 available. 
- ValueProvider.Deserializer - Class in org.apache.beam.sdk.options
- 
For internal use only; no backwards compatibility guarantees. 
- ValueProvider.NestedValueProvider<T,X> - Class in org.apache.beam.sdk.options
- 
- ValueProvider.RuntimeValueProvider<T> - Class in org.apache.beam.sdk.options
- 
- ValueProvider.Serializer - Class in org.apache.beam.sdk.options
- 
For internal use only; no backwards compatibility guarantees. 
- ValueProvider.StaticValueProvider<T> - Class in org.apache.beam.sdk.options
- 
- ValueProviders - Class in org.apache.beam.sdk.options
- 
- values() - Static method in enum org.apache.beam.io.debezium.Connectors
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.runners.dataflow.options.DataflowPipelineOptions.FlexResourceSchedulingGoal
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.Level
- 
Deprecated. Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.runners.local.ExecutionDriver.DriverState
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.coders.CannotProvideCoderException.ReasonCode
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.extensions.ordered.UnprocessedEvent.Reason
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.extensions.sbe.UnsignedOptions.Behavior
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options.SorterType
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.extensions.sql.impl.cep.CEPKind
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters.Kind
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamSetOperatorRelBase.OpType
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.extensions.sql.meta.ProjectSupport
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider.PushDownOptions
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.cassandra.CassandraIO.MutationType
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.cdap.PluginConstants.Format
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.cdap.PluginConstants.FormatProvider
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.cdap.PluginConstants.Hadoop
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.cdap.PluginConstants.PluginType
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.clickhouse.TableSchema.DefaultType
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.clickhouse.TableSchema.TypeName
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.CompressedSource.CompressionMode
- 
Deprecated. Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.Compression
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
- 
Deprecated. Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.FileBasedSource.Mode
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.FileIO.ReadMatches.DirectoryTreatment
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.FileSystem.LineageLevel
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.fs.EmptyMatchTreatment
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.fs.MatchResult.Status
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.fs.MoveOptions.StandardMoveOptions
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.fs.ResolveOptions.StandardResolveOptions
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead.Method
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead.QueryPriority
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.CreateDisposition
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.Method
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.SchemaUpdateOption
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.WriteDisposition
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService.TableMetadataView
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQuerySinkMetrics.RpcMethod
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.ConversionOptions.TruncateTimestamps
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.RowMutationInformation.MutationType
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ExistingPipelineOptions
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Import.ContentStructure
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write.WriteMethod
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.Write.WriteMethod
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient.FhirResourcePagesIterator.FhirMethod
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ModType
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.spanner.changestreams.model.PartitionMetadata.State
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ValueCaptureType
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.gcp.spanner.SpannerIO.FailureMode
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.iceberg.IcebergScanConfig.ScanType
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
- 
Writes just the values to Kafka. 
- values() - Static method in enum org.apache.beam.sdk.io.kafka.KafkaTimestampType
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.redis.RedisIO.Write.Method
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.snowflake.enums.CreateDisposition
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.snowflake.enums.StreamingLogLevel
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.snowflake.enums.WriteDisposition
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.solace.data.Solace.DestinationType
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.solace.SolaceIO.SubmissionMode
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.solace.SolaceIO.WriterType
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.TextIO.CompressionType
- 
Deprecated. Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.TFRecordIO.CompressionType
- 
Deprecated. Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.io.xml.XmlIO.Read.CompressionType
- 
Deprecated. Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.jmh.schemas.RowBundle.Action
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.metrics.Lineage.Type
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.options.PipelineOptions.CheckEnabled
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.options.SdkHarnessOptions.LogLevel
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.Pipeline.PipelineVisitor.CompositeBehavior
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.PipelineResult.State
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.ListQualifier
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.MapQualifier
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier.Kind
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.schemas.Schema.EquivalenceNullablePolicy
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.schemas.Schema.TypeName
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields.Fanout.Kind
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
-  
- values() - Method in interface org.apache.beam.sdk.state.MapState
- 
Returns an Iterableover the values contained in this map.
 
- values() - Static method in enum org.apache.beam.sdk.state.TimeDomain
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.testing.SourceTestUtils.ExpectedSplitOutcome
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.testing.TestStream.EventType
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in class org.apache.beam.sdk.transforms.Deduplicate
- 
Returns a deduplication transform that deduplicates values for up to 10 mins within the  processing time domain. 
- values() - Static method in enum org.apache.beam.sdk.transforms.display.DisplayData.Type
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.transforms.DoFnTester.CloningBehavior
- 
Deprecated. Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker.IsBounded
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- Values<V> - Class in org.apache.beam.sdk.transforms
- 
Values<V>takes aPCollectionofKV<K, V>s and returns aPCollection<V>of the values.
 
- values() - Static method in enum org.apache.beam.sdk.transforms.windowing.PaneInfo.Timing
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.transforms.windowing.Window.ClosingBehavior
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.transforms.windowing.Window.OnTimeBehavior
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.values.PCollection.IsBounded
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- values() - Static method in enum org.apache.beam.sdk.values.WindowingStrategy.AccumulationMode
- 
Returns an array containing the constants of this enum type, in
the order they are declared. 
- ValueState<T> - Interface in org.apache.beam.sdk.state
- 
- ValueWithRecordId<ValueT> - Class in org.apache.beam.sdk.values
- 
For internal use only; no backwards compatibility guarantees. 
- ValueWithRecordId(ValueT, byte[]) - Constructor for class org.apache.beam.sdk.values.ValueWithRecordId
-  
- ValueWithRecordId.StripIdsDoFn<T> - Class in org.apache.beam.sdk.values
- 
DoFn to turn a  ValueWithRecordId<T> back to the value  T. 
- ValueWithRecordId.ValueWithRecordIdCoder<ValueT> - Class in org.apache.beam.sdk.values
- 
A  Coder for  ValueWithRecordId, using a wrapped value  Coder. 
- ValueWithRecordIdCoder(Coder<ValueT>) - Constructor for class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
-  
- VARBINARY - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- VARCHAR - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
-  
- VariableBytes - Class in org.apache.beam.sdk.schemas.logicaltypes
- 
A LogicalType representing a variable-length byte array with specified maximum length. 
- VariableString - Class in org.apache.beam.sdk.schemas.logicaltypes
- 
A LogicalType representing a variable-length string with specified maximum length. 
- VarianceFn<T extends java.lang.Number> - Class in org.apache.beam.sdk.extensions.sql.impl.transform.agg
- 
Combine.CombineFnfor Variance onNumbertypes.
 
- VarIntBenchmark - Class in org.apache.beam.sdk.jmh.util
- 
Benchmarks for VarIntand variants.
 
- VarIntBenchmark() - Constructor for class org.apache.beam.sdk.jmh.util.VarIntBenchmark
-  
- VarIntBenchmark.BlackholeOutput - Class in org.apache.beam.sdk.jmh.util
- 
Output to Blackhole.
 
- VarIntBenchmark.Bytes - Class in org.apache.beam.sdk.jmh.util
- 
Input from randomly generated bytes. 
- VarIntBenchmark.ByteStringOutput - Class in org.apache.beam.sdk.jmh.util
- 
Output to ByteStringOutputStream.
 
- VarIntBenchmark.Longs - Class in org.apache.beam.sdk.jmh.util
- 
Input from randomly generated longs. 
- VarIntCoder - Class in org.apache.beam.sdk.coders
- 
A  Coder that encodes  Integers using between 1 and 5 bytes. 
- VarLongCoder - Class in org.apache.beam.sdk.coders
- 
A  Coder that encodes  Longs using between 1 and 10 bytes. 
- verifyBucketAccessible(GcsPath) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
- 
Checks whether the GCS bucket exists. 
- verifyCompatibility(Schema) - Method in class org.apache.beam.sdk.schemas.transforms.Cast
-  
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-  
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-  
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-  
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-  
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-  
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-  
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-  
- verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
- 
- verifyDeterministic() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.runners.fnexecution.wire.ByteStringCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.AtomicCoder
- 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BigEndianShortCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BitSetCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.ByteCoder
- 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.Coder
- 
- verifyDeterministic(Coder<?>, String, Iterable<Coder<?>>) - Static method in class org.apache.beam.sdk.coders.Coder
- 
Verifies all of the provided coders are deterministic. 
- verifyDeterministic(Coder<?>, String, Coder<?>...) - Static method in class org.apache.beam.sdk.coders.Coder
- 
Verifies all of the provided coders are deterministic. 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.CustomCoder
- 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.DelegateCoder
- 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.DequeCoder
- 
Deque sizes are always known, so DequeIterable may be deterministic while the general
 IterableLikeCoder is not. 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.DoubleCoder
- 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.DurationCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.FloatCoder
- 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.InstantCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
- 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.KvCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
- 
LengthPrefixCoderis deterministic if the nestedCoderis.
 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.ListCoder
- 
List sizes are always known, so ListIterable may be deterministic while the general
 IterableLikeCoder is not. 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.MapCoder
- 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.NullableCoder
- 
NullableCoderis deterministic if the nestedCoderis.
 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.SerializableCoder
- 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.SetCoder
- 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.ShardedKeyCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.SnappyCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.SortedMapCoder
- 
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.TimestampPrefixingWindowCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.VarIntCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.VarLongCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.VoidCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.coders.ZstdCoder
- 
- verifyDeterministic() - Method in class org.apache.beam.sdk.extensions.avro.coders.AvroCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertErrorCoder
- 
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV2
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV3
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
- 
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.hadoop.WritableCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.kafka.TopicPartitionCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.ReadableFileCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.io.splunk.SplunkEventCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.testing.TestStream.TestStreamCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo.PaneInfoCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.values.FailsafeValueInSingleWindow.Coder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.values.PCollectionViews.ValueOrMetadataCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
-  
- verifyDeterministic() - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
-  
- verifyFieldValue(Object, Schema.FieldType, String) - Static method in class org.apache.beam.sdk.values.SchemaVerification
-  
- verifyPAssertsSucceeded(Pipeline, PipelineResult) - Static method in class org.apache.beam.sdk.testing.TestPipeline
- 
Verifies all { PAsserts} in the pipeline have been executed and were successful. 
- verifyPath(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
-  
- verifyPath(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
-  
- verifyPath(String) - Method in interface org.apache.beam.sdk.extensions.gcp.storage.PathValidator
- 
Validate that a path is a valid path and that the path is accessible. 
- VersionDependentFlinkPipelineOptions - Interface in org.apache.beam.runners.flink
-  
- via(ProcessFunction<T, Long>) - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.Globally
-  
- via(ProcessFunction<KV<K, V>, KV<K, Long>>) - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.PerKey
-  
- via(Contextful<Contextful.Fn<UserT, OutputT>>, Contextful<Contextful.Fn<DestinationT, FileIO.Sink<OutputT>>>) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
Specifies how to create a  FileIO.Sink for a particular destination and how to map the
 element type to the sink's output type. 
- via(Contextful<Contextful.Fn<UserT, OutputT>>, FileIO.Sink<OutputT>) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- via(Contextful<Contextful.Fn<DestinationT, FileIO.Sink<UserT>>>) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- via(FileIO.Sink<UserT>) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- via(InferableFunction<? super InputT, ? extends Iterable<OutputT>>) - Static method in class org.apache.beam.sdk.transforms.FlatMapElements
- 
For a  InferableFunction<InputT, ? extends Iterable<OutputT>>fn, return a
  PTransform that applies  fn to every element of the input  PCollection<InputT> and outputs all of the elements to the output  PCollection<OutputT>. 
- via(SimpleFunction<? super InputT, ? extends Iterable<OutputT>>) - Static method in class org.apache.beam.sdk.transforms.FlatMapElements
- 
- via(ProcessFunction<NewInputT, ? extends Iterable<OutputT>>) - Method in class org.apache.beam.sdk.transforms.FlatMapElements
- 
For a  ProcessFunction<InputT, ? extends Iterable<OutputT>>fn, returns a  PTransform that applies  fn to every element of the input  PCollection<InputT>
 and outputs all of the elements to the output  PCollection<OutputT>. 
- via(SerializableFunction<NewInputT, ? extends Iterable<OutputT>>) - Method in class org.apache.beam.sdk.transforms.FlatMapElements
- 
- via(Contextful<Contextful.Fn<NewInputT, Iterable<OutputT>>>) - Method in class org.apache.beam.sdk.transforms.FlatMapElements
- 
- via(InferableFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.MapElements
- 
For InferableFunction<InputT, OutputT>fn, returns aPTransformthat
 takes an inputPCollection<InputT>and returns aPCollection<OutputT>containingfn.apply(v)for every elementvin the input.
 
- via(SimpleFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.MapElements
- 
- via(ProcessFunction<NewInputT, OutputT>) - Method in class org.apache.beam.sdk.transforms.MapElements
- 
For a ProcessFunction<InputT, OutputT>fnand output type descriptor, returns aPTransformthat takes an inputPCollection<InputT>and returns aPCollection<OutputT>containingfn.apply(v)for every elementvin the input.
 
- via(SerializableFunction<NewInputT, OutputT>) - Method in class org.apache.beam.sdk.transforms.MapElements
- 
- via(Contextful<Contextful.Fn<NewInputT, OutputT>>) - Method in class org.apache.beam.sdk.transforms.MapElements
- 
- via(SerializableFunction<NewKeyT, K2>) - Method in class org.apache.beam.sdk.transforms.MapKeys
- 
Returns a  MapKeys<K1, K2, V>PTransform for a  ProcessFunction<NewK1,
 K2> with predefined  MapKeys.outputType. 
- via(SerializableFunction<NewValueT, V2>) - Method in class org.apache.beam.sdk.transforms.MapValues
- 
- viaFlatMapFn(String, Coder<?>) - Static method in class org.apache.beam.sdk.extensions.python.transforms.PythonMap
-  
- viaMapFn(String, Coder<?>) - Static method in class org.apache.beam.sdk.extensions.python.transforms.PythonMap
-  
- viaRandomKey() - Static method in class org.apache.beam.sdk.transforms.Reshuffle
- 
Encapsulates the sequence "pair input with unique key, apply  Reshuffle.of(), drop the
 key" commonly used to break fusion. 
- VideoIntelligence - Class in org.apache.beam.sdk.extensions.ml
- 
Factory class for PTransforms integrating with Google Cloud AI - VideoIntelligence service. 
- VideoIntelligence() - Constructor for class org.apache.beam.sdk.extensions.ml.VideoIntelligence
-  
- VideoIntelligence.AnnotateVideoFromBytes - Class in org.apache.beam.sdk.extensions.ml
- 
A PTransform taking a PCollection of ByteStringand an optional side input with a
 context map and emitting lists ofVideoAnnotationResultsfor each element.
 
- VideoIntelligence.AnnotateVideoFromBytesWithContext - Class in org.apache.beam.sdk.extensions.ml
- 
A PTransform taking a PCollection of  KV of  ByteString and  VideoContext
 and emitting lists of  VideoAnnotationResults for each element. 
- VideoIntelligence.AnnotateVideoFromUri - Class in org.apache.beam.sdk.extensions.ml
- 
A PTransform taking a PCollection of Stringand an optional side input with a context
 map and emitting lists ofVideoAnnotationResultsfor each element.
 
- VideoIntelligence.AnnotateVideoFromURIWithContext - Class in org.apache.beam.sdk.extensions.ml
- 
A PTransform taking a PCollection of  KV of  String and  VideoContext and
 emitting lists of  VideoAnnotationResults for each element. 
- View - Class in org.apache.beam.sdk.transforms
- 
- View.AsIterable<T> - Class in org.apache.beam.sdk.transforms
- 
For internal use only; no backwards-compatibility guarantees. 
- View.AsList<T> - Class in org.apache.beam.sdk.transforms
- 
For internal use only; no backwards-compatibility guarantees. 
- View.AsMap<K,V> - Class in org.apache.beam.sdk.transforms
- 
For internal use only; no backwards-compatibility guarantees. 
- View.AsMultimap<K,V> - Class in org.apache.beam.sdk.transforms
- 
For internal use only; no backwards-compatibility guarantees. 
- View.AsSingleton<T> - Class in org.apache.beam.sdk.transforms
- 
For internal use only; no backwards-compatibility guarantees. 
- View.CreatePCollectionView<ElemT,ViewT> - Class in org.apache.beam.sdk.transforms
- 
For internal use only; no backwards-compatibility guarantees. 
- View.ToListViewDoFn<T> - Class in org.apache.beam.sdk.transforms
- 
Provides an index to value mapping using a random starting index and also provides an offset
 range for each window seen. 
- viewAsValues(PCollectionView<V>, Coder<V>) - Static method in class org.apache.beam.sdk.transforms.Reify
- 
Pairs each element in a collection with the value of a side input associated with the element's
 window. 
- ViewFn<PrimitiveViewT,ViewT> - Class in org.apache.beam.sdk.transforms
- 
For internal use only; no backwards-compatibility guarantees. 
- ViewFn() - Constructor for class org.apache.beam.sdk.transforms.ViewFn
-  
- viewInGlobalWindow(PCollectionView<V>, Coder<V>) - Static method in class org.apache.beam.sdk.transforms.Reify
- 
Returns a  PCollection consisting of a single element, containing the value of the given
 view in the global window. 
- ViewP - Class in org.apache.beam.runners.jet.processors
- 
Jet Processorimplementation for Beam's side input producing
 primitives.
 
- visitArrayQualifier(FieldSpecifierNotationParser.ArrayQualifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
- 
- visitArrayQualifier(FieldSpecifierNotationParser.ArrayQualifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
- 
- visitArrayQualifierList(FieldSpecifierNotationParser.ArrayQualifierListContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
- 
- visitArrayQualifierList(FieldSpecifierNotationParser.ArrayQualifierListContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
- 
- visitDotExpression(FieldSpecifierNotationParser.DotExpressionContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
- 
- visitDotExpression(FieldSpecifierNotationParser.DotExpressionContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
- 
- visitErrorNode(ErrorNode) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
- visitFieldSpecifier(FieldSpecifierNotationParser.FieldSpecifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
- 
- visitFieldSpecifier(FieldSpecifierNotationParser.FieldSpecifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
- 
- visitMapQualifier(FieldSpecifierNotationParser.MapQualifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
- 
- visitMapQualifier(FieldSpecifierNotationParser.MapQualifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
- 
- visitMapQualifierList(FieldSpecifierNotationParser.MapQualifierListContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
- 
- visitMapQualifierList(FieldSpecifierNotationParser.MapQualifierListContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
- 
- visitPrimitiveTransform(TransformHierarchy.Node) - Method in class org.apache.beam.runners.spark.SparkRunner.Evaluator
-  
- visitPrimitiveTransform(TransformHierarchy.Node) - Method in class org.apache.beam.runners.twister2.translators.Twister2BatchPipelineTranslator
-  
- visitPrimitiveTransform(TransformHierarchy.Node) - Method in class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
-  
- visitPrimitiveTransform(TransformHierarchy.Node) - Method in interface org.apache.beam.sdk.Pipeline.PipelineVisitor
- 
Called for each primitive transform after all of its topological predecessors and inputs have
 been visited. 
- visitQualifiedComponent(FieldSpecifierNotationParser.QualifiedComponentContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
- 
- visitQualifiedComponent(FieldSpecifierNotationParser.QualifiedComponentContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
- 
- visitQualifyComponent(FieldSpecifierNotationParser.QualifyComponentContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
- 
- visitQualifyComponent(FieldSpecifierNotationParser.QualifyComponentContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
- 
- visitSimpleIdentifier(FieldSpecifierNotationParser.SimpleIdentifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
- 
- visitSimpleIdentifier(FieldSpecifierNotationParser.SimpleIdentifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
- 
- visitTerminal(TerminalNode) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
- visitValue(PValue, TransformHierarchy.Node) - Method in class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
-  
- visitValue(PValue, TransformHierarchy.Node) - Method in interface org.apache.beam.sdk.Pipeline.PipelineVisitor
- 
Called for each value after the transform that produced the value has been visited. 
- visitWildcard(FieldSpecifierNotationParser.WildcardContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
- 
- visitWildcard(FieldSpecifierNotationParser.WildcardContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
- 
- VOCABULARY - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- VOCABULARY - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- VoidCoder - Class in org.apache.beam.sdk.coders
- 
- voids() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
-  
- vpnName(String) - Method in class org.apache.beam.sdk.io.solace.broker.BasicAuthJcsmpSessionServiceFactory.Builder
- 
Set Solace vpn name. 
- vpnName() - Method in class org.apache.beam.sdk.io.solace.broker.BasicAuthJcsmpSessionServiceFactory
- 
The name of the VPN to connect to. 
- vpnName(String) - Method in class org.apache.beam.sdk.io.solace.broker.BasicAuthSempClientFactory.Builder
- 
Set Solace vpn name. 
- vpnName(String) - Method in class org.apache.beam.sdk.io.solace.broker.GCPSecretSessionServiceFactory.Builder
- 
Optional. 
- vpnName() - Method in class org.apache.beam.sdk.io.solace.broker.GCPSecretSessionServiceFactory
-  
- Wait - Class in org.apache.beam.sdk.transforms
- 
Delays processing of each window in a  PCollection until signaled. 
- Wait() - Constructor for class org.apache.beam.sdk.transforms.Wait
-  
- Wait.OnSignal<T> - Class in org.apache.beam.sdk.transforms
- 
- waitForCompletion() - Method in class org.apache.beam.sdk.io.aws2.common.AsyncBatchWriteHandler
- 
Wait for all pending requests to complete and check for failures. 
- waitForNMessages(int, Duration) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
- 
- waitForPort(String, int, int) - Static method in class org.apache.beam.sdk.extensions.python.PythonService
-  
- waitForStart(Duration) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal
- 
Future that waits for a start signal for duration.
 
- waitForSuccess(Duration) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal
- 
Wait for a success signal for duration.
 
- waitForUpTo(Duration) - Method in interface org.apache.beam.sdk.io.gcp.pubsub.TestPubsub.PollingAssertion
-  
- waitTillUp(int) - Method in class org.apache.beam.sdk.transformservice.launcher.TransformServiceLauncher
-  
- waitUntilFinish() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-  
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
-  
- waitUntilFinish(Duration, MonitoringUtil.JobMessagesHandler) - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
- 
Waits until the pipeline finishes and returns the final status. 
- waitUntilFinish() - Method in class org.apache.beam.runners.direct.DirectRunner.DirectPipelineResult
- 
Waits until the pipeline finishes and returns the final status. 
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.direct.DirectRunner.DirectPipelineResult
- 
Waits until the pipeline finishes and returns the final status. 
- waitUntilFinish() - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
-  
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
-  
- waitUntilFinish() - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
-  
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
-  
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.jet.FailedRunningPipelineResults
-  
- waitUntilFinish() - Method in class org.apache.beam.runners.jet.FailedRunningPipelineResults
-  
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.jet.JetPipelineResult
-  
- waitUntilFinish() - Method in class org.apache.beam.runners.jet.JetPipelineResult
-  
- waitUntilFinish() - Method in class org.apache.beam.runners.spark.SparkPipelineResult
-  
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.spark.SparkPipelineResult
-  
- waitUntilFinish() - Method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineResult
-  
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineResult
-  
- waitUntilFinish(Duration) - Method in class org.apache.beam.runners.twister2.Twister2PipelineResult
-  
- waitUntilFinish() - Method in class org.apache.beam.runners.twister2.Twister2PipelineResult
-  
- waitUntilFinish(Duration) - Method in interface org.apache.beam.sdk.PipelineResult
- 
Waits until the pipeline finishes and returns the final status. 
- waitUntilFinish() - Method in interface org.apache.beam.sdk.PipelineResult
- 
Waits until the pipeline finishes and returns the final status. 
- WallTime(Instant) - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.WallTime
-  
- Watch - Class in org.apache.beam.sdk.transforms
- 
Given a "poll function" that produces a potentially growing set of outputs for an input, this
 transform simultaneously continuously watches the growth of output sets of all inputs, until a
 per-input termination condition is reached. 
- Watch() - Constructor for class org.apache.beam.sdk.transforms.Watch
-  
- Watch.Growth<InputT,OutputT,KeyT> - Class in org.apache.beam.sdk.transforms
- 
- Watch.Growth.PollFn<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
- 
- Watch.Growth.PollResult<OutputT> - Class in org.apache.beam.sdk.transforms
- 
- Watch.Growth.TerminationCondition<InputT,StateT> - Interface in org.apache.beam.sdk.transforms
- 
A strategy for determining whether it is time to stop polling the current input regardless of
 whether its output is complete or not. 
- Watch.WatchGrowthFn<InputT,OutputT,KeyT,TerminationStateT> - Class in org.apache.beam.sdk.transforms
-  
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Parse
- 
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>, boolean) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ParseAll
- 
Deprecated. Like Read#watchForNewFiles(Duration, TerminationCondition, boolean).
 
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ParseAll
- 
Deprecated. Like Read#watchForNewFiles(Duration, TerminationCondition).
 
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>, boolean) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Read
- 
Continuously watches for new files matching the filepattern, polling it at the given
 interval, until the given termination condition is reached. 
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Read
- 
Same as Read#watchForNewFiles(Duration, TerminationCondition, boolean)withmatchUpdatedFiles=false.
 
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadAll
- 
Deprecated. 
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>, boolean) - Method in class org.apache.beam.sdk.io.TextIO.Read
- 
See MatchConfiguration#continuously(Duration, TerminationCondition, boolean).
 
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.TextIO.Read
- 
Same as Read#watchForNewFiles(Duration, TerminationCondition, boolean)withmatchUpdatedFiles=false.
 
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>, boolean) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
- 
Deprecated. Same as Read#watchForNewFiles(Duration, TerminationCondition, boolean).
 
- watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
- 
Deprecated. Same as Read#watchForNewFiles(Duration, TerminationCondition).
 
- WatermarkAdvancingStreamingListener() - Constructor for class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.WatermarkAdvancingStreamingListener
-  
- WatermarkCache - Interface in org.apache.beam.sdk.io.gcp.spanner.changestreams.cache
-  
- WatermarkEstimator<WatermarkEstimatorStateT> - Interface in org.apache.beam.sdk.transforms.splittabledofn
- 
- WatermarkEstimators - Class in org.apache.beam.sdk.fn.splittabledofn
- 
- WatermarkEstimators() - Constructor for class org.apache.beam.sdk.fn.splittabledofn.WatermarkEstimators
-  
- WatermarkEstimators - Class in org.apache.beam.sdk.transforms.splittabledofn
- 
- WatermarkEstimators.Manual - Class in org.apache.beam.sdk.transforms.splittabledofn
- 
- WatermarkEstimators.MonotonicallyIncreasing - Class in org.apache.beam.sdk.transforms.splittabledofn
- 
A watermark estimator that observes timestamps of records output from a DoFn reporting the
 timestamp of the last element seen as the current watermark. 
- WatermarkEstimators.WallTime - Class in org.apache.beam.sdk.transforms.splittabledofn
- 
A watermark estimator that tracks wall time. 
- WatermarkEstimators.WatermarkAndStateObserver<WatermarkEstimatorStateT> - Interface in org.apache.beam.sdk.fn.splittabledofn
- 
Interface which allows for accessing the current watermark and watermark estimator state. 
- WatermarkEvent() - Constructor for class org.apache.beam.sdk.testing.TestStream.WatermarkEvent
-  
- WatermarkHoldState - Interface in org.apache.beam.sdk.state
- 
For internal use only; no backwards-compatibility guarantees. 
- WatermarkParameters - Class in org.apache.beam.sdk.io.aws2.kinesis
- 
WatermarkParameterscontains the parameters used for watermark computation.
 
- WatermarkParameters() - Constructor for class org.apache.beam.sdk.io.aws2.kinesis.WatermarkParameters
-  
- WatermarkPolicy - Interface in org.apache.beam.sdk.io.aws2.kinesis
- 
Implement this interface to define a custom watermark calculation heuristic. 
- WatermarkPolicyFactory - Interface in org.apache.beam.sdk.io.aws2.kinesis
- 
Implement this interface to create a WatermarkPolicy.
 
- WatermarkPolicyFactory.ArrivalTimeWatermarkPolicy - Class in org.apache.beam.sdk.io.aws2.kinesis
- 
- WatermarkPolicyFactory.CustomWatermarkPolicy - Class in org.apache.beam.sdk.io.aws2.kinesis
- 
- WatermarkPolicyFactory.ProcessingTimeWatermarkPolicy - Class in org.apache.beam.sdk.io.aws2.kinesis
- 
Watermark policy where the processing time is used as the event time. 
- watermarkStateInternal(TimestampCombiner) - Static method in class org.apache.beam.sdk.state.StateSpecs
- 
For internal use only; no backwards-compatibility guarantees. 
- WebPathParser - Class in org.apache.beam.sdk.io.gcp.healthcare
-  
- WebPathParser() - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.WebPathParser
-  
- WebPathParser.DicomWebPath - Class in org.apache.beam.sdk.io.gcp.healthcare
-  
- weeks(int, int) - Static method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows
- 
Returns a  WindowFn that windows elements into periods measured by weeks. 
- WeightedList<T> - Class in org.apache.beam.sdk.fn.data
- 
Facade for a Listthat keeps track of weight, for cache limit reasons.
 
- WeightedList(List<T>, long) - Constructor for class org.apache.beam.sdk.fn.data.WeightedList
-  
- where(TypeParameter<X>, TypeDescriptor<X>) - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
Returns a new TypeDescriptorwhere the type variable represented bytypeParameterare substituted bytype.
 
- where(Type, Type) - Method in class org.apache.beam.sdk.values.TypeDescriptor
- 
- whereFieldId(int, SerializableFunction<FieldT, Boolean>) - Method in class org.apache.beam.sdk.schemas.transforms.Filter.Inner
- 
Set a predicate based on the value of a field, where the field is specified by id. 
- whereFieldIds(List<Integer>, SerializableFunction<Row, Boolean>) - Method in class org.apache.beam.sdk.schemas.transforms.Filter.Inner
- 
Set a predicate based on the value of multipled fields, specified by id. 
- whereFieldName(String, SerializableFunction<FieldT, Boolean>) - Method in class org.apache.beam.sdk.schemas.transforms.Filter.Inner
- 
Set a predicate based on the value of a field, where the field is specified by name. 
- whereFieldNames(List<String>, SerializableFunction<Row, Boolean>) - Method in class org.apache.beam.sdk.schemas.transforms.Filter.Inner
- 
Set a predicate based on the value of multipled fields, specified by name. 
- widening(Schema) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
-  
- Widening() - Constructor for class org.apache.beam.sdk.schemas.transforms.Cast.Widening
-  
- WILDCARD - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- WILDCARD - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-  
- WILDCARD() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.WildcardContext
-  
- WildcardContext(FieldSpecifierNotationParser.DotExpressionComponentContext) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.WildcardContext
-  
- wildcardToRegexp(String) - Static method in class org.apache.beam.sdk.io.FileSystemUtils
- 
Expands glob expressions to regular expressions. 
- WINDMILL_SERVICE_EXPERIMENT - Static variable in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
- 
- WindmillServiceStreamingRpcBatchLimitFactory() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowStreamingPipelineOptions.WindmillServiceStreamingRpcBatchLimitFactory
-  
- window() - Method in class org.apache.beam.runners.twister2.utils.Twister2AssignContext
-  
- window() - Method in interface org.apache.beam.sdk.state.StateContext
- 
Returns the window corresponding to the state. 
- window() - Method in class org.apache.beam.sdk.transforms.DoFn.OnTimerContext
- 
Returns the window in which the timer is firing. 
- window() - Method in class org.apache.beam.sdk.transforms.DoFn.OnWindowExpirationContext
- 
Returns the window in which the window expiration is firing. 
- Window<T> - Class in org.apache.beam.sdk.transforms.windowing
- 
- Window() - Constructor for class org.apache.beam.sdk.transforms.windowing.Window
-  
- window() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn.AssignContext
- 
Returns the window of the current element prior to this WindowFnbeing called.
 
- Window.Assign<T> - Class in org.apache.beam.sdk.transforms.windowing
- 
- Window.ClosingBehavior - Enum in org.apache.beam.sdk.transforms.windowing
- 
Specifies the conditions under which a final pane will be created when a window is permanently
 closed. 
- Window.OnTimeBehavior - Enum in org.apache.beam.sdk.transforms.windowing
- 
Specifies the conditions under which an on-time pane will be created when a window is closed. 
- WINDOW_END - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.TVFStreamingUtils
-  
- WINDOW_START - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.TVFStreamingUtils
-  
- windowCoder() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.BagUserStateSpec
-  
- windowCoder() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.SideInputSpec
-  
- windowCoder(PCollection<T>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator
-  
- windowCoder() - Method in class org.apache.beam.sdk.extensions.sql.impl.TVFSlidingWindowFn
-  
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-  
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-  
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-  
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
-  
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
-  
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
-  
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
-  
- windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
- 
Returns the  Coder used for serializing the windows used by this windowFn. 
- WindowedContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.WindowedContext
-  
- windowedEncoder(Coder<T>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- windowedEncoder(Encoder<T>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- windowedEncoder(Coder<T>, Coder<W>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- windowedFilename(int, int, BoundedWindow, PaneInfo, FileBasedSink.OutputFileHints) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy
-  
- windowedFilename(int, int, BoundedWindow, PaneInfo, FileBasedSink.OutputFileHints) - Method in class org.apache.beam.sdk.io.FileBasedSink.FilenamePolicy
- 
When a sink has requested windowed or triggered output, this method will be invoked to return
 the file  resource to be created given the base output directory and a
  FileBasedSink.OutputFileHints containing information about the file, including a suggested
 extension (e.g. 
- windowedMultiReceiver(DoFn<?, ?>.WindowedContext, Map<TupleTag<?>, Coder<?>>) - Static method in class org.apache.beam.sdk.transforms.DoFnOutputReceivers
- 
- windowedMultiReceiver(DoFn<?, ?>.WindowedContext) - Static method in class org.apache.beam.sdk.transforms.DoFnOutputReceivers
- 
- windowedReceiver(DoFn<?, ?>.WindowedContext, TupleTag<T>) - Static method in class org.apache.beam.sdk.transforms.DoFnOutputReceivers
- 
- windowedValueEncoder(Encoder<T>, Encoder<W>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers
- 
Creates a Spark EncoderforTofStructTypewith fieldsvalue,timestamp,windowsandpane.
 
- windowedValues(Iterable<WindowedValue<T>>) - Static method in class org.apache.beam.sdk.transforms.Create
- 
Returns a new  Create.WindowedValues transform that produces a  PCollection
 containing the elements of the provided  Iterable with the specified windowing metadata. 
- windowedValues(WindowedValue<T>, WindowedValue<T>...) - Static method in class org.apache.beam.sdk.transforms.Create
- 
- windowedWrites - Variable in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
- 
Whether windowed writes are being used. 
- windowEncoder() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator.Context
-  
- WindowFn<T,W extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
- 
The argument to the  Window transform used to assign elements into windows and to
 determine how windows are merged. 
- WindowFn() - Constructor for class org.apache.beam.sdk.transforms.windowing.WindowFn
-  
- WindowFn.AssignContext - Class in org.apache.beam.sdk.transforms.windowing
- 
- WindowFn.MergeContext - Class in org.apache.beam.sdk.transforms.windowing
- 
- WindowFnTestUtils - Class in org.apache.beam.sdk.testing
- 
- WindowFnTestUtils() - Constructor for class org.apache.beam.sdk.testing.WindowFnTestUtils
-  
- WindowGroupP<K,V> - Class in org.apache.beam.runners.jet.processors
- 
Jet Processorimplementation for Beam's GroupByKeyOnly +
 GroupAlsoByWindow primitives.
 
- WINDOWING_STRATEGY - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
-  
- WindowingStrategy<T,W extends BoundedWindow> - Class in org.apache.beam.sdk.values
- 
A WindowingStrategydescribes the windowing behavior for a specific collection of values.
 
- WindowingStrategy.AccumulationMode - Enum in org.apache.beam.sdk.values
- 
The accumulation modes that can be used with windowing. 
- WindowIntoTransformProvider - Class in org.apache.beam.sdk.expansion.service
- 
- WindowIntoTransformProvider() - Constructor for class org.apache.beam.sdk.expansion.service.WindowIntoTransformProvider
-  
- WindowIntoTransformProvider.Configuration - Class in org.apache.beam.sdk.expansion.service
-  
- WindowIntoTransformProvider.Configuration.Builder - Class in org.apache.beam.sdk.expansion.service
-  
- WindowMappingFn<TargetWindowT extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
- 
A function that takes the windows of elements in a main input and maps them to the appropriate
 window in a  PCollectionView consumed as a  side input. 
- WindowMappingFn() - Constructor for class org.apache.beam.sdk.transforms.windowing.WindowMappingFn
- 
- WindowMappingFn(Duration) - Constructor for class org.apache.beam.sdk.transforms.windowing.WindowMappingFn
- 
- windowOnlyContext(W) - Static method in class org.apache.beam.sdk.state.StateContexts
-  
- windows() - Static method in class org.apache.beam.sdk.transforms.Reify
- 
- windows() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn.MergeContext
- 
Returns the current set of windows. 
- windowsInValue() - Static method in class org.apache.beam.sdk.transforms.Reify
- 
Create a  PTransform that will output all input  KVs with the window pane info
 inside the value. 
- WireCoders - Class in org.apache.beam.runners.fnexecution.wire
- 
Helpers to construct coders for gRPC port reads and writes. 
- with(PCollection<KV<K, V2>>, V1, V2) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join.FullOuterJoin
-  
- with(PCollection<KV<K, V2>>) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join.InnerJoin
-  
- with(PCollection<KV<K, V2>>, V2) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join.LeftOuterJoin
-  
- with(PCollection<KV<K, V2>>, V1) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join.RightOuterJoin
-  
- with(SimpleFunction<DataT, InputT>, Combine.CombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposeCombineFnBuilder
- 
- with(SimpleFunction<DataT, InputT>, Coder, Combine.CombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposeCombineFnBuilder
- 
Like #with(SimpleFunction, CombineFn, TupleTag)but with an explicit input coder.
 
- with(SimpleFunction<DataT, InputT>, CombineWithContext.CombineFnWithContext<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposeCombineFnBuilder
- 
- with(SimpleFunction<DataT, InputT>, Coder, CombineWithContext.CombineFnWithContext<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposeCombineFnBuilder
- 
Like #with(SimpleFunction, CombineFnWithContext, TupleTag)but with input coder.
 
- with(SimpleFunction<DataT, InputT>, Combine.CombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
- 
- with(SimpleFunction<DataT, InputT>, CombineWithContext.CombineFnWithContext<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
- 
- with(SimpleFunction<DataT, InputT>, Coder, Combine.CombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
- 
- with(SimpleFunction<DataT, InputT>, Coder, CombineWithContext.CombineFnWithContext<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
- 
- with(SimpleFunction<DataT, InputT>, CombineFnBase.GlobalCombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
- 
- with(SimpleFunction<DataT, InputT>, Coder<InputT>, CombineFnBase.GlobalCombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
- 
- with(PCollection<T>) - Static method in class org.apache.beam.sdk.transforms.Flatten
- 
- with(PTransform<PBegin, PCollection<T>>) - Static method in class org.apache.beam.sdk.transforms.Flatten
- 
- withAccelerator(String) - Method in class org.apache.beam.sdk.transforms.resourcehints.ResourceHints
- 
Declares hardware accelerators that are desired to have in the execution environment. 
- withAccuracy(double, double) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.CountMinSketchFn
- 
- withAddresses(List<String>) - Method in class org.apache.beam.sdk.io.amqp.AmqpIO.Read
- 
Define the AMQP addresses where to receive messages. 
- withAdminUrl(String) - Method in class org.apache.beam.sdk.io.pulsar.PulsarIO.Read
-  
- withAllFields() - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
-  
- withAllowableResponseErrors(Set<String>) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BulkIO
- 
Provide a set of textual error types which can be contained in Bulk API response
 items[].error.type field. 
- withAllowableResponseErrors(Set<String>) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withAllowDuplicates(Boolean) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- withAllowDuplicates() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
-  
- withAllowDuplicates(boolean) - Method in class org.apache.beam.sdk.transforms.Redistribute.RedistributeArbitrarily
-  
- withAllowDuplicates(boolean) - Method in class org.apache.beam.sdk.transforms.Redistribute.RedistributeByKey
-  
- withAllowedLateness(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.Window
- 
Override the amount of lateness allowed for data elements in the output  PCollection and
 downstream  PCollections until explicitly set again. 
- withAllowedLateness(Duration, Window.ClosingBehavior) - Method in class org.apache.beam.sdk.transforms.windowing.Window
- 
Override the amount of lateness allowed for data elements in the pipeline. 
- withAllowedLateness(Duration) - Method in class org.apache.beam.sdk.values.WindowingStrategy
- 
Returns a  WindowingStrategy identical to  this but with the allowed lateness set
 to  allowedLateness. 
- withAllowedTimestampSkew(Duration) - Method in class org.apache.beam.sdk.transforms.WithTimestamps
- 
- withAlreadyMerged(boolean) - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- withApiKey(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
If Elasticsearch authentication is enabled, provide an API key. 
- withAppendOnly(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.DocToBulk
- 
Provide an instruction to control whether the target index should be considered append-only. 
- withAppendOnly(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withAppProfileId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
Returns a new  BigtableIO.Read that will read using the specified app profile id. 
- withAppProfileId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
Returns a new  BigtableIO.Read that will read using the specified app profile id. 
- withAppProfileId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
- withAppProfileId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
Returns a new  BigtableIO.Write that will write using the specified app profile id. 
- withAppProfileId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
Returns a new  BigtableIO.Write that will write using the specified app profile id. 
- withApproximateTrim(boolean) - Method in class org.apache.beam.sdk.io.redis.RedisIO.WriteStreams
- 
- withArgs(Object...) - Method in class org.apache.beam.sdk.extensions.python.PythonExternalTransform
- 
Positional arguments for the Python cross-language transform. 
- withArrivalTimePolicy() - Static method in interface org.apache.beam.sdk.io.aws2.kinesis.WatermarkPolicyFactory
- 
Returns an ArrivalTimeWatermarkPolicy. 
- withArrivalTimePolicy(Duration) - Static method in interface org.apache.beam.sdk.io.aws2.kinesis.WatermarkPolicyFactory
- 
Returns an ArrivalTimeWatermarkPolicy. 
- withArrivalTimeWatermarkPolicy() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specifies the WatermarkPolicyFactoryas ArrivalTimeWatermarkPolicyFactory.
 
- withArrivalTimeWatermarkPolicy(Duration) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specifies the WatermarkPolicyFactoryas ArrivalTimeWatermarkPolicyFactory.
 
- withAttemptTimeout(Duration) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
- withAttemptTimeout(Duration) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withAuth(String) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
- 
Define the password to authenticate on the Redis server. 
- withAuth(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
- 
- withAuth(String) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
-  
- withAuth(String) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadKeyPatterns
-  
- withAuth(String) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
-  
- withAuth(String) - Method in class org.apache.beam.sdk.io.redis.RedisIO.WriteStreams
- 
Use the redis AUTH command when connecting to the server; the format of the string can be
 either just a password or a username and password separated by a space. 
- withAuthenticator(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets authenticator for Snowflake. 
- withAutoLoading(boolean) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
-  
- withAutoScaler(AutoScaler) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
- 
Sets the  AutoScaler to use for reporting backlog during the execution of this source. 
- withAutoSchemaUpdate(boolean) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
If true, enables automatically detecting BigQuery table schema updates. 
- withAutoSharding() - Method in class org.apache.beam.sdk.io.FileIO.Write
-  
- withAutoSharding() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
If true, enables using a dynamically determined number of shards to write to BigQuery. 
- withAutoSharding() - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
- 
- withAutoSharding() - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
- 
If true, enables using a dynamically determined number of shards to write. 
- withAutoSharding() - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteWithResults
- 
If true, enables using a dynamically determined number of shards to write. 
- withAutoSharding() - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
-  
- withAutoSharding() - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withAutoSharding() - Method in class org.apache.beam.sdk.io.WriteFiles
-  
- withAvroDataModel(GenericData) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Read
- 
Define the Avro data model; see AvroParquetReader.Builder#withDataModel(GenericData).
 
- withAvroDataModel(GenericData) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.ReadFiles
- 
Define the Avro data model; see AvroParquetReader.Builder#withDataModel(GenericData).
 
- withAvroDataModel(GenericData) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Sink
- 
Define the Avro data model; see AvroParquetWriter.Builder#withDataModel(GenericData).
 
- withAvroFormatFunction(SerializableFunction<AvroWriteRequest<T>, GenericRecord>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Formats the user's type into a  GenericRecord to be written to BigQuery. 
- withAvroSchemaFactory(SerializableFunction<TableSchema, Schema>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- withAvroWriter(SerializableFunction<Schema, DatumWriter<T>>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Writes the user's type as avro using the supplied  DatumWriter. 
- withAvroWriter(SerializableFunction<AvroWriteRequest<T>, AvroT>, SerializableFunction<Schema, DatumWriter<AvroT>>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Convert's the user's type to an avro record using the supplied avroFormatFunction. 
- withBackendVersion(int) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.DocToBulk
- 
Use to set explicitly which version of Elasticsearch the destination cluster is running. 
- withBackendVersion(int) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withBacklogReplicationAdjustment(Duration) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
- withBadRecordErrorHandler(ErrorHandler<BadRecord, ?>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
See FileIO.Write#withBadRecordErrorHandler(ErrorHandler)for details on usage.
 
- withBadRecordErrorHandler(ErrorHandler<BadRecord, ?>) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- withBadRecordErrorHandler(ErrorHandler<BadRecord, ?>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- withBadRecordErrorHandler(ErrorHandler<BadRecord, ?>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
-  
- withBadRecordErrorHandler(ErrorHandler<BadRecord, ?>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
- 
- withBadRecordErrorHandler(ErrorHandler<BadRecord, ?>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- withBadRecordErrorHandler(ErrorHandler<BadRecord, ?>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
- withBadRecordErrorHandler(ErrorHandler<BadRecord, ?>) - Method in class org.apache.beam.sdk.io.WriteFiles
- 
- withBaseFilename(ResourceId) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
- 
Sets the base filename. 
- withBaseFilename(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
- 
- withBasicCredentials(String, String) - Method in class org.apache.beam.sdk.io.solr.SolrIO.ConnectionConfiguration
- 
If Solr basic authentication is enabled, provide the username and password. 
- withBatchCount(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.splunk.SplunkIO.Write
- 
- withBatchCount(Integer) - Method in class org.apache.beam.sdk.io.splunk.SplunkIO.Write
- 
Sets batchCount for sending multiple events in a single request to the HEC. 
- withBatching(boolean) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
- 
If true the uses Cloud Spanner batch API. 
- withBatching(boolean) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
- 
- withBatchInitialCount(int) - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions.Builder
- 
Configure the initial size of a batch; used in the absence of the QoS system having
 significant data to determine a better batch size. 
- withBatchMaxBufferingDuration(Duration) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Returns a new  TextIO.TypedWrite that will batch the input records using specified max
 buffering duration. 
- withBatchMaxBufferingDuration(Duration) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withBatchMaxBufferingDuration(Duration) - Method in class org.apache.beam.sdk.io.WriteFiles
- 
Returns a new  WriteFiles that will batch the input records using specified max
 buffering duration. 
- withBatchMaxBytes(int) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Write
- 
Max. 
- withBatchMaxBytes(long) - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions.Builder
- 
Configure the maximum number of bytes to include in a batch. 
- withBatchMaxCount(int) - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions.Builder
- 
Configure the maximum number of writes to include in a batch. 
- withBatchMaxRecords(int) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Write
- 
Max. 
- withBatchSize(Integer) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportCatalogItems
-  
- withBatchSize(Integer) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportUserEvents
-  
- withBatchSize(int) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.WriteBatches
- 
The batch size to use, default (and AWS limit) is 10.
 
- withBatchSize(long) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
- 
Provide a size for the scroll read. 
- withBatchSize(long) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
- 
Sets batch size for the write operation. 
- withBatchSize(int) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Write
- 
Number of elements to batch. 
- withBatchSize(long) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
- 
- withBatchSize(long) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
- 
Provide a maximum size in number of SQL statement for the batch. 
- withBatchSize(int) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Read
- 
Reads from the table in batches of the specified size. 
- withBatchSize(long) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
- 
Define the size of the batch to group write operations. 
- withBatchSize(long) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- withBatchSize(ValueProvider<Long>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- withBatchSize(int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
-  
- withBatchSize(int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadKeyPatterns
-  
- withBatchSize(Integer) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.Write
- 
Provide a maximum number of rows that is written by one SQL statement. 
- withBatchSize(Integer) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Returns a new  TextIO.TypedWrite that will batch the input records using specified batch
 size. 
- withBatchSize(Integer) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withBatchSize(Integer) - Method in class org.apache.beam.sdk.io.WriteFiles
- 
Returns a new  WriteFiles that will batch the input records using specified batch size. 
- withBatchSizeBytes(long) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the batch size limit (max number of bytes mutated per batch). 
- withBatchSizeBytes(Integer) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Returns a new  TextIO.TypedWrite that will batch the input records using specified batch size
 in bytes. 
- withBatchSizeBytes(Integer) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withBatchSizeBytes(Integer) - Method in class org.apache.beam.sdk.io.WriteFiles
- 
Returns a new  WriteFiles that will batch the input records using specified batch size
 in bytes. 
- withBatchTargetLatency(Duration) - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions.Builder
- 
Target latency for batch requests. 
- withBatchTimeout(Duration) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.WriteBatches
- 
The duration to accumulate records before timing out, default is 3 secs. 
- withBatchTimeout(Duration, boolean) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.WriteBatches
- 
The duration to accumulate records before timing out, default is 3 secs. 
- withBeamRowConverters(TypeDescriptor<T>, BigQueryIO.TypedRead.ToBeamRowFunction<T>, BigQueryIO.TypedRead.FromBeamRowFunction<T>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
Sets the functions to convert elements to/from  Row objects. 
- withBeamRowConverters(TypeDescriptor<Struct>, SpannerIO.Read.ToBeamRowFunction, SpannerIO.Read.FromBeamRowFunction) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withBeamSchemas(boolean) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Read
- 
If set to true, a Beam schema will be inferred from the AVRO schema. 
- withBeamSchemas(boolean) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadAll
- 
Deprecated. If set to true, a Beam schema will be inferred from the AVRO schema. 
- withBeamSchemas(boolean) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadFiles
- 
If set to true, a Beam schema will be inferred from the AVRO schema. 
- withBeamSchemas(boolean) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Read
-  
- withBeamSchemas(boolean) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.ReadFiles
-  
- withBearerToken(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
If Elasticsearch authentication is enabled, provide a bearer token. 
- withBigLakeConfiguration(Map<String, String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Specifies a configuration to create BigLake tables. 
- withBigtableOptions(BigtableOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableConfig
- 
- withBigtableOptions(BigtableOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
- withBigtableOptions(BigtableOptions.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
- withBigtableOptions(BigtableOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withBigtableOptions(BigtableOptions.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withBigtableOptionsConfigurator(SerializableFunction<BigtableOptions.Builder, BigtableOptions.Builder>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableConfig
- 
- withBigtableOptionsConfigurator(SerializableFunction<BigtableOptions.Builder, BigtableOptions.Builder>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
- withBigtableOptionsConfigurator(SerializableFunction<BigtableOptions.Builder, BigtableOptions.Builder>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withBootstrapServers(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Sets the bootstrap servers for the Kafka consumer. 
- withBootstrapServers(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Sets the bootstrap servers to use for the Kafka consumer if unspecified via
 KafkaSourceDescriptor#getBootStrapServers()}. 
- withBootstrapServers(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
- 
- withBootstrapServers(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
- 
Returns a new  KafkaIO.Write transform with Kafka producer pointing to  bootstrapServers. 
- withBucket(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-  
- withBucket(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-  
- withBucketAuto(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
- 
Sets weather to use $bucketAuto or not. 
- withBulkDirective(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Document
- 
Sets the bulk directive representation of an input document. 
- withByteSize(long) - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches
-  
- withByteSize(long, SerializableFunction<InputT, Long>) - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches
-  
- withCache(Cache.Pair<RequestT, ResponseT>) - Method in class org.apache.beam.io.requestresponse.RequestResponseIO
- 
Configures  RequestResponseIO for reading and writing  RequestT and  ResponseT pairs using a cache. 
- withCallShouldBackoff(CallShouldBackoff<ResponseT>) - Method in class org.apache.beam.io.requestresponse.RequestResponseIO
- 
Overrides the private no-op implementation of  CallShouldBackoff that determines whether
 the  DoFn should hold  RequestTs. 
- withCatalogName(String) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAICreateCatalogItem
-  
- withCatalogName(String) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportCatalogItems
-  
- withCatalogName(String) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportUserEvents
-  
- withCatalogName(String) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIPredict
-  
- withCatalogName(String) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIWriteUserEvent
-  
- withCdapPlugin(Plugin<K, V>) - Method in class org.apache.beam.sdk.io.cdap.CdapIO.Read
- 
- withCdapPlugin(Plugin<K, V>) - Method in class org.apache.beam.sdk.io.cdap.CdapIO.Write
- 
- withCdapPluginClass(Class<?>) - Method in class org.apache.beam.sdk.io.cdap.CdapIO.Read
- 
Sets a CDAP Plugin class. 
- withCdapPluginClass(Class<?>) - Method in class org.apache.beam.sdk.io.cdap.CdapIO.Write
- 
Sets a CDAP Plugin class. 
- withChangeStreamName(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
- withChangeStreamName(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
Specifies the change stream name. 
- withCharset(Charset) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
- 
Sets the XML file charset. 
- withCharset(Charset) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
- 
- withCharset(Charset) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Sink
-  
- withCharset(Charset) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
- 
Sets the charset used to write the file. 
- withCheckStopReadingFn(CheckStopReadingFn) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
- withCheckStopReadingFn(SerializableFunction<TopicPartition, Boolean>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
- withCheckStopReadingFn(CheckStopReadingFn) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
- withCheckStopReadingFn(SerializableFunction<TopicPartition, Boolean>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
- withChunkSize(Long) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-  
- withClientConfiguration(ClientConfiguration) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
- 
Configuration of DynamoDB client. 
- withClientConfiguration(ClientConfiguration) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Write
- 
Configuration of DynamoDB client. 
- withClientConfiguration(ClientConfiguration) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Configuration of Kinesis & Cloudwatch clients. 
- withClientConfiguration(ClientConfiguration) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Write
- 
Configuration of Kinesis client. 
- withClientConfiguration(ClientConfiguration) - Method in class org.apache.beam.sdk.io.aws2.sns.SnsIO.Write
- 
Configuration of SNS client. 
- withClientConfiguration(ClientConfiguration) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Read
- 
Configuration of SQS client. 
- withClientConfiguration(ClientConfiguration) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Write
- 
Deprecated. Configuration of SQS client. 
- withClientConfiguration(ClientConfiguration) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.WriteBatches
- 
Configuration of SQS client. 
- withClientFactory(PubsubClient.PubsubClientFactory) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
- 
The default client to write to Pub/Sub is the  PubsubJsonClient, created by the  PubsubJsonClient.PubsubJsonClientFactory. 
- withClientFactory(PubsubClient.PubsubClientFactory) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
- 
The default client to write to Pub/Sub is the  PubsubJsonClient, created by the  PubsubJsonClient.PubsubJsonClientFactory. 
- withClientId(String) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
- 
Set up the client ID prefix, which is used to construct a unique client ID. 
- withClientUrl(String) - Method in class org.apache.beam.sdk.io.pulsar.PulsarIO.Read
-  
- withClientUrl(String) - Method in class org.apache.beam.sdk.io.pulsar.PulsarIO.Write
-  
- withCloseTimeout(Duration) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
- 
Sets the amount of time to wait for callbacks from the runner stating that the output has
 been durably persisted before closing the connection to the JMS broker. 
- withClosingBehavior(Window.ClosingBehavior) - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- withClustering(Clustering) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Specifies the clustering fields to use when writing to a single output table. 
- withClustering() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- withCodec(CodecFactory) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Sink
- 
Specifies to use the given  CodecFactory for each generated file. 
- withCodec(CodecFactory) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
Writes to Avro file(s) compressed using specified codec. 
- withCodec(CodecFactory) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- withCoder(Coder<T>) - Method in class org.apache.beam.io.debezium.DebeziumIO.Read
- 
Applies a  Coder to the connector. 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Parse
- 
Sets a coder for the result of the parse function. 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ParseAll
- 
Deprecated. Specifies the coder for the result of the parseFn.
 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ParseFiles
- 
Specifies the coder for the result of the parseFn.
 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Read
- 
Sets a coder for the result of the read function. 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadFiles
- 
Sets a coder for the result of the read function. 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
- 
Specifies the coder for the result of the AvroSource.
 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
-  
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.azure.cosmos.CosmosIO.Read
- 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.ReadAll
- 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
Sets a  Coder for the result of the parse function. 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
- 
- withCoder(Coder<OutputT>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
- 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadWithPartitions
- 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
-  
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Read
- 
Sets a  Coder for the result of the parse function. 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-  
- withCoder(Coder<OutputT>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.ReadAll
-  
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Parse
- 
Specify the output coder to use for output of the ParseFn.
 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.ParseFiles
- 
Specify the output coder to use for output of the ParseFn.
 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
- 
A Coder to be used by the output PCollection generated by the source. 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.transforms.Create.TimestampedValues
- 
Returns a  Create.TimestampedValues PTransform like this one that uses the given
  Coder<T> to decode each of the objects into a value of type  T. 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.transforms.Create.Values
- 
Returns a  Create.Values PTransform like this one that uses the given  Coder<T>
 to decode each of the objects into a value of type  T. 
- withCoder(Coder<T>) - Method in class org.apache.beam.sdk.transforms.Create.WindowedValues
- 
Returns a  Create.WindowedValues PTransform like this one that uses the given  Coder<T> to decode each of the objects into a value of type  T. 
- withCoderAndParseFn(Coder<T>, SimpleFunction<PubsubMessage, T>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
- 
Causes the source to return a PubsubMessage that includes Pubsub attributes, and uses the
 given parsing function to transform the PubsubMessage into an output type. 
- withCollection(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
- 
Sets the collection to consider in the database. 
- withCollection(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
- 
Sets the collection where to write data in the database. 
- withColumns(String...) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- withColumns(List<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- withColumns(String...) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withColumns(List<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withCommitDeadline(Duration) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the commit deadline. 
- withCommitDeadline(ValueProvider<Duration>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the commit deadline. 
- withCommitDeadline(Duration) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the deadline for the Commit API call. 
- withCommitRetrySettings(RetrySettings) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the commit retry settings. 
- withCompression(double) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.GlobalDigest
- 
Sets the compression factor cf.
 
- withCompression(double) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.PerKeyDigest
- 
Sets the compression factor cf.
 
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.CompressedSource
- 
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.Read
- 
Reads from input sources using the specified compression type. 
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.csv.CsvIO.Write
- 
Specifies the  Compression of all generated shard files. 
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.FileIO.ReadMatches
- 
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
Specifies to compress all generated shard files using the given  Compression and, by
 default, append the respective extension to the filename. 
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.json.JsonIO.Write
- 
Specifies the  Compression of all generated shard files. 
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.TextIO.Read
- 
Reads from input sources using the specified compression type. 
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
- 
Deprecated. Reads from input sources using the specified compression type. 
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Returns a transform for writing to text files like this one but that compresses output using
 the given  Compression. 
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
- 
Returns a transform for reading TFRecord files that decompresses all input files using the
 specified compression type. 
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
- 
Writes to output files using the specified compression type. 
- withCompression(Compression) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
- 
Decompresses all input files using the specified compression type. 
- withCompressionCodec(CompressionCodecName) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Sink
- 
Specifies compression codec. 
- withCompressionEnabled(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
Configure whether the REST client should compress requests using gzip content encoding and
 add the "Accept-Encoding: gzip". 
- withCompressionType(TextIO.CompressionType) - Method in class org.apache.beam.sdk.io.TextIO.Read
- 
- withCompressionType(TextIO.CompressionType) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
- 
- withCompressionType(TFRecordIO.CompressionType) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
- 
- withCompressionType(TFRecordIO.CompressionType) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
- 
- withCompressionType(XmlIO.Read.CompressionType) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
- 
- withConcurrentRequests(int) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Write
- 
Max number of concurrent batch write requests per bundle. 
- withConcurrentRequests(int) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.WriteBatches
- 
Max number of concurrent batch write requests per bundle, default is 5.
 
- withConfidence(double) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.GlobalSketch
- 
Sets the confidencevalue, i.e.
 
- withConfidence(double) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.PerKeySketch
- 
Sets the confidencevalue, i.e.
 
- withConfig(PluginConfig) - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Sets a plugin config. 
- withConfig(Config) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverConfiguration
-  
- withConfig(Map<String, Object>) - Method in class org.apache.beam.sdk.managed.Managed.ManagedTransform
- 
Use the input Map of configuration arguments to build and instantiate the underlying
 transform. 
- withConfigProperties(Map<String, String>) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
- 
Sets the configuration properties like metastore URI. 
- withConfigProperties(Map<String, String>) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
- 
Sets the configuration properties like metastore URI. 
- withConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.FileIO.Match
- 
- withConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.FileIO.MatchAll
- 
- withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
- 
Reads from the source using the options provided by the given configuration. 
- withConfiguration(Configuration) - Method in interface org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write.WriteBuilder
- 
Writes to the sink using the options provided by the given hadoop configuration. 
- withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
- 
Reads from the HBase instance indicated by the* given configuration. 
- withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
- 
Writes to the HBase instance indicated by the* given Configuration. 
- withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.WriteRowMutations
- 
Writes to the HBase instance indicated by the given Configuration. 
- withConfiguration(Map<String, String>) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Parse
- 
Specify Hadoop configuration for ParquetReader. 
- withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Parse
- 
Specify Hadoop configuration for ParquetReader. 
- withConfiguration(Map<String, String>) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.ParseFiles
- 
Specify Hadoop configuration for ParquetReader. 
- withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.ParseFiles
- 
Specify Hadoop configuration for ParquetReader. 
- withConfiguration(Map<String, String>) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Read
- 
Specify Hadoop configuration for ParquetReader. 
- withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Read
- 
Specify Hadoop configuration for ParquetReader. 
- withConfiguration(Map<String, String>) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.ReadFiles
- 
Specify Hadoop configuration for ParquetReader. 
- withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.ReadFiles
- 
Specify Hadoop configuration for ParquetReader. 
- withConfiguration(Map<String, String>) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Sink
- 
Specify Hadoop configuration for ParquetWriter. 
- withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Sink
- 
Specify Hadoop configuration for ParquetWriter. 
- withConfigurationTransform(PTransform<PCollection<? extends KV<KeyT, ValueT>>, PCollectionView<Configuration>>) - Method in interface org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write.WriteBuilder
- 
Writes to the sink using configuration created by provided configurationTransformation.
 
- withConfigUrl(String) - Method in class org.apache.beam.sdk.managed.Managed.ManagedTransform
- 
- withConnectionConfiguration(ElasticsearchIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BulkIO
- 
Provide the Elasticsearch connection configuration object. 
- withConnectionConfiguration(ElasticsearchIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.DocToBulk
- 
Provide the Elasticsearch connection configuration object. 
- withConnectionConfiguration(ElasticsearchIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
- 
Provide the Elasticsearch connection configuration object. 
- withConnectionConfiguration(ElasticsearchIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withConnectionConfiguration(MqttIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Read
- 
Define the MQTT connection configuration used to connect to the MQTT broker. 
- withConnectionConfiguration(MqttIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Write
- 
Define MQTT connection configuration used to connect to the MQTT broker. 
- withConnectionConfiguration(RedisConnectionConfiguration) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
-  
- withConnectionConfiguration(RedisConnectionConfiguration) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadKeyPatterns
-  
- withConnectionConfiguration(RedisConnectionConfiguration) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
-  
- withConnectionConfiguration(RedisConnectionConfiguration) - Method in class org.apache.beam.sdk.io.redis.RedisIO.WriteStreams
- 
- withConnectionConfiguration(SolrIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Read
- 
Provide the Solr connection configuration object. 
- withConnectionConfiguration(SolrIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Write
- 
Provide the Solr connection configuration object. 
- withConnectionFactory(ConnectionFactory) - Method in interface org.apache.beam.sdk.io.jms.JmsIO.ConnectionFactoryContainer
-  
- withConnectionFactory(ConnectionFactory) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
- 
Specify the JMS connection factory to connect to the JMS broker. 
- withConnectionFactory(ConnectionFactory) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
- 
Specify the JMS connection factory to connect to the JMS broker. 
- withConnectionFactoryProviderFn(SerializableFunction<Void, ? extends ConnectionFactory>) - Method in interface org.apache.beam.sdk.io.jms.JmsIO.ConnectionFactoryContainer
-  
- withConnectionFactoryProviderFn(SerializableFunction<Void, ? extends ConnectionFactory>) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
- 
Specify a JMS connection factory provider function to connect to the JMS broker. 
- withConnectionFactoryProviderFn(SerializableFunction<Void, ? extends ConnectionFactory>) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
- 
Specify a JMS connection factory provider function to connect to the JMS broker. 
- withConnectionInitSqls(Collection<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
- 
Sets the connection init sql statements to driver.connect(...). 
- withConnectionInitSqls(ValueProvider<Collection<String>>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
- 
- withConnectionProperties(Map<String, String>) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Sets a custom property to be used within the connection to your database. 
- withConnectionProperties(ValueProvider<Map<String, String>>) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Sets a custom property to be used within the connection to your database. 
- withConnectionProperties(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
- 
Sets the connection properties passed to driver.connect(...). 
- withConnectionProperties(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
- 
- withConnectionProperties(String) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.DataSourceConfiguration
- 
Sets the connection properties passed to driver.connect(...). 
- withConnectionProperty(String, String) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Sets a custom property to be used within the connection to your database. 
- withConnectorClass(Class<?>) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Applies the connectorClass to be used to connect to your database. 
- withConnectorClass(ValueProvider<Class<?>>) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Sets the connectorClass to be used to connect to your database. 
- withConnectorConfiguration(DebeziumIO.ConnectorConfiguration) - Method in class org.apache.beam.io.debezium.DebeziumIO.Read
- 
Applies the given configuration to the connector. 
- withConnectTimeout(Integer) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the Cassandra client connect timeout in ms. 
- withConnectTimeout(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the Cassandra client connect timeout in ms. 
- withConnectTimeout(Integer) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Cassandra client socket option for connect timeout in ms. 
- withConnectTimeout(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Cassandra client socket option for connect timeout in ms. 
- withConnectTimeout(Integer) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
If set, overwrites the default connect timeout (1000ms) in the  RequestConfig of the
 Elastic  RestClient. 
- withConsistencyLevel(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the consistency level for the request (e.g. 
- withConsistencyLevel(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the consistency level for the request (e.g. 
- withConsistencyLevel(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Specify the consistency level for the request (e.g. 
- withConsistencyLevel(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Specify the consistency level for the request (e.g. 
- withConsistencyLevel(InfluxDB.ConsistencyLevel) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Write
- 
Sets the consistency level to use. 
- withConstructorArgs(Object...) - Method in class org.apache.beam.sdk.io.sparkreceiver.ReceiverBuilder
- 
- withConsumerArn(String) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specify consumer ARN to enable Enhanced Fan-Out. 
- withConsumerConfigOverrides(Map<String, Object>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Replaces the configuration for the main consumer. 
- withConsumerConfigUpdates(Map<String, Object>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Update configuration for the backend main consumer. 
- withConsumerConfigUpdates(Map<String, Object>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Updates configuration for the main consumer. 
- withConsumerFactoryFn(SerializableFunction<Map<String, Object>, Consumer<byte[], byte[]>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
A factory to create Kafka  Consumer from consumer configuration. 
- withConsumerFactoryFn(SerializableFunction<Map<String, Object>, Consumer<byte[], byte[]>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
A factory to create Kafka  Consumer from consumer configuration. 
- withConsumerFactoryFn(SerializableFunction<Map<String, Object>, ? extends Consumer<?, ?>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
- 
- withConsumerFactoryFn(SerializableFunction<Map<String, Object>, ? extends Consumer<?, ?>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
- 
- withConsumerPollingTimeout(long) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Sets the timeout time in seconds for Kafka consumer polling request in the ReadFromKafkaDoFn.
 
- withConsumerPollingTimeout(long) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Sets the timeout time in seconds for Kafka consumer polling request in the ReadFromKafkaDoFn.
 
- withContainer(String) - Method in class org.apache.beam.sdk.io.azure.cosmos.CosmosIO.Read
- 
Specify the Cosmos container to read from. 
- withContentTypeHint(String) - Method in class org.apache.beam.sdk.io.tika.TikaIO.ParseFiles
- 
Sets a content type hint to make the file parser detection more efficient. 
- withCPUCount(int) - Method in class org.apache.beam.sdk.transforms.resourcehints.ResourceHints
- 
Sets desired minimal CPU or vCPU count to have in transform's execution environment. 
- withCreateDisposition(BigQueryIO.Write.CreateDisposition) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Specifies whether the table should be created if it does not exist. 
- withCreateDisposition(CreateDisposition) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
A disposition to be used during table preparation. 
- withCreateOrUpdateMetadataTable(boolean) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
- withCreateTime(Duration) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
- withCreateTime() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Use the creation time of  KafkaRecord as the output timestamp. 
- withCreateTime(Duration) - Static method in interface org.apache.beam.sdk.io.kafka.TimestampPolicyFactory
- 
- withCreatWatermarkEstimatorFn(SerializableFunction<Instant, WatermarkEstimator<Instant>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
- withCredentials(Credentials) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the credentials. 
- withCredentials(ValueProvider<Credentials>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the credentials. 
- withCsvMapper(SnowflakeIO.CsvMapper<T>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
- 
User-defined function mapping CSV lines into user data. 
- withCustomBeamRequirement(String) - Method in class org.apache.beam.sdk.extensions.python.PythonService
- 
Override the Beam version to be installed in the service environment. 
- withCustomGcsTempLocation(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Provides a custom location on GCS for storing temporary files to be loaded via BigQuery batch
 load jobs. 
- withCustomRateLimitPolicy(RateLimitPolicyFactory) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specifies the RateLimitPolicyFactoryfor a custom rate limiter.
 
- withCustomRecordParsing(String, SerializableFunction<String, OutputT>) - Method in class org.apache.beam.sdk.io.csv.CsvIOParse
- 
Configures custom cell parsing. 
- withCustomWatermarkPolicy(WatermarkPolicyFactory) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specifies the WatermarkPolicyFactoryas a custom watermarkPolicyFactory.
 
- withCustomWatermarkPolicy(WatermarkParameters) - Static method in interface org.apache.beam.sdk.io.aws2.kinesis.WatermarkPolicyFactory
- 
Returns an custom WatermarkPolicyFactory. 
- withCypher(String) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.ReadAll
-  
- withCypher(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.ReadAll
-  
- withCypher(String) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- withCypher(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- withCypherLogging() - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.ReadAll
-  
- withCypherLogging() - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- withDatabase(String) - Method in class org.apache.beam.sdk.io.azure.cosmos.CosmosIO.Read
- 
Specify the Cosmos database to read from. 
- withDatabase(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
- 
Sets the database name. 
- withDatabase(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
- 
Sets the database name. 
- withDatabase(String) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Read
- 
Reads from the specified database. 
- withDatabase(String) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Write
- 
Writes to the specified database. 
- withDatabase(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-  
- withDatabase(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-  
- withDatabase(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
- 
Sets the database to use. 
- withDatabase(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
- 
Sets the database to use. 
- withDatabase(String) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.DataSourceConfiguration
-  
- withDatabase(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets database to use. 
- withDatabase(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
- 
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
- 
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntityWithSummary
- 
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntityWithSummary
- 
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
- 
Returns a new  DatastoreV1.DeleteKey that deletes entities from the Cloud Datastore for the
 specified database. 
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
- 
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKeyWithSummary
- 
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKeyWithSummary
- 
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
- 
Returns a new  DatastoreV1.Read that reads from the Cloud Datastore for the specified
 database. 
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
- 
Returns a new  DatastoreV1.Write that writes to the Cloud Datastore for the database id. 
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
- 
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.WriteWithSummary
- 
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.WriteWithSummary
- 
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the Cloud Spanner database ID. 
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the Cloud Spanner database ID. 
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
- 
Specifies the Cloud Spanner database. 
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
- 
Specifies the Cloud Spanner database. 
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
- 
Specifies the Cloud Spanner database. 
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
- 
Specifies the Cloud Spanner database. 
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
- 
Specifies the Cloud Spanner database. 
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
- 
Specifies the Cloud Spanner database. 
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
Specifies the Cloud Spanner database. 
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
Specifies the Cloud Spanner database. 
- withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the Cloud Spanner database. 
- withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the Cloud Spanner database. 
- withDatabaseRole(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the Cloud Spanner database role. 
- withDataBoostEnabled(ValueProvider<Boolean>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies if the pipeline has to be run on the independent compute resource. 
- withDatasetService(FakeDatasetService) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
-  
- withDataSourceConfiguration(InfluxDbIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Read
- 
Reads from the InfluxDB instance indicated by the given configuration. 
- withDataSourceConfiguration(InfluxDbIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Write
- 
Writes to the InfluxDB instance indicated by the given configuration. 
- withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-  
- withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-  
- withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
-  
- withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadWithPartitions
-  
- withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
- 
See WriteVoid#withDataSourceConfiguration(DataSourceConfiguration).
 
- withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
-  
- withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteWithResults
-  
- withDataSourceConfiguration(SingleStoreIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.Read
-  
- withDataSourceConfiguration(SingleStoreIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.ReadWithPartitions
-  
- withDataSourceConfiguration(SingleStoreIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.Write
-  
- withDataSourceConfiguration(SnowflakeIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
- 
Setting information about Snowflake server. 
- withDataSourceConfiguration(SnowflakeIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
Setting information about Snowflake server. 
- withDataSourceProviderFn(SerializableFunction<Void, DataSource>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-  
- withDataSourceProviderFn(SerializableFunction<Void, DataSource>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-  
- withDataSourceProviderFn(SerializableFunction<Void, DataSource>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
-  
- withDataSourceProviderFn(SerializableFunction<Void, DataSource>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadWithPartitions
-  
- withDataSourceProviderFn(SerializableFunction<Void, DataSource>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
- 
- withDataSourceProviderFn(SerializableFunction<Void, DataSource>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
-  
- withDataSourceProviderFn(SerializableFunction<Void, DataSource>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteWithResults
-  
- withDataSourceProviderFn(SerializableFunction<Void, DataSource>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
- 
- withDataSourceProviderFn(SerializableFunction<Void, DataSource>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
- withDatumReaderFactory(AvroSource.DatumReaderFactory<T>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Read
- 
- withDatumReaderFactory(AvroSource.DatumReaderFactory<T>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadFiles
- 
- withDatumReaderFactory(AvroSource.DatumReaderFactory<?>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
- 
- withDatumWriterFactory(AvroSink.DatumWriterFactory<ElementT>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Sink
- 
- withDatumWriterFactory(AvroSink.DatumWriterFactory<OutputT>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
- withDatumWriterFactory(AvroSink.DatumWriterFactory<T>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- withDdlString(String) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
-  
- withDeadLetterQueue() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.BatchWriteWithSummary.Builder
-  
- withDeadLetterTopic(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
- 
Creates and returns a transform for writing read failures out to a dead-letter topic. 
- withDeadLetterTopic(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
- 
- withDebugMode(StreamingLogLevel) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
The option to verbose info (or only errors) of loaded files while streaming. 
- withDecompression(CompressedSource.DecompressingChannelFactory) - Method in class org.apache.beam.sdk.io.CompressedSource
- 
- withDeduplicateKeys(List<String>) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Write
-  
- withDeduplicateRecords(boolean) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Read
- 
Optional, default: false. 
- WithDefault() - Constructor for class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
-  
- withDefaultConfig(boolean) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverConfiguration
-  
- withDefaultConfig(ValueProvider<Boolean>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverConfiguration
-  
- withDefaultHeaders(Header[]) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
For authentication or custom requirements, provide a set if default headers for the client. 
- withDefaultMissingValueInterpretation(AppendRowsRequest.MissingValueInterpretation) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Specify how missing values should be interpreted when there is a default value in the schema. 
- withDefaultRateLimiter() - Static method in interface org.apache.beam.sdk.io.aws2.kinesis.RateLimitPolicyFactory
-  
- withDefaultRateLimiter(Duration, Duration, Duration) - Static method in interface org.apache.beam.sdk.io.aws2.kinesis.RateLimitPolicyFactory
-  
- withDefaultTableProvider(String, TableProvider) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
-  
- withDefaultThreadPool() - Static method in class org.apache.beam.runners.dataflow.util.PackageUtil
-  
- withDefaultValue(T) - Method in class org.apache.beam.sdk.transforms.View.AsSingleton
- 
Default value to return for windows with no value in them. 
- withDelay(Supplier<Duration>) - Static method in interface org.apache.beam.sdk.io.aws2.kinesis.RateLimitPolicyFactory
-  
- withDelimiter(byte[]) - Method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.Read
- 
Set the custom delimiter to be used in place of the default ones ('\r', '\n' or '\r\n'). 
- withDelimiter(byte[]) - Method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.ReadFiles
- 
- withDelimiter(byte[]) - Method in class org.apache.beam.sdk.io.TextIO.Read
- 
Set the custom delimiter to be used in place of the default ones ('\r', '\n' or '\r\n'). 
- withDelimiter(byte[]) - Method in class org.apache.beam.sdk.io.TextIO.ReadFiles
- 
Like Read#withDelimiter.
 
- withDelimiter(char[]) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Specifies the delimiter after each string written. 
- withDelimiter(char[]) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withDeliveryMode(DeliveryMode) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Write
- 
Set the delivery mode. 
- withDescription(String) - Method in class org.apache.beam.sdk.schemas.Schema.Field
- 
Returns a copy of the Field with the description set. 
- withDesiredBundleSizeBytes(long) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ParseAll
- 
Deprecated. Set a value for the bundle size for parallel reads. 
- withDesiredBundleSizeBytes(long) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ParseFiles
- 
Set a value for the bundle size for parallel reads. 
- withDesiredBundleSizeBytes(long) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadAll
- 
Deprecated. Set a value for the bundle size for parallel reads. 
- withDesiredBundleSizeBytes(long) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadFiles
- 
Set a value for the bundle size for parallel reads. 
- withDestinationCoder(Coder<DestinationT>) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- withDeterministicRecordIdFn(SerializableFunction<T, String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-  
- withDeveloperToken(String) - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17.Read
- 
Creates and returns a new  GoogleAdsV17.Read transform with the specified developer token. 
- withDeveloperToken(String) - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17.ReadAll
- 
- withDialectView(PCollectionView<Dialect>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-  
- withDirectExecutor() - Method in class org.apache.beam.sdk.fn.channel.ManagedChannelFactory
- 
Returns a  ManagedChannelFactory like this one, but will construct the channel to use
 the direct executor. 
- withDirectoryTreatment(FileIO.ReadMatches.DirectoryTreatment) - Method in class org.apache.beam.sdk.io.FileIO.ReadMatches
- 
Controls how to handle directories in the input  PCollection. 
- withDirectWriteProtos(boolean) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-  
- withDisableAutoCommit(boolean) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
- 
Whether to disable auto commit on read. 
- withDisableAutoCommit(boolean) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
- 
Whether to disable auto commit on read. 
- withDisableAutoCommit(boolean) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
- 
Whether to disable auto commit on read. 
- withDisableAutoCommit(boolean) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadWithPartitions
- 
Whether to disable auto commit on read. 
- withDisableCertificateValidation(boolean) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Read
- 
Disable SSL certification validation. 
- withDisableCertificateValidation(boolean) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Write
- 
Disable SSL certification validation. 
- withDisableCertificateValidation(ValueProvider<Boolean>) - Method in class org.apache.beam.sdk.io.splunk.SplunkIO.Write
- 
- withDisableCertificateValidation(Boolean) - Method in class org.apache.beam.sdk.io.splunk.SplunkIO.Write
- 
Disables ssl certificate validation. 
- withDocVersionFn(ElasticsearchIO.Write.FieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.DocToBulk
- 
Provide a function to extract the doc version from the document. 
- withDocVersionFn(ElasticsearchIO.Write.FieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withDocVersionType(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.DocToBulk
- 
Provide a function to extract the doc version from the document. 
- withDocVersionType(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withDriverClassLoader(ClassLoader) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
- 
Sets the class loader instance to be used to load the JDBC driver. 
- withDriverConfiguration(Neo4jIO.DriverConfiguration) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.ReadAll
-  
- withDriverConfiguration(Neo4jIO.DriverConfiguration) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- withDriverJars(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
- 
Comma separated paths for JDBC drivers. 
- withDriverJars(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
- 
- withDuration(Duration) - Method in class org.apache.beam.sdk.transforms.Deduplicate.KeyedValues
- 
Returns a  KeyedValuesPTransform like this one but with the specified
 duration. 
- withDuration(Duration) - Method in class org.apache.beam.sdk.transforms.Deduplicate.Values
- 
Returns a  ValuesPTransform like this one but with the specified duration. 
- withDuration(Duration) - Method in class org.apache.beam.sdk.transforms.Deduplicate.WithRepresentativeValues
- 
Return a  WithRepresentativeValuesPTransform that is like this one, but with
 the specified deduplication duration. 
- withDynamicDelayRateLimitPolicy(Supplier<Duration>) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specifies a dynamic delay rate limit policy with the given function being called at each
 polling interval to get the next delay value. 
- withDynamicRead(Duration) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Configure the KafkaIO to use  WatchForKafkaTopicPartitions to detect and emit any new
 available  TopicPartition for  ReadFromKafkaDoFn to consume during pipeline
 execution time. 
- withEarlyFirings(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-  
- withEarlyFirings(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
- 
Creates a new Triggerlike the this, except that it fires repeatedly whenever the
 givenTriggerfires before the watermark has passed the end of the window.
 
- withElementTimestamp() - Static method in interface org.apache.beam.sdk.io.kafka.KafkaPublishTimestampFunction
- 
- withEmptyGlobalWindowDestination(DestinationT) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Parse
- 
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ParseAll
- 
Deprecated. 
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Read
- 
Configures whether or not a filepattern matching no files is allowed. 
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadAll
- 
Deprecated. 
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
-  
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.Read
- 
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.FileIO.Match
- 
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.FileIO.MatchAll
- 
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
- 
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.TextIO.Read
- 
- withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
- 
Deprecated. Same as Read#withEmptyMatchTreatment.
 
- withEmulator(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableConfig
-  
- withEmulator(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
- withEmulator(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withEmulatorHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the Cloud Spanner host, when an emulator is used. 
- withEmulatorHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
- 
Specifies the Cloud Spanner emulator host. 
- withEmulatorHost(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
-  
- withEmulatorHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
- 
Specifies the Cloud Spanner emulator host. 
- withEmulatorHost(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withEmulatorHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
- 
Specifies the Cloud Spanner emulator host. 
- withEmulatorHost(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-  
- withEmulatorHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the Cloud Spanner emulator host. 
- withEmulatorHost(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-  
- withEnableBatchLogs(ValueProvider<Boolean>) - Method in class org.apache.beam.sdk.io.splunk.SplunkIO.Write
- 
Same as  Builder#withEnableBatchLogs(ValueProvider) but without a  ValueProvider. 
- withEnableBatchLogs(Boolean) - Method in class org.apache.beam.sdk.io.splunk.SplunkIO.Write
- 
Method to enable batch logs. 
- withEnableGzipHttpCompression(ValueProvider<Boolean>) - Method in class org.apache.beam.sdk.io.splunk.SplunkIO.Write
- 
Same as  Builder#withEnableGzipHttpCompression(ValueProvider) but without a  ValueProvider. 
- withEnableGzipHttpCompression(Boolean) - Method in class org.apache.beam.sdk.io.splunk.SplunkIO.Write
- 
Method to specify if HTTP requests sent to Splunk should be GZIP encoded. 
- withEndKey(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
- 
Returns new  ByteKeyRange like this one, but with the specified end key. 
- withEndMessageId(MessageId) - Method in class org.apache.beam.sdk.io.pulsar.PulsarIO.Read
-  
- withEndpoint(String, int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
-  
- withEndpoint(String, int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadKeyPatterns
-  
- withEndpoint(String, int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
-  
- withEndpoint(String, int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.WriteStreams
- 
Set the hostname and port of the Redis server to connect to. 
- withEndTimestamp(Long) - Method in class org.apache.beam.sdk.io.pulsar.PulsarIO.Read
-  
- withEntity(Class<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the entity class (annotated POJO). 
- withEntity(Class<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
- withEntries - Variable in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
-  
- withEntryMapper(SqsIO.WriteBatches.EntryMapperFn<T>) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.WriteBatches
- 
Optional mapper to create a batch entry from a unique entry id and the input T,
 otherwise inferred from the schema.
 
- withEntryMapper(SqsIO.WriteBatches.EntryMapperFn.Builder<T>) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.WriteBatches
- 
Optional mapper to create a batch entry from the input Tusing a builder, otherwise
 inferred from the schema.
 
- withEnvironmentId(String) - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- withEOS(int, String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
- 
- withEOS(int, String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
- 
Provides exactly-once semantics while writing to Kafka, which enables applications with
 end-to-end exactly-once guarantees on top of exactly-once semantics within Beam
 pipelines. 
- withEpsilon(double) - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
- 
Returns an ApproximateQuantilesCombineFnthat's like this one except that it uses the
 specifiedepsilonvalue.
 
- withErrorHandler(ErrorHandler<BadRecord, ?>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-  
- withErrorHandler(ErrorHandler<BadRecord, ?>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-  
- withErrorHandler(ErrorHandler<BadRecord, ?>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
-  
- withErrorHandler(ErrorHandler<BadRecord, ?>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
- 
Configures the PubSub read with an alternate error handler. 
- withErrorHandler(ErrorHandler<BadRecord, ?>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
- 
Writes any serialization failures out to the Error Handler. 
- withErrorHandler(ErrorHandler<BadRecord, ?>) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Write
- 
An optional error handler for handling records that failed to publish to Solace. 
- withErrorsTransformer(PTransform<PCollection<Row>, POutput>) - Method in interface org.apache.beam.sdk.extensions.sql.impl.rel.BeamRelNode
-  
- withErrorsTransformer(PTransform<PCollection<Row>, ? extends POutput>) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
-  
- withEvent(String) - Method in class org.apache.beam.sdk.io.splunk.SplunkEvent.Builder
- 
Assigns the event payload to be sent to the HEC endpoint. 
- withEventStore(String) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportUserEvents
-  
- withEventStore(String) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIPredict
-  
- withEventStore(String) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIWriteUserEvent
-  
- withEverythingCounted() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
Turns on all monitoring. 
- withEverythingCountedExceptedCaching() - Method in class org.apache.beam.io.requestresponse.Monitoring
- 
Turns on all monitoring except for cache related metrics. 
- withExceptionReporting(Schema) - Static method in class org.apache.beam.sdk.transforms.JsonToRow
- 
Enable Exception Reporting support. 
- withExchange(String, String, String) - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Read
- 
In AMQP, messages are published to an exchange and routed to queues based on the exchange
 type and a queue binding. 
- withExchange(String, String) - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Read
- 
In AMQP, messages are published to an exchange and routed to queues based on the exchange
 type and a queue binding. 
- withExchange(String, String) - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Write
- 
Defines the to-be-declared exchange where the messages will be sent. 
- withExchange(String) - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Write
- 
Defines the existing exchange where the messages will be sent. 
- withExecuteStreamingSqlRetrySettings(RetrySettings) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the ExecuteStreamingSql retry settings. 
- withExecutorService(ExecutorService) - Static method in class org.apache.beam.runners.dataflow.util.PackageUtil
-  
- withExistingPipelineOptions(BigtableIO.ExistingPipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
- withExpansionService(String) - Method in class org.apache.beam.sdk.extensions.python.transforms.DataframeTransform
- 
Sets an expansion service endpoint for DataframeTransform. 
- withExpansionService(String) - Method in class org.apache.beam.sdk.extensions.python.transforms.PythonMap
-  
- withExpansionService(String) - Method in class org.apache.beam.sdk.extensions.python.transforms.RunInference
- 
Sets an expansion service endpoint for RunInference. 
- withExpireTime(Long) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
-  
- withExtendedErrorInfo() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- withExtendedErrorInfo(boolean) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StreamingInserts
- 
Specify whether to use extended error info or not. 
- withExtendedErrorInfo() - Method in class org.apache.beam.sdk.transforms.JsonToRow.JsonToRowWithErrFn
- 
Adds the error message to the returned error Row. 
- withExtensionsFrom(Iterable<Class<?>>) - Method in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
- 
Returns a  DynamicProtoCoder like this one, but with the extensions from the given
 classes registered. 
- withExtensionsFrom(Iterable<Class<?>>) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
- 
Returns a  ProtoCoder like this one, but with the extensions from the given classes
 registered. 
- withExtensionsFrom(Class<?>...) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
- 
- withExternalSorterType(ExternalSorter.Options.SorterType) - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
- 
Sets the external sorter type. 
- withExternalSynchronization(ExternalSynchronization) - Method in interface org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write.ExternalSynchronizationBuilder
- 
Specifies class which will provide external synchronization required for hadoop write
 operation. 
- withExtractOutputTimestampFn(SerializableFunction<KafkaRecord<K, V>, Instant>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
A function to calculate output timestamp for a given  KafkaRecord. 
- withExtractOutputTimestampFn(SerializableFunction<Message<byte[]>, Instant>) - Method in class org.apache.beam.sdk.io.pulsar.PulsarIO.Read
-  
- withExtraPackages(List<String>) - Method in class org.apache.beam.sdk.extensions.python.PythonExternalTransform
- 
Specifies that the given Python packages are required for this transform, which will cause them
 to be installed in both the construction-time and execution time environment. 
- withExtraPackages(List<String>) - Method in class org.apache.beam.sdk.extensions.python.PythonService
- 
Specifies that the given Python packages should be installed for this service environment. 
- withExtraPackages(List<String>) - Method in class org.apache.beam.sdk.extensions.python.transforms.PythonMap
- 
Specifies any extra packages required by the Python function. 
- withExtraPackages(List<String>) - Method in class org.apache.beam.sdk.extensions.python.transforms.RunInference
- 
Specifies any extra packages required by the RunInference model handler. 
- withFailedInsertRetryPolicy(InsertRetryPolicy) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Specifies a policy for handling failed inserts. 
- withFailureMode(SpannerIO.FailureMode) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies failure mode. 
- WithFailures - Class in org.apache.beam.sdk.transforms
- 
A collection of utilities for writing transforms that can handle exceptions raised during
 processing of elements. 
- WithFailures() - Constructor for class org.apache.beam.sdk.transforms.WithFailures
-  
- WithFailures.ExceptionAsMapHandler<T> - Class in org.apache.beam.sdk.transforms
- 
A simple handler that extracts information from an exception to a  Map<String, String>
 and returns a  KV where the key is the input element that failed processing, and the
 value is the map of exception attributes. 
- WithFailures.ExceptionElement<T> - Class in org.apache.beam.sdk.transforms
- 
The value type passed as input to exception handlers. 
- WithFailures.Result<OutputT extends POutput,FailureElementT> - Class in org.apache.beam.sdk.transforms
- 
An intermediate output type for PTransforms that allows an output collection to live alongside
 a collection of elements that failed the transform. 
- WithFailures.ThrowableHandler<T> - Class in org.apache.beam.sdk.transforms
- 
A handler that holds onto the  Throwable that led to the exception, returning it along
 with the original value as a  KV. 
- withFanout(int) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
-  
- withFanout(int) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineGlobally
-  
- withFanout(int) - Method in class org.apache.beam.sdk.transforms.Combine.Globally
- 
Returns a  PTransform identical to this, but that uses an intermediate node to combine
 parts of the data to reduce load on the final global combine step. 
- withFaultTolerent(boolean) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Read
- 
Instructs the read scan to resume a scan on another tablet server if the current server fails
 and faultTolerant is set to true. 
- withFetchSize(int) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
- 
This method is used to set the size of the data that is going to be fetched and loaded in
 memory per every database call. 
- withFetchSize(int) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
- 
This method is used to set the size of the data that is going to be fetched and loaded in
 memory per every database call. 
- withFetchSize(int) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
- 
This method is used to set the size of the data that is going to be fetched and loaded in
 memory per every database call. 
- withFetchSize(int) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadWithPartitions
- 
The number of rows to fetch from the database in the same ResultSetround-trip.
 
- withFieldAccessDescriptors(Map<FieldAccessDescriptor, Object>) - Method in class org.apache.beam.sdk.values.Row.FieldValueBuilder
- 
Sets field values using the FieldAccessDescriptors. 
- withFieldIds(FieldAccessDescriptor, Integer...) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Return a descriptor that accesses the specified field ids as nested subfields of the
 baseDescriptor. 
- withFieldIds(FieldAccessDescriptor, Iterable<Integer>) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Return a descriptor that accesses the specified field ids as nested subfields of the
 baseDescriptor. 
- withFieldIds(Integer...) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Return a descriptor that access the specified fields. 
- withFieldIds(Iterable<Integer>) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Return a descriptor that access the specified fields. 
- withFieldNameAs(String, String) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Add a field with a new name. 
- withFieldNameAs(String, String) - Method in class org.apache.beam.sdk.schemas.transforms.Select.Fields
- 
Add a single field to the selection, along with the name the field should take in the
 selected schema. 
- withFieldNameAs(String, String) - Method in class org.apache.beam.sdk.schemas.transforms.Select.Flattened
- 
Allows renaming a specific nested field. 
- withFieldNames(String...) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Return a descriptor that access the specified fields. 
- withFieldNames(Iterable<String>) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Return a descriptor that access the specified fields. 
- withFieldNames(FieldAccessDescriptor, String...) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Return a descriptor that accesses the specified field names as nested subfields of the
 baseDescriptor. 
- withFieldNames(FieldAccessDescriptor, Iterable<String>) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Return a descriptor that accesses the specified field names as nested subfields of the
 baseDescriptor. 
- withFieldNamesAs(Map<String, String>) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Return a descriptor that accesses the specified fields, renaming those fields. 
- withFieldReordering() - Method in enum org.apache.beam.sdk.extensions.sql.meta.ProjectSupport
-  
- withFields(JsonObject) - Method in class org.apache.beam.sdk.io.splunk.SplunkEvent.Builder
- 
Assigns fields value to the event metadata. 
- withFields(FieldAccessDescriptor.FieldDescriptor...) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
- withFields(Iterable<FieldAccessDescriptor.FieldDescriptor>) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
- withFieldValue(String, Object) - Method in class org.apache.beam.sdk.values.Row.Builder
- 
Set a field value using the field name. 
- withFieldValue(Integer, Object) - Method in class org.apache.beam.sdk.values.Row.Builder
- 
Set a field value using the field id. 
- withFieldValue(FieldAccessDescriptor, Object) - Method in class org.apache.beam.sdk.values.Row.Builder
- 
Set a field value using a FieldAccessDescriptor. 
- withFieldValue(String, Object) - Method in class org.apache.beam.sdk.values.Row.FieldValueBuilder
- 
Set a field value using the field name. 
- withFieldValue(Integer, Object) - Method in class org.apache.beam.sdk.values.Row.FieldValueBuilder
- 
Set a field value using the field id. 
- withFieldValue(FieldAccessDescriptor, Object) - Method in class org.apache.beam.sdk.values.Row.FieldValueBuilder
- 
Set a field value using a FieldAccessDescriptor. 
- withFieldValueGetters(Factory<List<FieldValueGetter<T, Object>>>, T) - Method in class org.apache.beam.sdk.values.Row.Builder
-  
- withFieldValues(Map<String, Object>) - Method in class org.apache.beam.sdk.values.Row.Builder
- 
Sets field values using the field names. 
- withFieldValues(Map<String, Object>) - Method in class org.apache.beam.sdk.values.Row.FieldValueBuilder
- 
Sets field values using the field names. 
- withFileExceptionHandler(ReadAllViaFileBasedSource.ReadFileRangesFnExceptionHandler) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ParseFiles
- 
Specifies if exceptions should be logged only for streaming pipelines. 
- withFileExceptionHandler(ReadAllViaFileBasedSource.ReadFileRangesFnExceptionHandler) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadFiles
- 
Specifies if exceptions should be logged only for streaming pipelines. 
- withFilename(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-  
- withFileNameTemplate(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
A template name for files saved to GCP. 
- withFilter(Filter) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
- 
Filters the rows read from HBase using the given* row filter. 
- withFilter(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
- 
Sets the filter details. 
- withFilter(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-  
- withFilters(Bson) - Method in class org.apache.beam.sdk.io.mongodb.FindQuery
- 
Sets the filters to find. 
- withFindKey(String) - Method in class org.apache.beam.sdk.io.mongodb.UpdateConfiguration
- 
Sets the filters to find. 
- withFixedDelay() - Static method in interface org.apache.beam.sdk.io.aws2.kinesis.RateLimitPolicyFactory
-  
- withFixedDelay(Duration) - Static method in interface org.apache.beam.sdk.io.aws2.kinesis.RateLimitPolicyFactory
-  
- withFixedDelayRateLimitPolicy() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specifies a fixed delay rate limit policy with the default delay of 1 second. 
- withFixedDelayRateLimitPolicy(Duration) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specifies a fixed delay rate limit policy with the given delay. 
- withFlowControl(boolean) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
Returns a new  BigtableIO.Write with flow control enabled if enableFlowControl is
 true. 
- withFlushRowLimit(Integer) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
Sets number of row limit that will be saved to the staged file and then loaded to Snowflake. 
- withFlushTimeLimit(Duration) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
Sets duration how often staged files will be created and then how often ingested by Snowflake
 during streaming. 
- withFooter(String) - Method in class org.apache.beam.sdk.io.TextIO.Sink
-  
- withFooter(String) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Adds a footer string to each file. 
- withFooter(String) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withFormat(DataFormat) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
See DataFormat.
 
- withFormatFn(KuduIO.FormatFunction<T>) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Write
- 
Writes using the given function to create the mutation operations from the input. 
- withFormatFunction(SourceRecordMapper<T>) - Method in class org.apache.beam.io.debezium.DebeziumIO.Read
- 
- withFormatFunction(SerializableFunction<UserT, OutputT>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
Specifies a format function to convert UserTto the output type.
 
- withFormatFunction(SerializableFunction<T, TableRow>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Formats the user's type into a  TableRow to be written to BigQuery. 
- withFormatFunction(SerializableFunction<UserT, String>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
- withFormatRecordOnFailureFunction(SerializableFunction<T, TableRow>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
If an insert failure occurs, this function is applied to the originally supplied T element. 
- withFromDateTime(String) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Read
- 
Read metric data from the fromDateTime. 
- withGapDuration(Duration) - Static method in class org.apache.beam.sdk.transforms.windowing.Sessions
- 
Creates a  SessionsWindowFn with the specified gap duration. 
- withGCPApplicationDefaultCredentials() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Creates and sets the Application Default Credentials for a Kafka consumer. 
- withGCPApplicationDefaultCredentials() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
- 
Creates and sets the Application Default Credentials for a Kafka producer. 
- withGetOffsetFn(SerializableFunction<V, Long>) - Method in class org.apache.beam.sdk.io.sparkreceiver.SparkReceiverIO.Read
- 
A function to get offset in order to start Receiverfrom it.
 
- withGoogleAdsClientFactory(GoogleAdsClientFactory) - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17.Read
- 
Creates and returns a new  GoogleAdsV17.Read transform with the specified client factory. 
- withGoogleAdsClientFactory(GoogleAdsClientFactory) - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17.ReadAll
- 
- withGroupingFactor(int) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the multiple of max mutation (in terms of both bytes per batch and cells per batch)
 that is used to select a set of mutations to sort by key for batching. 
- withHadoopConfiguration(Class<K>, Class<V>) - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Sets a plugin Hadoop configuration. 
- withHadoopConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.cdap.Plugin
- 
Sets a plugin Hadoop configuration. 
- withHasError(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Document
- 
Used to set whether or not there was an error for a given document as indicated by the
 response from Elasticsearch. 
- withHasMultilineCSVRecords(Boolean) - Method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.Read
- 
When reading RFC4180 CSV files that have values that span multiple lines, set this to true. 
- withHeader(String) - Method in class org.apache.beam.sdk.io.TextIO.Sink
-  
- withHeader(String) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Adds a header string to each file. 
- withHeader(String) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withHighPriority() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withHighPriority() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-  
- withHighPriority() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-  
- withHint(String, ResourceHint) - Method in class org.apache.beam.sdk.transforms.resourcehints.ResourceHints
- 
Declares a custom resource hint that has a specified URN. 
- withHintMatchesManyFiles() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Parse
- 
- withHintMatchesManyFiles() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Read
- 
- withHintMatchesManyFiles() - Method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.Read
- 
- withHintMatchesManyFiles() - Method in class org.apache.beam.sdk.io.TextIO.Read
- 
- withHintMaxNumWorkers(int) - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions.Builder
- 
Provide a hint to the QoS system for the intended max number of workers for a pipeline. 
- withHintNumWorkers(int) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
- 
- withHintNumWorkers(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
- 
- withHintNumWorkers(int) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntityWithSummary
- 
- withHintNumWorkers(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntityWithSummary
- 
- withHintNumWorkers(int) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
- 
- withHintNumWorkers(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
- 
- withHintNumWorkers(int) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKeyWithSummary
- 
- withHintNumWorkers(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKeyWithSummary
- 
- withHintNumWorkers(int) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
- 
Returns a new  DatastoreV1.Write with a different worker count hint for ramp-up throttling. 
- withHintNumWorkers(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
- 
- withHintNumWorkers(int) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.WriteWithSummary
- 
- withHintNumWorkers(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.WriteWithSummary
- 
- withHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the Cloud Spanner host. 
- withHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
- 
Specifies the Cloud Spanner host. 
- withHost(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
-  
- withHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
- 
Specifies the Cloud Spanner host. 
- withHost(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
- 
Specifies the Cloud Spanner host. 
- withHost(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-  
- withHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the Cloud Spanner host. 
- withHost(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the Cloud Spanner host. 
- withHost(String) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
- 
Define the host name of the Redis server. 
- withHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
- 
- withHost(String) - Method in class org.apache.beam.sdk.io.splunk.SplunkEvent.Builder
- 
Assigns host value to the event metadata. 
- withHostName(String) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Sets the host name to be used on the database. 
- withHostName(ValueProvider<String>) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Sets the host name to be used on the database. 
- withHosts(List<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the hosts of the Apache Cassandra instances. 
- withHosts(ValueProvider<List<String>>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the hosts of the Apache Cassandra instances. 
- withHosts(List<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Specify the Cassandra instance hosts where to write data. 
- withHosts(ValueProvider<List<String>>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Specify the hosts of the Apache Cassandra instances. 
- withHotKeyFanout(int) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
-  
- withHotKeyFanout(SerializableFunction<Row, Integer>) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
-  
- withHotKeyFanout(SerializableFunction<? super K, Integer>) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
- 
If a single key has disproportionately many values, it may become a bottleneck, especially in
 streaming mode. 
- withHotKeyFanout(int) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
- 
- withIdAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
- 
When reading from Cloud Pub/Sub where unique record identifiers are provided as Pub/Sub
 message attributes, specifies the name of the attribute containing the unique identifier. 
- withIdAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
- 
Writes to Pub/Sub, adding each record's unique identifier to the published messages in an
 attribute with the specified name. 
- withIdFn(ElasticsearchIO.Write.FieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.DocToBulk
- 
Provide a function to extract the id from the document. 
- withIdFn(ElasticsearchIO.Write.FieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withIdGenerator(IdGenerator) - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient
-  
- withIgnoreSSLCertificate(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
- 
Enable ignoreSSLCertificate for ssl for connection (allow for self signed certificates). 
- withIgnoreSSLCertificate(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
- 
Enable ignoreSSLCertificate for ssl for connection (allow for self signed certificates). 
- withIgnoreVersionConflicts(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BulkIO
- 
Whether or not to suppress version conflict errors in a Bulk API response. 
- withIgnoreVersionConflicts(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withIgnoreWindowing() - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- withInclusiveEndAt(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
Specifies the end time of the change stream. 
- withInclusiveStartAt(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
Specifies the time that the change stream should be read from. 
- withIndex(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- withIndex(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withIndex(String) - Method in class org.apache.beam.sdk.io.splunk.SplunkEvent.Builder
- 
Assigns index value to the event metadata. 
- withIndexes() - Method in class org.apache.beam.sdk.extensions.python.transforms.DataframeTransform
- 
Sets include_indexes option for DataframeTransform. 
- withIndexFn(ElasticsearchIO.Write.FieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.DocToBulk
- 
Provide a function to extract the target index from the document allowing for dynamic
 document routing. 
- withIndexFn(ElasticsearchIO.Write.FieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withInitialBackoff(Duration) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
- 
Set initial backoff duration. 
- withInitialBackoff(Duration) - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions.Builder
- 
Configure the initial backoff duration to be used before retrying a request for the first
 time. 
- withInitialPositionInStream(InitialPositionInStream) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specify reading from some initial position in stream. 
- withInitialSplitDuration(Duration) - Method in class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.ListHL7v2Messages
-  
- withInitialTimestampInStream(Instant) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specify reading beginning at given  Instant. 
- withInputDoc(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Document
- 
Sets the input document i.e. 
- withInputMetadata(Metadata) - Method in class org.apache.beam.sdk.io.tika.TikaIO.ParseFiles
- 
Sets the input metadata for Parser.parse(java.io.InputStream, org.xml.sax.ContentHandler, org.apache.tika.metadata.Metadata, org.apache.tika.parser.ParseContext).
 
- withInputTimestamp() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
- 
- withInputTimestamp() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
- 
The timestamp for each record being published is set to timestamp of the element in the
 pipeline. 
- withInsertDeduplicate(Boolean) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
- 
For INSERT queries in the replicated table, specifies that deduplication of inserting blocks
 should be performed. 
- withInsertDistributedSync(Boolean) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
- 
If setting is enabled, insert query into distributed waits until data will be sent to all
 nodes in cluster. 
- withInsertQuorum(Long) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
- 
For INSERT queries in the replicated table, wait writing for the specified number of replicas
 and linearize the addition of the data. 
- withInsertRetryPolicy(InsertRetryPolicy) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StreamingInserts
- 
Specify a retry policy for failed inserts. 
- withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableConfig
-  
- withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
- withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
- withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
- withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the Cloud Spanner instance ID. 
- withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the Cloud Spanner instance ID. 
- withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
- 
Specifies the Cloud Spanner instance. 
- withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
- 
Specifies the Cloud Spanner instance. 
- withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
- 
Specifies the Cloud Spanner instance. 
- withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
- 
Specifies the Cloud Spanner instance. 
- withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
- 
Specifies the Cloud Spanner instance. 
- withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
- 
Specifies the Cloud Spanner instance. 
- withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
Specifies the Cloud Spanner instance. 
- withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
Specifies the Cloud Spanner instance. 
- withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the Cloud Spanner instance. 
- withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the Cloud Spanner instance. 
- withInterceptors(List<ClientInterceptor>) - Method in class org.apache.beam.sdk.fn.channel.ManagedChannelFactory
- 
Returns a  ManagedChannelFactory like this one, but which will apply the provided  ClientInterceptors to any channel it creates. 
- withInterpolateFunction(SerializableFunction<FillGaps.InterpolateData<ValueT>, ValueT>) - Method in class org.apache.beam.sdk.extensions.timeseries.FillGaps
- 
This function can be used to modify elements before propagating to the next bucket. 
- withInterval(Duration) - Method in class org.apache.beam.sdk.transforms.PeriodicImpulse
-  
- withIsDeleteFn(ElasticsearchIO.Write.BooleanFieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.DocToBulk
- 
Provide a function to extract the target operation either upsert or delete from the document
 fields allowing dynamic bulk operation decision. 
- withIsDeleteFn(ElasticsearchIO.Write.BooleanFieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withIsLocalChannelProvider(ValueProvider<Boolean>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies whether a local channel provider should be used. 
- withIsReady(Supplier<Boolean>) - Method in class org.apache.beam.sdk.fn.test.TestStreams.Builder
- 
Returns a new  TestStreams.Builder like this one with the specified  CallStreamObserver.isReady() callback. 
- withIsUpsert(boolean) - Method in class org.apache.beam.sdk.io.mongodb.UpdateConfiguration
-  
- withJobService(BigQueryServices.JobService) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
-  
- withJsonClustering(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- withJsonSchema(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- withJsonSchema(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- withJsonTimePartitioning(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- withKeyClass(Class<K>) - Method in class org.apache.beam.sdk.io.cdap.CdapIO.Read
- 
Sets a key class. 
- withKeyClass(Class<K>) - Method in class org.apache.beam.sdk.io.cdap.CdapIO.Write
- 
Sets a key class. 
- withKeyDeserializer(Class<? extends Deserializer<K>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Sets a Kafka  Deserializer to interpret key bytes read from Kafka. 
- withKeyDeserializer(DeserializerProvider<K>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- withKeyDeserializer(Class<? extends Deserializer<K>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Sets a Kafka  Deserializer to interpret key bytes read from Kafka. 
- withKeyDeserializerAndCoder(Class<? extends Deserializer<K>>, Coder<K>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Sets a Kafka  Deserializer for interpreting key bytes read from Kafka along with a
  Coder for helping the Beam runner materialize key objects at runtime if necessary. 
- withKeyDeserializerAndCoder(Class<? extends Deserializer<K>>, Coder<K>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Sets a Kafka  Deserializer for interpreting key bytes read from Kafka along with a
  Coder for helping the Beam runner materialize key objects at runtime if necessary. 
- withKeyDeserializerProvider(DeserializerProvider<K>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
-  
- withKeyDeserializerProviderAndCoder(DeserializerProvider<K>, Coder<K>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- withKeyField(String) - Method in class org.apache.beam.sdk.schemas.transforms.CoGroup.Impl
-  
- withKeyField(String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
-  
- withKeyField(String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
- 
Set the name of the key field in the resulting schema. 
- withKeyPairAuth(String, PrivateKey) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets key pair authentication. 
- withKeyPairPathAuth(ValueProvider<String>, String, ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets key pair authentication. 
- withKeyPairPathAuth(ValueProvider<String>, String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets key pair authentication. 
- withKeyPairPathAuth(String, String, String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets key pair authentication. 
- withKeyPairPathAuth(String, String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets key pair authentication. 
- withKeyPairRawAuth(ValueProvider<String>, ValueProvider<String>, ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets key pair authentication. 
- withKeyPairRawAuth(ValueProvider<String>, ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets key pair authentication. 
- withKeyPairRawAuth(String, String, String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets key pair authentication. 
- withKeyPairRawAuth(String, String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets key pair authentication. 
- withKeyPattern(String) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
-  
- withKeyRange(ByteKeyRange) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
Returns a new  BigtableIO.Read that will read only rows in the specified range. 
- withKeyRange(ByteKeyRange) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
- 
Reads only rows in the specified range. 
- withKeyRange(byte[], byte[]) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
- 
Reads only rows in the specified range. 
- withKeyRanges(ValueProvider<List<ByteKeyRange>>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
Returns a new  BigtableIO.Read that will read only rows in the specified ranges. 
- withKeyRanges(List<ByteKeyRange>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
Returns a new  BigtableIO.Read that will read only rows in the specified ranges. 
- WithKeys<T> - Class in org.apache.beam.sdk.schemas.transforms
-  
- WithKeys<K,V> - Class in org.apache.beam.sdk.transforms
- 
WithKeys<K, V>takes aPCollection<V>, and either a constant key of typeKor a function fromVtoK, and returns aPCollection<KV<K, V>>, where
 each of the values in the inputPCollectionhas been paired with either the constant key
 or a key computed from the value.
 
- withKeySerializer(Class<? extends Serializer<K>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
- 
- withKeySerializer(Class<? extends Serializer<K>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
- 
Sets a  Serializer for serializing key (if any) to bytes. 
- withKeySet(KeySet) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- withKeySet(KeySet) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withKeyspace(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the Cassandra keyspace where to read data. 
- withKeyspace(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the Cassandra keyspace where to read data. 
- withKeyspace(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Specify the Cassandra keyspace where to write data. 
- withKeyspace(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Specify the Cassandra keyspace where to read data. 
- withKeystorePassword(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
If Elasticsearch uses SSL/TLS with mutual authentication (via shield), provide the password
 to open the client keystore. 
- withKeystorePath(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
If Elasticsearch uses SSL/TLS with mutual authentication (via shield), provide the keystore
 containing the client key. 
- withKeyTranslation(SimpleFunction<?, K>) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
- 
Transforms the keys read from the source using the given key translation function. 
- withKeyTranslation(SimpleFunction<?, K>, Coder<K>) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
- 
Transforms the keys read from the source using the given key translation function. 
- withKeyType(TypeDescriptor<K>) - Method in class org.apache.beam.sdk.transforms.WithKeys
- 
Return a  WithKeys that is like this one with the specified key type descriptor. 
- withKmsKey(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
For query sources, use this Cloud KMS key to encrypt any temporary tables created. 
- withKmsKey(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-  
- withKwarg(String, Object) - Method in class org.apache.beam.sdk.extensions.python.PythonExternalTransform
- 
Specifies a single keyword argument for the Python cross-language transform. 
- withKwarg(String, Object) - Method in class org.apache.beam.sdk.extensions.python.transforms.RunInference
- 
Sets keyword arguments for the model loader. 
- withKwargs(Map<String, Object>) - Method in class org.apache.beam.sdk.extensions.python.PythonExternalTransform
- 
Specifies keyword arguments for the Python cross-language transform. 
- withKwargs(Row) - Method in class org.apache.beam.sdk.extensions.python.PythonExternalTransform
- 
Specifies keyword arguments as a Row objects. 
- withLabel(String) - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
- 
- withLateFirings(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
-  
- withLateFirings(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
- 
Creates a new Triggerlike the this, except that it fires repeatedly whenever the
 givenTriggerfires after the watermark has passed the end of the window.
 
- withLimit(int) - Method in class org.apache.beam.sdk.io.mongodb.FindQuery
- 
Sets the limit of documents to find. 
- withLinkUrl(String) - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
- 
- withLiteralGqlQuery(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
- 
Returns a new  DatastoreV1.Read that reads the results of the specified GQL query. 
- withLiteralGqlQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
- 
- withLoadJobProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Set the project the BigQuery load job will be initiated from. 
- withLoadJobProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-  
- withLocalDc(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the local DC used for the load balancing. 
- withLocalDc(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the local DC used for the load balancing. 
- withLocalDc(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Specify the local DC used by the load balancing policy. 
- withLocalDc(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Specify the local DC used for the load balancing. 
- withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
- 
Returns a new  DatastoreV1.DeleteEntity that deletes entities from the Cloud Datastore Emulator
 running locally on the specified host port. 
- withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntityWithSummary
- 
- withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
- 
Returns a new  DatastoreV1.DeleteKey that deletes entities from the Cloud Datastore Emulator
 running locally on the specified host port. 
- withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKeyWithSummary
- 
- withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
- 
Returns a new  DatastoreV1.Read that reads from a Datastore Emulator running at the
 given localhost address. 
- withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
- 
Returns a new  DatastoreV1.Write that writes to the Cloud Datastore Emulator running locally on
 the specified host port. 
- withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.WriteWithSummary
- 
- withLocksDirPath(String) - Method in class org.apache.beam.sdk.io.cdap.CdapIO.Write
- 
Sets path to directory where locks will be stored. 
- withLogAppendTime() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
- withLogAppendTime() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Use the log append time as the output timestamp. 
- withLogAppendTime() - Static method in interface org.apache.beam.sdk.io.kafka.TimestampPolicyFactory
- 
A  TimestampPolicy that assigns Kafka's log append time (server side ingestion time) to
 each record. 
- withLoginCustomerId(Long) - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17.Read
- 
Creates and returns a new  GoogleAdsV17.Read transform with the specified login customer ID. 
- withLoginCustomerId(Long) - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17.ReadAll
- 
- withLoginTimeout(Integer) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets loginTimeout that will be used in SnowflakeBasicDataSource.setLoginTimeout(int).
 
- withLowerBound(PartitionColumnT) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadWithPartitions
-  
- withLowPriority() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withLowPriority() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-  
- withLowPriority() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-  
- withManualWatermarkEstimator() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
- withMapper(ObjectMapper) - Method in class org.apache.beam.sdk.extensions.jackson.AsJsons
- 
- withMapper(ObjectMapper) - Method in class org.apache.beam.sdk.extensions.jackson.ParseJsons
- 
- withMapperFactoryFn(SerializableFunction<Session, Mapper>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
A factory to create a specific  Mapper for a given Cassandra Session. 
- withMapperFactoryFn(SerializableFunction<Session, Mapper>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
-  
- withMasterAddresses(String) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Read
- 
Reads from the Kudu cluster on the specified master addresses. 
- withMasterAddresses(String) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Write
- 
Writes to the Kudu cluster on the specified master addresses. 
- withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Parse
- 
- withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ParseAll
- 
Deprecated. 
- withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Read
- 
- withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadAll
- 
Deprecated. 
- withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.Read
- 
- withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.TextIO.Read
- 
- withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
- 
Deprecated. 
- withMaxAttempts(int) - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions.Builder
- 
Configure the maximum number of times a request will be attempted for a complete successful
 result. 
- withMaxBatchBufferingDuration(long) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
- 
- withMaxBatchBufferingDuration(long) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
- 
Provide maximum buffering time to batch elements before committing SQL statement. 
- withMaxBatchBytesSize(int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
- 
Writes to Pub/Sub are limited by 10mb in general. 
- withMaxBatchSize(long) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BulkIO
- 
Provide a maximum size in number of documents for the batch see bulk API
 (https://www.elastic.co/guide/en/elasticsearch/reference/7.17/docs-bulk.html). 
- withMaxBatchSize(long) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withMaxBatchSize(int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
- 
Writes to Pub/Sub are batched to efficiently send data. 
- withMaxBatchSize(int) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Write
- 
Provide a maximum size in number of documents for the batch. 
- withMaxBatchSizeBytes(long) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BulkIO
- 
Provide a maximum size in bytes for the batch see bulk API
 (https://www.elastic.co/guide/en/elasticsearch/reference/7.17/docs-bulk.html). 
- withMaxBatchSizeBytes(long) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withMaxBufferElementCount(Integer) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
Returns a new  BigtableIO.Read that will break up read requests into smaller batches. 
- withMaxBufferingDuration(Duration) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BulkIO
- 
- withMaxBufferingDuration(Duration) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withMaxBufferingDuration(Duration) - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches
- 
Sets a time limit (in processing time) on how long an incomplete batch of elements is allowed
 to be buffered. 
- withMaxBytesPerBatch(long) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withMaxBytesPerPartition(long) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Control how much data will be assigned to a single BigQuery load job. 
- withMaxCapacityPerShard(Integer) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specifies the maximum number of messages per one shard. 
- withMaxCommitDelay(long) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
-  
- withMaxCommitDelay(Duration) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the max commit delay for high throughput writes. 
- withMaxCommitDelay(ValueProvider<Duration>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the max commit delay for high throughput writes. 
- withMaxCommitDelay(long) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies max commit delay for the Commit API call for throughput optimized writes. 
- withMaxConnectionIdleTime(int) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
- 
Sets the maximum idle time for a pooled connection. 
- withMaxConnectionIdleTime(int) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
- 
Sets the maximum idle time for a pooled connection. 
- withMaxConnections(Integer) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
- 
Sets the maximum total number of connections. 
- withMaxConnections(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
- 
- withMaxCumulativeBackoff(Duration) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
- 
Limits total time spent in backoff. 
- withMaxCumulativeBackoff(Duration) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the maximum cumulative backoff. 
- withMaxCumulativeBackoff(ValueProvider<Duration>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the maximum cumulative backoff. 
- withMaxCumulativeBackoff(Duration) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the maximum cumulative backoff time when retrying after DEADLINE_EXCEEDED errors. 
- withMaxElementsPerBatch(long) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withMaxFileSize(long) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Controls the maximum byte size per file to be loaded into BigQuery. 
- withMaxFilesPerBundle(int) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Control how many files will be written concurrently by a single worker when using BigQuery
 load jobs before spilling to a shuffle. 
- withMaxFilesPerPartition(int) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Controls how many files will be assigned to a single BigQuery load job. 
- withMaxGapFillBuckets(Long) - Method in class org.apache.beam.sdk.extensions.timeseries.FillGaps
-  
- withMaxInputSize(long) - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
- 
Returns an ApproximateQuantilesCombineFnthat's like this one except that it uses the
 specifiedmaxNumElementsvalue.
 
- withMaxInsertBlockSize(long) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
- 
The maximum block size for insertion, if we control the creation of blocks for insertion. 
- withMaxLen(long) - Method in class org.apache.beam.sdk.io.redis.RedisIO.WriteStreams
- 
When appending (XADD) to a stream, set a MAXLEN option. 
- withMaxNumberOfRecords(Integer) - Method in class org.apache.beam.io.debezium.DebeziumIO.Read
- 
Once the specified number of records has been reached, it will stop fetching them. 
- withMaxNumConnections(Integer) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Read
- 
Optional. 
- withMaxNumMutations(long) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the cell mutation limit (maximum number of mutated cells per batch). 
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.amqp.AmqpIO.Read
- 
Define the max number of records received by the  AmqpIO.Read. 
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specifies to read at most a given number of records. 
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Read
- 
Define the max number of records received by the  SqsIO.Read. 
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.BoundedReadFromUnboundedSource
- 
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
- 
Define the max number of records that the source will read. 
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Read
- 
Define the max number of records received by the  MqttIO.Read. 
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Read
- 
- withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.Read.Unbounded
- 
- withMaxNumRows(long) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the row mutation limit (maximum number of mutated rows per batch). 
- withMaxNumWritersPerBundle(int) - Method in class org.apache.beam.sdk.io.WriteFiles
- 
Set the maximum number of writers created in a bundle before spilling to shuffle. 
- withMaxOutstandingBytes(long) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
Returns a new  BigtableIO.Write with the max number of outstanding bytes allowed
 before enforcing flow control. 
- withMaxOutstandingElements(long) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
Returns a new  BigtableIO.Write with the max number of outstanding elements allowed
 before enforcing flow control. 
- withMaxParallelRequests(int) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BulkIO
- 
- withMaxParallelRequests(int) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withMaxParallelRequestsPerWindow(int) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BulkIO
- 
- withMaxParallelRequestsPerWindow(int) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.amqp.AmqpIO.Read
- 
Define the max read time (duration) while the  AmqpIO.Read will receive messages. 
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specifies to read records during maxReadTime.
 
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Read
- 
Define the max read time (duration) while the  SqsIO.Read will receive messages. 
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.BoundedReadFromUnboundedSource
- 
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.GenerateSequence
- 
Specifies to stop generating elements after the given time. 
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
- 
Define the max read time that the source will read. 
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Read
- 
Define the max read time (duration) while the  MqttIO.Read will receive messages. 
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Read
- 
Define the max read time (duration) while the  RabbitMqIO.Read will receive messages. 
- withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.Read.Unbounded
- 
- withMaxRetries(int) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
- 
Maximum number of retries per insert. 
- withMaxRetryJobs(int) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
If set, this will set the max number of retry of batch load jobs. 
- withMaxTimeToRun(Long) - Method in class org.apache.beam.io.debezium.DebeziumIO.Read
- 
Once the connector has run for the determined amount of time, it will stop. 
- withMemoryMB(int) - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
- 
Sets the size of the memory buffer in megabytes. 
- withMergeFunction(SerializableBiFunction<TimestampedValue<ValueT>, TimestampedValue<ValueT>, TimestampedValue<ValueT>>) - Method in class org.apache.beam.sdk.extensions.timeseries.FillGaps
- 
If there are multiple values in a single timeseries bucket, this function is used to specify
 what to propagate to the next bucket. 
- withMessageMapper(JmsIO.MessageMapper<T>) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
-  
- withMetadata(Map<String, Object>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Sink
- 
Specifies to put the given metadata into each generated file. 
- withMetadata(Map<String, Object>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
Writes to Avro file(s) with the specified metadata. 
- withMetadata(Map<String, Object>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- withMetadata() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
- 
Include metadata in result json documents. 
- withMetadata(Map<String, byte[]>) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
- withMetadata(String, byte[]) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
- withMetadata(String, String) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
- 
- withMetadataDatabase(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
Specifies the metadata database. 
- withMetadataInstance(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
Specifies the metadata database. 
- withMetadataTable(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
Specifies the metadata table name. 
- withMetadataTableAppProfileId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
- withMetadataTableInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
Returns a new  BigtableIO.ReadChangeStream that will use the Cloud Bigtable instance
 indicated by given parameter to manage the metadata of the stream. 
- withMetadataTableProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
Returns a new  BigtableIO.ReadChangeStream that will use the Cloud Bigtable project
 indicated by given parameter to manage the metadata of the stream. 
- withMetadataTableTableId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
- withMethod(BigQueryIO.TypedRead.Method) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- withMethod(BigQueryIO.Write.Method) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Choose the method used to write data to BigQuery. 
- withMethod(RedisIO.Write.Method) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
-  
- withMetric(String) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Read
- 
Sets the metric to use. 
- WithMetricsSupport - Class in org.apache.beam.runners.spark.metrics
- 
A MetricRegistrydecorator-like that supportsAggregatorMetricandSparkBeamMetricasGauges.
 
- WithMetricsSupport - Class in org.apache.beam.runners.spark.structuredstreaming.metrics
- 
A MetricRegistrydecorator-like that supportsBeamMetricSets asGauges.
 
- withMinBundleSize(long) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
- 
Sets the minimum bundle size. 
- withMinBundleSize(long) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
- 
Sets a parameter minBundleSizefor the minimum bundle size of the source.
 
- withMinNumberOfSplits(Integer) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
It's possible that system.size_estimates isn't populated or that the number of splits
 computed by Beam is still to low for Cassandra to handle it. 
- withMinNumberOfSplits(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
It's possible that system.size_estimates isn't populated or that the number of splits
 computed by Beam is still to low for Cassandra to handle it. 
- withMinRam(long) - Method in class org.apache.beam.sdk.transforms.resourcehints.ResourceHints
- 
Sets desired minimal available RAM size to have in transform's execution environment. 
- withMinRam(String) - Method in class org.apache.beam.sdk.transforms.resourcehints.ResourceHints
- 
Sets desired minimal available RAM size to have in transform's execution environment. 
- withMode(WindowingStrategy.AccumulationMode) - Method in class org.apache.beam.sdk.values.WindowingStrategy
- 
Returns a  WindowingStrategy identical to  this but with the accumulation mode
 set to  mode. 
- withMongoDbPipeline(List<BsonDocument>) - Method in class org.apache.beam.sdk.io.mongodb.AggregationQuery
-  
- withMonitoringConfiguration(Monitoring) - Method in class org.apache.beam.io.requestresponse.RequestResponseIO
-  
- withMonotonicallyIncreasingWatermarkEstimator() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
- withName(String) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
-  
- withName(String) - Method in class org.apache.beam.sdk.schemas.Schema.Field
- 
Returns a copy of the Field with the name set. 
- withNamedParameters(Map<String, ?>) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
-  
- withNameOnlyQuery() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.PartitionQuery.Builder
- 
Update produced queries to only retrieve their __name__thereby not retrieving any
 fields and reducing resource requirements.
 
- withNamespace(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
- 
- withNamespace(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
- 
- withNamespace(Class<?>) - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
- 
- withNaming(FileIO.Write.FileNaming) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
Specifies a custom strategy for generating filenames. 
- withNaming(SerializableFunction<DestinationT, FileIO.Write.FileNaming>) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- withNaming(Contextful<Contextful.Fn<DestinationT, FileIO.Write.FileNaming>>) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- withNestedField(int, FieldAccessDescriptor) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Return a descriptor that access the specified nested field. 
- withNestedField(String, FieldAccessDescriptor) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
Return a descriptor that access the specified nested field. 
- withNestedField(FieldAccessDescriptor.FieldDescriptor, FieldAccessDescriptor) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
-  
- withNestedFieldAs(String, String, FieldAccessDescriptor) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
- 
- withNoOutputTimestamp() - Method in interface org.apache.beam.sdk.state.Timer
- 
Asserts that there is no output timestamp. 
- withNoSpilling() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
- withNoSpilling() - Method in class org.apache.beam.sdk.io.csv.CsvIO.Write
- 
Whether to skip the spilling of data. 
- withNoSpilling() - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- withNoSpilling() - Method in class org.apache.beam.sdk.io.json.JsonIO.Write
- 
Whether to skip the spilling of data. 
- withNoSpilling() - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
- withNoSpilling() - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withNoSpilling() - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
- 
- withNoSpilling() - Method in class org.apache.beam.sdk.io.WriteFiles
- 
Returns a new  WriteFiles that writes all data without spilling, simplifying the
 pipeline. 
- withNullable(boolean) - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
-  
- withNullable(boolean) - Method in class org.apache.beam.sdk.schemas.Schema.Field
- 
Returns a copy of the Field with isNullable set. 
- withNullable(boolean) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
-  
- withNullBehavior(RowJson.RowJsonDeserializer.NullBehavior) - Method in class org.apache.beam.sdk.transforms.JsonToRow.JsonToRowWithErrFn
- 
Sets the behavior of the deserializer according to RowJson.RowJsonDeserializer.NullBehavior.
 
- withNumberOfClientsPerWorker(int) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Write
- 
The number of clients that each worker will create. 
- withNumberOfRecordsRead(long) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata.Builder
- 
Sets the number of records read in the partition change stream query before reading this
 record. 
- withNumBuckets(Integer) - Method in class org.apache.beam.sdk.transforms.Redistribute.RedistributeArbitrarily
-  
- withNumBuckets(Integer) - Method in class org.apache.beam.sdk.transforms.Reshuffle.ViaRandomKey
-  
- withNumFileShards(int) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Control how many file shards are written when using BigQuery load jobs. 
- withNumPartitions(int) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadWithPartitions
- 
The number of partitions. 
- withNumQuerySplits(int) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
- 
Returns a new  DatastoreV1.Read that reads by splitting the given  query into
  numQuerySplits. 
- withNumShards(int) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
Configures the number of output shards produced overall (when using unwindowed writes) or
 per-window (when using windowed writes). 
- withNumShards(int) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- withNumShards(Integer) - Method in class org.apache.beam.sdk.io.csv.CsvIO.Write
- 
Specifies to use a given fixed number of shards per window. 
- withNumShards(int) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
Specifies to use a given fixed number of shards per window. 
- withNumShards(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- withNumShards(Integer) - Method in class org.apache.beam.sdk.io.json.JsonIO.Write
- 
Specifies to use a given fixed number of shards per window. 
- withNumShards(int) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Write
- 
The number of workers used by the job to write to Solace. 
- withNumShards(int) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Configures the number of output shards produced overall (when using unwindowed writes) or
 per-window (when using windowed writes). 
- withNumShards(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
- withNumShards(int) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withNumShards(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withNumShards(int) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
- 
Writes to the provided number of shards. 
- withNumShards(int) - Method in class org.apache.beam.sdk.io.WriteFiles
- 
- withNumShards(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.WriteFiles
- 
- withNumSplits(int) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
- 
Sets the user defined number of splits. 
- withNumStorageWriteApiStreams(int) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Control how many parallel streams are used when using Storage API writes. 
- withOAuth(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets OAuth authentication. 
- withOAuth(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets OAuth authentication. 
- withOffset(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
- 
Partitions the timestamp space into half-open intervals of the form [N * size + offset, (N + 1)
 * size + offset), where 0 is the epoch. 
- withOffset(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
- 
Assigns timestamps into half-open intervals of the form [N * period + offset, N * period +
 offset + size). 
- withOffsetConsumerConfigOverrides(Map<String, Object>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Set additional configuration for the backend offset consumer. 
- withOffsetConsumerConfigOverrides(Map<String, Object>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Set additional configuration for the offset consumer. 
- withOffsetDeduplication(Boolean) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- withOnCompleted(Runnable) - Method in class org.apache.beam.sdk.fn.test.TestStreams.Builder
- 
Returns a new  TestStreams.Builder like this one with the specified  StreamObserver.onCompleted() callback. 
- withOnError(Runnable) - Method in class org.apache.beam.sdk.fn.test.TestStreams.Builder
- 
Returns a new  TestStreams.Builder like this one with the specified  StreamObserver.onError(java.lang.Throwable)
 callback. 
- withOnError(Consumer<Throwable>) - Method in class org.apache.beam.sdk.fn.test.TestStreams.Builder
- 
Returns a new  TestStreams.Builder like this one with the specified  StreamObserver.onError(java.lang.Throwable)
 consumer. 
- withOnNext(Consumer<T>) - Static method in class org.apache.beam.sdk.fn.test.TestStreams
- 
Creates a test  CallStreamObserverTestStreams.Builder that forwards  StreamObserver.onNext(V) calls to the supplied  Consumer. 
- withOnTimeBehavior(Window.OnTimeBehavior) - Method in class org.apache.beam.sdk.transforms.windowing.Window
- 
- withOnTimeBehavior(Window.OnTimeBehavior) - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- withOperationTimeout(Duration) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
- withOperationTimeout(Duration) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withOptionalParticipation() - Method in class org.apache.beam.sdk.schemas.transforms.CoGroup.By
- 
Means that this field will participate in a join even when not present, similar to SQL
 outer-join semantics. 
- withOptions(Schema.Options) - Method in class org.apache.beam.sdk.schemas.Schema.Field
- 
Returns a copy of the Field with the options set. 
- withOptions(Schema.Options.Builder) - Method in class org.apache.beam.sdk.schemas.Schema.Field
- 
Returns a copy of the Field with the options set. 
- withOptions(Schema.Options) - Method in class org.apache.beam.sdk.schemas.Schema
- 
Returns a copy of the Schema with the options set. 
- withOptions(Schema.Options.Builder) - Method in class org.apache.beam.sdk.schemas.Schema
- 
Returns a copy of the Schema with the options set. 
- withOrdered(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
- 
Enables ordered bulk insertion (default: true). 
- withOrdinality - Variable in class org.apache.beam.sdk.extensions.sql.zetasql.unnest.ZetaSqlUnnest
-  
- withoutDefaults() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
- 
Returns a  PTransform identical to this, but that does not attempt to provide a
 default value in the case of empty input. 
- withoutLimiter() - Static method in interface org.apache.beam.sdk.io.aws2.kinesis.RateLimitPolicyFactory
-  
- withoutMetadata() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Returns a  PTransform for PCollection of  KV, dropping Kafka metatdata. 
- withoutPartitioning() - Method in interface org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write.PartitionedWriterBuilder
- 
Writes to the sink without need to partition output into specified number of partitions. 
- withOutputCoder(Coder<?>) - Method in class org.apache.beam.sdk.extensions.python.PythonExternalTransform
- 
- withOutputCoder(Coder<OutputT>) - Method in class org.apache.beam.sdk.transforms.Watch.Growth
- 
Specifies a  Coder to use for the outputs. 
- withOutputCoders(Map<String, Coder<?>>) - Method in class org.apache.beam.sdk.extensions.python.PythonExternalTransform
- 
Specifies the keys and  Coders of the output  PCollections produced by this
 transform. 
- withOutputFilenames() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
Specify that output filenames are wanted. 
- withOutputFilenames() - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
Specify that output filenames are wanted. 
- withOutputKeyCoder(Coder<KeyT>) - Method in class org.apache.beam.sdk.transforms.Watch.Growth
- 
Specifies the coder for the output key. 
- withOutputParallelization(boolean) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
- 
Whether to reshuffle the resulting PCollection so results are distributed to all workers. 
- withOutputParallelization(boolean) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
- 
Whether to reshuffle the resulting PCollection so results are distributed to all workers. 
- withOutputParallelization(boolean) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
- 
Whether to reshuffle the resulting PCollection so results are distributed to all workers. 
- withOutputParallelization(boolean) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
- 
Whether to reshuffle the resulting PCollection so results are distributed to all workers. 
- withOutputParallelization(boolean) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadKeyPatterns
- 
Whether to reshuffle the resulting PCollection so results are distributed to all workers. 
- withOutputParallelization(Boolean) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.Read
- 
Whether to reshuffle the resulting PCollection so results are distributed to all workers. 
- withOutputs(List<TimestampedValue<OutputT>>) - Method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
- 
- withOutputSchema(Schema) - Method in class org.apache.beam.sdk.schemas.transforms.Select.Fields
- 
Rename all output fields to match the specified schema. 
- withOutputSchema(Schema) - Method in class org.apache.beam.sdk.schemas.transforms.Select.Flattened
- 
Rename all output fields to match the specified schema. 
- withOutputTags(TupleTag<OutputT>, TupleTagList) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
- 
- withOutputTimestamp(Instant) - Method in interface org.apache.beam.sdk.state.Timer
- 
Sets event time timer's output timestamp. 
- withoutRepeater() - Method in class org.apache.beam.io.requestresponse.RequestResponseIO
- 
- withoutResultFlattening() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
- 
- withoutResultFlattening() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- withoutSharding() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
Forces a single file as output and empty shard name template. 
- withoutSharding() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- withoutSharding() - Method in class org.apache.beam.sdk.io.csv.CsvIO.Write
- 
Forces a single file as output and empty shard name template. 
- withoutSharding() - Method in class org.apache.beam.sdk.io.json.JsonIO.Write
- 
Forces a single file as output and empty shard name template. 
- withoutSharding() - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Forces a single file as output and empty shard name template. 
- withoutSharding() - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withoutSharding() - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
- 
Forces a single file as output. 
- withoutStrictParsing() - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
- 
During parsing of the arguments, we will skip over improperly formatted and unknown
 arguments. 
- withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
- 
Disable validation that the table exists or the query succeeds prior to pipeline submission. 
- withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Disables BigQuery table validation. 
- withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
Disables validation that the table being read from exists. 
- withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
Disables validation that the table being read and the metadata table exists, and that the app
 profile used is single cluster and single row transaction enabled. 
- withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
Disables validation that the table being written to exists. 
- withoutValidation() - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
- 
Returns a transform for reading TFRecord files that has GCS path validation on pipeline
 creation disabled. 
- withOverloadRatio(double) - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions.Builder
- 
The target ratio between requests sent and successful requests. 
- withParallelism(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.splunk.SplunkIO.Write
- 
- withParallelism(Integer) - Method in class org.apache.beam.sdk.io.splunk.SplunkIO.Write
- 
Sets the number of parallel http client connections to the HEC. 
- withParameterSetter(JdbcIO.PreparedStatementSetter<ParameterT>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-  
- withParametersFunction(SerializableFunction<ParameterT, Map<String, Object>>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.ReadAll
-  
- withParametersFunction(SerializableFunction<ParameterT, Map<String, Object>>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- withParams(Map<String, Object>) - Method in class org.apache.beam.sdk.io.cdap.ConfigWrapper
- 
- withParent(Schema.TypeName) - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold.Context
-  
- withParseFn(SerializableFunction<GenericRecord, X>, Coder<X>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
- 
Reads  GenericRecord of unspecified schema and maps them to instances of a custom type
 using the given  parseFn and encoded using the given coder. 
- withParseFn(SerializableFunction<RowResult, T>) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Read
- 
Provides the function to parse a row from Kudu into the typed object. 
- withParser(MongoDbGridFSIO.Parser<X>) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-  
- withPartition(Map<String, String>) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
- 
Sets the partition details. 
- withPartitionCols(List<String>) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
- 
Set the names of the columns that are partitions. 
- withPartitionColumn(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadWithPartitions
- 
The name of a column of numeric type that will be used for partitioning. 
- withPartitionCreatedAt(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata.Builder
- 
Sets the time at which this partition was first detected and created in the metadata table. 
- withPartitionEndTimestamp(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata.Builder
- 
Sets the end time for the partition change stream query that originated this record. 
- withPartitioner(KinesisPartitioner<T>) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Write
- 
Specify how to partition records among all stream shards (required). 
- withPartitioning() - Method in interface org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write.PartitionedWriterBuilder
- 
Writes to the sink with partitioning by Task Id. 
- withPartitionOptions(PartitionOptions) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- withPartitionOptions(PartitionOptions) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
- 
Note that PartitionOptionsare currently ignored.
 
- withPartitionQueryTimeout(Duration) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the PartitionQuery timeout. 
- withPartitionQueryTimeout(ValueProvider<Duration>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the PartitionQuery timeout. 
- withPartitionReadTimeout(Duration) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the PartitionRead timeout. 
- withPartitionReadTimeout(ValueProvider<Duration>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the PartitionRead timeout. 
- withPartitionRunningAt(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata.Builder
- 
Sets the time at which the connector started processing this partition. 
- withPartitionScheduledAt(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata.Builder
- 
Sets the time at which this partition was scheduled to be queried. 
- withPartitionStartTimestamp(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata.Builder
- 
Sets the start time for the partition change stream query that originated this record. 
- withPartitionToken(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata.Builder
- 
Sets the partition token where this record originated from. 
- withPassword(String) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Sets the password to connect to your database. 
- withPassword(ValueProvider<String>) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Sets the password to connect to your database. 
- withPassword(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the password used for authentication. 
- withPassword(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the password used for authentication. 
- withPassword(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Specify the password used for authentication. 
- withPassword(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Specify the password used for authentication. 
- withPassword(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
If Elasticsearch authentication is enabled, provide the password. 
- withPassword(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
-  
- withPassword(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
-  
- withPassword(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
- 
Define the password to connect to the JMS broker (authenticated). 
- withPassword(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
- 
Define the password to connect to the JMS broker (authenticated). 
- withPassword(String) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
-  
- withPassword(String) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverConfiguration
-  
- withPassword(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverConfiguration
-  
- withPassword(String) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.DataSourceConfiguration
-  
- withPathPart(String) - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold.Context
-  
- withPathPrefix(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
If Elasticsearch is not running at the root path, e.g. 
- withPayload(String) - Method in class org.apache.beam.sdk.io.splunk.SplunkWriteError.Builder
- 
Assigns the payload to be used for reprocessing. 
- withPayloadFn(SerializableFunction<InputT, byte[]>) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Write
-  
- withPercision(Integer) - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.Globally
-  
- withPercision(Integer) - Method in class org.apache.beam.sdk.extensions.zetasketch.ApproximateCountDistinct.PerKey
-  
- withPlacementId(String) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIPredict
-  
- withPluginConfig(PluginConfig) - Method in class org.apache.beam.sdk.io.cdap.CdapIO.Read
- 
- withPluginConfig(PluginConfig) - Method in class org.apache.beam.sdk.io.cdap.CdapIO.Write
- 
- withPointInTimeSearch() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
- 
Configures the source to user Point In Time search iteration while reading data from
 Elasticsearch. 
- withPointInTimeSearchAndSortConfiguration(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
- 
Similar to  the default PIT search but setting a specific
 sorting configuration which Elasticsearch will use to sort for the results. 
- withPointInTimeSearchAndTimestampSortProperty(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
- 
Similar to  the default PIT search but setting an existing
 timestamp based property name which Elasticsearch will use to sort for the results. 
- withPollingInterval(Duration) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
- 
If specified, polling for new partitions will happen at this periodicity. 
- withPollInterval(Duration) - Method in class org.apache.beam.sdk.transforms.Watch.Growth
- 
- withPort(String) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Sets the port on which your database is listening. 
- withPort(ValueProvider<String>) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Sets the port on which your database is listening. 
- withPort(int) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the port number of the Apache Cassandra instances. 
- withPort(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the port number of the Apache Cassandra instances. 
- withPort(int) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Specify the Cassandra instance port number where to write data. 
- withPort(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Specify the port number of the Apache Cassandra instances. 
- withPort(int) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
- 
Define the port number of the Redis server. 
- withPort(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
- 
- withPortNumber(Integer) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets port number to use to connect to Snowflake. 
- withPositionalParameters(List<?>) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
-  
- withPrecision(int) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.ApproximateDistinctFn
- 
- withPrecision(int) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.GloballyDistinct
- 
Sets the precision p.
 
- withPrecision(int) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.PerKeyDistinct
- 
Sets the precision p.
 
- withPrecision(int) - Method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Init.Builder
- 
Explicitly set the precisionparameter used to compute HLL++ sketch.
 
- withPrecombining(boolean) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
- 
Enable precombining. 
- withPredicates(List<KuduPredicate>) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Read
- 
Filters the rows read from Kudu using the given predicates. 
- withPrefix(String) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
Specifies a common prefix to use for all generated filenames, if using the default file
 naming. 
- withPrefix(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- withPreparedStatementSetter(JdbcIO.PreparedStatementSetter<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
- 
See WriteVoid#withPreparedStatementSetter(PreparedStatementSetter).
 
- withPreparedStatementSetter(JdbcIO.PreparedStatementSetter<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
-  
- withPreparedStatementSetter(JdbcIO.PreparedStatementSetter<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteWithResults
-  
- withPrimaryKey(List<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-  
- withProcessingTime() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
- withProcessingTime() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Use the processing time as the output timestamp. 
- withProcessingTime() - Static method in interface org.apache.beam.sdk.io.kafka.TimestampPolicyFactory
- 
- withProcessingTime() - Method in class org.apache.beam.sdk.io.pulsar.PulsarIO.Read
-  
- withProcessingTimePolicy() - Static method in interface org.apache.beam.sdk.io.aws2.kinesis.WatermarkPolicyFactory
- 
Returns an ProcessingTimeWatermarkPolicy. 
- withProcessingTimeWatermarkPolicy() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specifies the WatermarkPolicyFactoryas ProcessingTimeWatermarkPolicyFactory.
 
- withProducerConfigUpdates(Map<String, Object>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
- 
Update configuration for the producer. 
- withProducerConfigUpdates(Map<String, Object>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
- 
Update configuration for the producer. 
- withProducerFactoryFn(SerializableFunction<Map<String, Object>, Producer<K, V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
- 
- withProducerFactoryFn(SerializableFunction<Map<String, Object>, Producer<K, V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
- 
Sets a custom function to create Kafka producer. 
- withProjectedColumns(List<String>) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Read
- 
Filters the columns read from the table to include only those specified. 
- withProjectId(String) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAICreateCatalogItem
-  
- withProjectId(String) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportCatalogItems
-  
- withProjectId(String) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIImportUserEvents
-  
- withProjectId(String) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIPredict
-  
- withProjectId(String) - Method in class org.apache.beam.sdk.extensions.ml.RecommendationAIWriteUserEvent
-  
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableConfig
-  
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
- 
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
- 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntityWithSummary
- 
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntityWithSummary
- 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
- 
Returns a new  DatastoreV1.DeleteKey that deletes entities from the Cloud Datastore for the
 specified project. 
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
- 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKeyWithSummary
- 
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKeyWithSummary
- 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
- 
Returns a new  DatastoreV1.Read that reads from the Cloud Datastore for the specified
 project. 
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
- 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
- 
Returns a new  DatastoreV1.Write that writes to the Cloud Datastore for the default database. 
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
- 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.WriteWithSummary
- 
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.WriteWithSummary
- 
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the Cloud Spanner project ID. 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the Cloud Spanner project ID. 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
- 
Specifies the Cloud Spanner project. 
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
- 
Specifies the Cloud Spanner project. 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
- 
Specifies the Cloud Spanner project. 
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
- 
Specifies the Cloud Spanner project. 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
- 
Specifies the Cloud Spanner project. 
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
- 
Specifies the Cloud Spanner project. 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
Specifies the Cloud Spanner project. 
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
Specifies the Cloud Spanner project. 
- withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the Cloud Spanner project. 
- withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the Cloud Spanner project. 
- withProjection(List<String>) - Method in class org.apache.beam.sdk.io.mongodb.FindQuery
- 
Sets the projection. 
- withProjection(Schema, Schema) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Read
- 
Enable the reading with projection. 
- withProjection(Schema, Schema) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.ReadFiles
-  
- withPropagateSuccessfulStorageApiWrites(boolean) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- withPropagateSuccessfulStorageApiWrites(Predicate<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- withProtocol(TProtocolFactory) - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.ReadFiles
- 
Specifies the TProtocolFactoryto be used to decode Thrift objects.
 
- withPublishRequestBuilder(SerializableFunction<T, PublishRequest.Builder>) - Method in class org.apache.beam.sdk.io.aws2.sns.SnsIO.Write
- 
Function to convert a message into a PublishRequest.Builder(mandatory).
 
- withPublishRequestFn(SerializableFunction<T, PublishRequest>) - Method in class org.apache.beam.sdk.io.aws2.sns.SnsIO.Write
- 
- withPublishTime() - Method in class org.apache.beam.sdk.io.pulsar.PulsarIO.Read
-  
- withPublishTimestampFunction(KafkaPublishTimestampFunction<KV<K, V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
- 
- withPublishTimestampFunction(KafkaPublishTimestampFunction<ProducerRecord<K, V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
- 
- withPubsubRootUrl(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
-  
- withPullFrequencySec(Long) - Method in class org.apache.beam.sdk.io.cdap.CdapIO.Read
- 
Delay in seconds between polling for new records updates. 
- withPullFrequencySec(Long) - Method in class org.apache.beam.sdk.io.sparkreceiver.SparkReceiverIO.Read
- 
Delay in seconds between polling for new records updates. 
- withPulsarClient(SerializableFunction<String, PulsarClient>) - Method in class org.apache.beam.sdk.io.pulsar.PulsarIO.Read
-  
- withQuery(String) - Method in class org.apache.beam.sdk.io.azure.cosmos.CosmosIO.Read
- 
Specify the query to read data. 
- withQuery(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the query to read data. 
- withQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the query to read data. 
- withQuery(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
- 
Provide a query used while reading from Elasticsearch. 
- withQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
- 
Provide a  ValueProvider that provides the query used while reading from
 Elasticsearch. 
- withQuery(Query) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
- 
- withQuery(Statement) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- withQuery(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- withQuery(Statement) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withQuery(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withQuery(String) - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17.Read
- 
- withQuery(String) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Read
- 
Sets the query to use. 
- withQuery(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-  
- withQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-  
- withQuery(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-  
- withQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-  
- withQuery(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
-  
- withQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
-  
- withQuery(String) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.Read
-  
- withQuery(String) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.ReadWithPartitions
-  
- withQuery(String) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Read
- 
Provide a query used while reading from Solr. 
- withQueryFn(SerializableFunction<MongoCollection<Document>, MongoCursor<Document>>) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
- 
Sets a queryFn. 
- withQueryLocation(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
BigQuery geographic location where the query  job will be
 executed. 
- withQueryName(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- withQueryName(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withQueryPlannerClass(Class<? extends QueryPlanner>) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
-  
- withQueryPriority(BigQueryIO.TypedRead.QueryPriority) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- withQueryStartedAt(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata.Builder
- 
Sets the time that the change stream query which produced this record started. 
- withQueryTempDataset(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- withQueryTempProjectAndDataset(String, String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- withQueryTransformation(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
A query to be executed in Snowflake. 
- withQueryTransformation(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
-  
- withQueue(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
- 
Specify the JMS queue destination name where to read messages from. 
- withQueue(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
- 
Specify the JMS queue destination name where to send messages to. 
- withQueue(String) - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Read
- 
If you want to directly consume messages from a specific queue, you just have to specify the
 queue name. 
- withQueue(String) - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Write
- 
Defines the queue where the messages will be sent. 
- withQueueDeclare(boolean) - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Read
- 
You can "force" the declaration of a queue on the RabbitMQ broker. 
- withQueueDeclare(boolean) - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Write
- 
If the queue is not declared by another application,  RabbitMqIO can declare the queue
 itself. 
- withQueueUrl(String) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Read
- 
Define the queueUrl used by the  SqsIO.Read to receive messages from SQS. 
- withQuotationMark(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
- 
Sets Snowflake-specific quotations around strings. 
- withQuotationMark(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
-  
- withQuotationMark(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
Sets Snowflake-specific quotations around strings. 
- withRampupThrottlingDisabled() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
- 
- withRampupThrottlingDisabled() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntityWithSummary
- 
- withRampupThrottlingDisabled() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
- 
- withRampupThrottlingDisabled() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKeyWithSummary
- 
- withRampupThrottlingDisabled() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
- 
- withRampupThrottlingDisabled() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.WriteWithSummary
- 
- withRandomAccess() - Method in class org.apache.beam.sdk.transforms.View.AsList
- 
Returns a PCollection view like this one, but whose resulting list will have RandomAccess
 (aka fast indexing). 
- withRandomAccess(boolean) - Method in class org.apache.beam.sdk.transforms.View.AsList
- 
Returns a PCollection view like this one, but whose resulting list will have RandomAccess
 (aka fast indexing) according to the input parameter. 
- withRate(long, Duration) - Method in class org.apache.beam.sdk.io.GenerateSequence
- 
Specifies to generate at most a given number of elements per a given period. 
- withRateLimitPolicy(GoogleAdsV17.RateLimitPolicyFactory) - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17.Read
- 
Creates and returns a new  GoogleAdsV17.Read transform with the specified rate limit policy
 factory. 
- withRateLimitPolicy(GoogleAdsV17.RateLimitPolicyFactory) - Method in class org.apache.beam.sdk.io.googleads.GoogleAdsV17.ReadAll
- 
Creates and returns a new  GoogleAdsV17.ReadAll transform with the specified rate limit policy
 factory. 
- withReadChangeStreamTimeout(Duration) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
- withReadCommitted() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Sets "isolation_level" to "read_committed" in Kafka consumer configuration. 
- withReadCommitted() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Sets "isolation_level" to "read_committed" in Kafka consumer configuration. 
- withReadOperation(ReadOperation) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withReadTime(Instant) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
- 
- withReadTimeout(Integer) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the Cassandra client read timeout in ms. 
- withReadTimeout(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the Cassandra client read timeout in ms. 
- withReadTimeout(Integer) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Cassandra client socket option to set the read timeout in ms. 
- withReadTimeout(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Cassandra client socket option to set the read timeout in ms. 
- withReadTransaction() - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.ReadAll
-  
- withReceiveTimeout(Duration) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
- 
If set, block for the Duration of timeout for each poll to new JMS record if the previous
 poll returns no new record. 
- withRecordAggregation(KinesisIO.RecordAggregation) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Write
- 
Enable record aggregation that is compatible with the KPL / KCL. 
- withRecordAggregation(Consumer<KinesisIO.RecordAggregation.Builder>) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Write
- 
Enable record aggregation that is compatible with the KPL / KCL. 
- withRecordAggregationDisabled() - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Write
- 
Disable KPL / KCL like record aggregation. 
- withRecordClass(Class<T>) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
- 
Sets a JAXB annotated class that can be populated using a record of the provided XML file. 
- withRecordClass(Class<T>) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
- 
- withRecordClass(Class<T>) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
- 
Writes objects of the given class mapped to XML elements using JAXB. 
- withRecordElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
- 
Sets name of the record element of the XML document. 
- withRecordElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
- 
- withRecordNumMetadata() - Method in class org.apache.beam.sdk.io.contextualtextio.ContextualTextIO.Read
- 
Allows the user to opt into getting recordNums associated with each record. 
- withRecordReadAt(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata.Builder
- 
Sets the time at which the record was fully read. 
- withRecordStreamEndedAt(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata.Builder
- 
Sets the time at which the record finished streaming. 
- withRecordStreamStartedAt(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata.Builder
- 
Sets the time at which the record started to be streamed. 
- withRecordTimestamp(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata.Builder
- 
Sets the timestamp of when this record occurred. 
- withRedistribute() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Sets redistribute transform that hints to the runner to try to redistribute the work evenly. 
- withRedistribute() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Enable Redistribute. 
- withRedistributeNumKeys(int) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- withRedistributeNumKeys(int) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
-  
- withRelativeError(double) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.GlobalSketch
- 
Sets the relative error epsilon.
 
- withRelativeError(double) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.PerKeySketch
- 
Sets the relative error epsilon.
 
- withReplicaInfo(SolrIO.ReplicaInfo) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Read
- 
Read from a specific Replica (partition). 
- withReportDiagnosticMetrics() - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions.Builder
- 
Whether additional diagnostic metrics should be reported for a Transform. 
- withRepresentativeCoder(Coder<IdT>) - Method in class org.apache.beam.sdk.transforms.Deduplicate.WithRepresentativeValues
- 
Return a  WithRepresentativeValuesPTransform that is like this one, but with
 the specified id type coder. 
- withRepresentativeType(TypeDescriptor<IdT>) - Method in class org.apache.beam.sdk.transforms.Deduplicate.WithRepresentativeValues
- 
Return a  WithRepresentativeValuesPTransform that is like this one, but with
 the specified id type descriptor. 
- withRepresentativeType(TypeDescriptor<IdT>) - Method in class org.apache.beam.sdk.transforms.Distinct.WithRepresentativeValues
- 
Return a  WithRepresentativeValuesPTransform that is like this one, but with
 the specified output type descriptor. 
- withRepresentativeValueFn(SerializableFunction<T, IdT>) - Static method in class org.apache.beam.sdk.transforms.Deduplicate
- 
Returns a deduplication transform that deduplicates values using the supplied representative
 value for up to 10 mins within the  processing time domain. 
- withRepresentativeValueFn(SerializableFunction<T, IdT>) - Static method in class org.apache.beam.sdk.transforms.Distinct
- 
Returns a Distinct<T, IdT>PTransform.
 
- withRequestRecordsLimit(int) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specifies the maximum number of records in GetRecordsResult returned by GetRecords call which
 is limited by 10K records. 
- withRequiresDeduping() - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
- 
If set, requires runner deduplication for the messages. 
- withResponseItemJson(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Document
- 
Sets the element from Elasticsearch Bulk API response "items" pertaining to this
 WriteSummary. 
- withResults() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
- 
- withResults() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
- 
- withResults() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
- 
- withResults() - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
- 
- withResumeDelay(Duration) - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
- 
- withRetained(boolean) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Write
- 
Whether or not the publish message should be retained by the messaging engine. 
- withRetentionPolicy(String) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Read
- 
Sets the retention policy to use. 
- withRetentionPolicy(String) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Write
- 
Sets the retention policy to use. 
- withRetryableCodes(ImmutableSet<StatusCode.Code>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the errors that will be retried by the client library for all operations. 
- withRetryConfiguration(ElasticsearchIO.RetryConfiguration) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BulkIO
- 
Provides configuration to retry a failed batch call to Elasticsearch. 
- withRetryConfiguration(ElasticsearchIO.RetryConfiguration) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withRetryConfiguration(JdbcIO.RetryConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
- 
See WriteVoid#withRetryConfiguration(RetryConfiguration).
 
- withRetryConfiguration(JdbcIO.RetryConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
- 
- withRetryConfiguration(JdbcIO.RetryConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteWithResults
- 
- withRetryConfiguration(RetryConfiguration) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
- 
Specify the JMS retry configuration. 
- withRetryConfiguration(SolrIO.RetryConfiguration) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Write
- 
Provides configuration to retry a failed batch call to Solr. 
- withRetryStrategy(JdbcIO.RetryStrategy) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
- 
See WriteVoid#withRetryStrategy(RetryStrategy).
 
- withRetryStrategy(JdbcIO.RetryStrategy) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
- 
- withRetryStrategy(JdbcIO.RetryStrategy) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteWithResults
- 
- withRingRanges(Set<RingRange>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-  
- withRingRanges(ValueProvider<Set<RingRange>>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
-  
- withRole(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets user's role to be used when running queries on Snowflake. 
- withRole(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets user's role to be used when running queries on Snowflake. 
- withRootCaCertificatePath(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.splunk.SplunkIO.Write
- 
Same as  Builder#withRootCaCertificatePath(ValueProvider) but without a  ValueProvider. 
- withRootCaCertificatePath(String) - Method in class org.apache.beam.sdk.io.splunk.SplunkIO.Write
- 
Method to set the root CA certificate. 
- withRootElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
- 
Sets name of the root element of the XML document. 
- withRootElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
- 
- withRootElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Sink
-  
- withRootElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
- 
Sets the enclosing root element for the generated XML files. 
- withRoutingFn(ElasticsearchIO.Write.FieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.DocToBulk
- 
Provide a function to extract the target routing from the document allowing for dynamic
 document routing. 
- withRoutingFn(ElasticsearchIO.Write.FieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withRowFilter(ValueProvider<RowFilter>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
Returns a new  BigtableIO.Read that will filter the rows read from Cloud Bigtable
 using the given row filter. 
- withRowFilter(RowFilter) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
Returns a new  BigtableIO.Read that will filter the rows read from Cloud Bigtable
 using the given row filter. 
- withRowGroupSize(int) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Sink
- 
Specify row-group size; if not set or zero, a default is used by the underlying writer. 
- withRowMapper(JdbcIO.RowMapper<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-  
- withRowMapper(JdbcIO.RowMapper<OutputT>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
-  
- withRowMapper(JdbcIO.RowMapper<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadWithPartitions
-  
- withRowMapper(JdbcIO.RowMapper<V>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteWithResults
-  
- withRowMapper(Neo4jIO.RowMapper<OutputT>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.ReadAll
-  
- withRowMapper(SingleStoreIO.RowMapper<T>) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.Read
-  
- withRowMapper(SingleStoreIO.RowMapper<T>) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.ReadWithPartitions
-  
- withRowMutationInformationFn(SerializableFunction<T, RowMutationInformation>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Allows upserting and deleting rows for tables with a primary key defined. 
- withRowOutput() - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadWithPartitions
- 
Data output type is  Row, and schema is auto-inferred from the database. 
- withRowRestriction(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- withRowRestriction(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
Read only rows which match the specified filter, which must be a SQL expression compatible
 with  Google standard
 SQL. 
- withRowSchema(Schema) - Method in class org.apache.beam.sdk.transforms.Create.Values
- 
Returns a  Create.Values PTransform like this one that uses the given  Schema
 to represent objects. 
- withRpcPriority(Options.RpcPriority) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the RPC priority. 
- withRpcPriority(ValueProvider<Options.RpcPriority>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
- 
Specifies the RPC priority. 
- withRpcPriority(Options.RpcPriority) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
Specifies the priority of the change stream queries. 
- withRunnerDeterminedSharding() - Method in class org.apache.beam.sdk.io.WriteFiles
- 
- withSamplePeriod(Duration) - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions.Builder
- 
Configure the length of time sampled request data will be retained. 
- withSamplePeriodBucketSize(Duration) - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions.Builder
- 
Configure the size of buckets within the specified  samplePeriod. 
- withScan(Scan) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
- 
Filters the rows read from HBase using the given* scan. 
- withScanRequestFn(SerializableFunction<Void, ScanRequest>) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
- 
Can't pass ScanRequest object directly from client since this object is not full
 serializable. 
- withScanResponseMapperFn(SerializableFunction<ScanResponse, T>) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
-  
- withSchema(Schema) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
Sets the output schema. 
- withSchema(Schema) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- withSchema(String) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
- 
Reads files containing records that conform to the given schema. 
- withSchema(Schema) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
- 
- withSchema(Class<X>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroSource
- 
Reads files containing records of the given class. 
- withSchema(Schema) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable.DocumentToRow
-  
- withSchema(TableSchema) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Uses the specified schema for rows to be written. 
- withSchema(ValueProvider<TableSchema>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- withSchema(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets schema to use when connecting to Snowflake. 
- withSchema(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- withSchema(Schema, TypeDescriptor<T>, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Method in class org.apache.beam.sdk.transforms.Create.TimestampedValues
- 
- withSchema(Schema, TypeDescriptor<T>, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Method in class org.apache.beam.sdk.transforms.Create.Values
- 
Returns a  Create.Values PTransform like this one that uses the given  Schema
 to represent objects. 
- withSchema(Schema, TypeDescriptor<T>, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Method in class org.apache.beam.sdk.transforms.Create.WindowedValues
- 
Returns a  Create.WindowedValues PTransform like this one that uses the given  Schema to represent objects. 
- withSchema(Schema) - Static method in class org.apache.beam.sdk.transforms.JsonToRow
-  
- withSchema(Schema) - Static method in class org.apache.beam.sdk.values.Row
- 
- withSchemaAndNullBehavior(Schema, RowJson.RowJsonDeserializer.NullBehavior) - Static method in class org.apache.beam.sdk.transforms.JsonToRow
-  
- withSchemaFromView(PCollectionView<Map<String, String>>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Allows the schemas for each table to be computed within the pipeline itself. 
- withSchemaReadySignal(PCollection<?>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies an optional input PCollection that can be used as the signal for Wait.OnSignalto indicate when the database schema is ready to be read.
 
- withSchemaUpdateOptions(Set<BigQueryIO.Write.SchemaUpdateOption>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Allows the schema of the destination table to be updated as a side effect of the write. 
- withScrollKeepalive(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
- 
Provide a scroll keepalive. 
- withSelectedFields(List<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
- withSelectedFields(ValueProvider<List<String>>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
- 
Read only the specified fields (columns) from a BigQuery table. 
- withSempClientFactory(SempClientFactory) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Read
- 
- withSerializer(SerializableFunction<T, byte[]>) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Write
- 
Specify how to serialize records to bytes on the stream (required). 
- withServerName(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets the name of the Snowflake server. 
- withServerName(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
-  
- withServerUri(String) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
- 
Set up the MQTT broker URI. 
- withSessionConfig(SessionConfig) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.ReadAll
-  
- withSessionConfig(SessionConfig) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- withSessionConfig(ValueProvider<SessionConfig>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- withSessionServiceFactory(SessionServiceFactory) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Read
- 
- withSessionServiceFactory(SessionServiceFactory) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Write
- 
Set the provider used to obtain the properties to initialize a new session in the broker. 
- withShard(int) - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
-  
- withShardedKey() - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches
- 
Outputs batched elements associated with sharded input keys. 
- withSharding(PTransform<PCollection<UserT>, PCollectionView<Integer>>) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
Specifies a  PTransform to use for computing the desired number of shards in each
 window. 
- withSharding(PTransform<PCollection<UserT>, PCollectionView<Integer>>) - Method in class org.apache.beam.sdk.io.WriteFiles
- 
- withShardingFunction(ShardingFunction<UserT, DestinationT>) - Method in class org.apache.beam.sdk.io.WriteFiles
- 
Returns a new  WriteFiles that will write to the current  FileBasedSink using the
 specified sharding function to assign shard for inputs. 
- withShardNameTemplate(String) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
- withShardNameTemplate(String) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- withShardNameTemplate(String) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
- withShardNameTemplate(String) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withShardNameTemplate(String) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
- 
Uses the given shard name template. 
- withShardsNumber(Integer) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
Number of shards that are created per window. 
- withShardTemplate(String) - Method in class org.apache.beam.sdk.io.csv.CsvIO.Write
- 
- withShardTemplate(String) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
- 
Sets the shard template. 
- withShardTemplate(String) - Method in class org.apache.beam.sdk.io.json.JsonIO.Write
- 
- withSideInput() - Method in class org.apache.beam.sdk.schemas.transforms.CoGroup.By
-  
- withSideInput(String, PCollectionView<?>) - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
- 
- withSideInput(String, PCollectionView<?>) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
- 
- withSideInputs(List<PCollectionView<?>>) - Method in class org.apache.beam.sdk.io.WriteFiles
-  
- withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.Combine.Globally
- 
- withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.Combine.Globally
- 
- withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
-  
- withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
-  
- withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
- 
- withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
- 
- withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
- 
- withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
-  
- withSideInputs(Map<String, PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
- 
- withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
- 
- withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
- 
- withSideInputs(Map<String, PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
- 
- withSingletonValues() - Method in class org.apache.beam.sdk.transforms.View.AsMap
- 
- withSize(int) - Static method in class org.apache.beam.sdk.extensions.avro.schemas.utils.AvroUtils.FixedBytesField
- 
- withSize(long) - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches
-  
- withSkew(Duration) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-  
- withSkipHeaderLines(int) - Method in class org.apache.beam.sdk.io.TextIO.Read
-  
- withSkipHeaderLines(int) - Method in class org.apache.beam.sdk.io.TextIO.ReadFiles
-  
- withSkipIfEmpty(boolean) - Method in class org.apache.beam.sdk.io.WriteFiles
- 
Set this sink to skip writing any files if the PCollection is empty. 
- withSkipIfEmpty() - Method in class org.apache.beam.sdk.io.WriteFiles
-  
- withSkipKeyClone(boolean) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
- 
Determines if key clone should be skipped or not (default is 'false'). 
- withSkipValueClone(boolean) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
- 
Determines if value clone should be skipped or not (default is 'false'). 
- withSnowflakeServices(SnowflakeServices) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
- withSnowPipe(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
Sets name of  SnowPipe
 which can be created in Snowflake dashboard or cli: 
- withSnowPipe(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
- withSocketTimeout(Integer) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
If set, overwrites the default max retry timeout (30000ms) in the Elastic  RestClient
 and the default socket timeout (30000ms) in the  RequestConfig of the Elastic  RestClient. 
- withSoftTimeout(Duration) - Static method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.restriction.RestrictionInterrupter
- 
Sets a soft timeout from now for processing new positions. 
- withSource(String) - Method in class org.apache.beam.sdk.io.splunk.SplunkEvent.Builder
- 
Assigns source value to the event metadata. 
- withSourceConnector(SourceConnector) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Sets the SourceConnectorto be used.
 
- withSourceConnector(ValueProvider<SourceConnector>) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
-  
- withSourceType(String) - Method in class org.apache.beam.sdk.io.splunk.SplunkEvent.Builder
- 
Assigns sourceType value to the event metadata. 
- withSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
- 
Specifies the Cloud Spanner configuration. 
- withSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
- 
Specifies the Cloud Spanner configuration. 
- withSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
- 
Specifies the Cloud Spanner configuration. 
- withSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
Specifies the Cloud Spanner configuration. 
- withSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
- 
Specifies the Cloud Spanner configuration. 
- withSparkReceiverBuilder(ReceiverBuilder<V, ? extends Receiver<V>>) - Method in class org.apache.beam.sdk.io.sparkreceiver.SparkReceiverIO.Read
- 
- withSparsePrecision(int) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.GloballyDistinct
- 
Sets the sparse representation's precision sp.
 
- withSparsePrecision(int) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.PerKeyDistinct
- 
Sets the sparse representation's precision sp.
 
- withSparseRepresentation(int) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.ApproximateDistinctFn
- 
- withSsl(SSLOptions) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Optionally, specify  SSLOptions configuration to utilize SSL. 
- withSsl(ValueProvider<SSLOptions>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Optionally, specify  SSLOptions configuration to utilize SSL. 
- withSsl(SSLOptions) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Optionally, specify  SSLOptions configuration to utilize SSL. 
- withSsl(ValueProvider<SSLOptions>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Optionally, specify  SSLOptions configuration to utilize SSL. 
- withSSL(ValueProvider<Boolean>) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
- 
Define if a SSL connection to Redis server should be used. 
- withSSLEnabled(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
- 
Enable ssl for connection. 
- withSSLEnabled(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
- 
Enable ssl for connection. 
- withSSLInvalidHostNameAllowed(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
- 
Enable invalidHostNameAllowed for ssl for connection. 
- withSSLInvalidHostNameAllowed(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
- 
Enable invalidHostNameAllowed for ssl for connection. 
- withStagingBucketName(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
- 
Name of the cloud bucket (GCS by now) to use as tmp location of CSVs during COPY statement. 
- withStagingBucketName(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
-  
- withStagingBucketName(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
Name of the cloud bucket (GCS by now) to use as tmp location of CSVs during COPY statement. 
- withStagingBucketName(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
-  
- withStartingDay(int, int, int) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-  
- withStartingMonth(int, int) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-  
- withStartingYear(int) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-  
- withStartKey(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
- 
Returns new  ByteKeyRange like this one, but with the specified start key. 
- withStartOffset(Long) - Method in class org.apache.beam.sdk.io.cdap.CdapIO.Read
- 
Inclusive start offset from which the reading should be started. 
- withStartOffset(Long) - Method in class org.apache.beam.sdk.io.sparkreceiver.SparkReceiverIO.Read
- 
Inclusive start offset from which the reading should be started. 
- withStartPollTimeoutSec(Long) - Method in class org.apache.beam.sdk.io.cdap.CdapIO.Read
- 
Delay in seconds before start polling. 
- withStartPollTimeoutSec(Long) - Method in class org.apache.beam.sdk.io.sparkreceiver.SparkReceiverIO.Read
- 
Waiting time after the Receiverstarts.
 
- withStartReadTime(Instant) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Use timestamp to set up start offset. 
- withStartTime(Instant) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
- withStartTimestamp(Long) - Method in class org.apache.beam.sdk.io.pulsar.PulsarIO.Read
-  
- withStatement(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
- 
- withStatement(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
-  
- withStatement(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
-  
- withStatement(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteWithResults
-  
- withStatement(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteWithResults
-  
- withStatementPreparator(JdbcIO.StatementPreparator) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
-  
- withStatementPreparator(JdbcIO.StatementPreparator) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
-  
- withStatementPreparator(SingleStoreIO.StatementPreparator) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.Read
-  
- withStatusCode(Integer) - Method in class org.apache.beam.sdk.io.splunk.SplunkWriteError.Builder
- 
Assigns a return status code to assist with debugging. 
- withStatusMessage(String) - Method in class org.apache.beam.sdk.io.splunk.SplunkWriteError.Builder
- 
Assigns a return status message to assist with debugging. 
- withStopReadTime(Instant) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Use timestamp to set up stop offset. 
- withStopTime(Instant) - Method in class org.apache.beam.sdk.extensions.timeseries.FillGaps
-  
- withStorageClient(BigQueryServices.StorageClient) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
-  
- withStorageIntegrationName(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
- 
Name of the Storage Integration in Snowflake to be used. 
- withStorageIntegrationName(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Read
-  
- withStorageIntegrationName(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
Name of the Storage Integration in Snowflake to be used. 
- withStorageIntegrationName(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
-  
- withStreamName(String) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specify reading from streamName. 
- withStreamName(String) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Write
- 
Kinesis stream name which will be used for writing (required). 
- withSubmissionMode(SolaceIO.SubmissionMode) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Write
- 
This setting controls the JCSMP property MESSAGE_CALLBACK_ON_REACTOR. 
- withSuccessfulInsertsPropagation(boolean) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- withSuffix(String) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
Configures the filename suffix for written files. 
- withSuffix(String) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- withSuffix(String) - Method in class org.apache.beam.sdk.io.csv.CsvIO.Write
- 
Configures the filename suffix for written files. 
- withSuffix(String) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
- 
Sets the suffix. 
- withSuffix(String) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
Specifies a common suffix to use for all generated filenames, if using the default file
 naming. 
- withSuffix(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- withSuffix(String) - Method in class org.apache.beam.sdk.io.json.JsonIO.Write
- 
Configures the filename suffix for written files. 
- withSuffix(String) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Configures the filename suffix for written files. 
- withSuffix(String) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withSuffix(String) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
- 
Writes to the file(s) with the given filename suffix. 
- withSyncInterval(int) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
Sets the approximate number of uncompressed bytes to write in each block for the AVRO
 container format. 
- withSyncInterval(int) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- withTable(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the Cassandra table where to read data. 
- withTable(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the Cassandra table where to read data. 
- withTable(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
-  
- withTable(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withTable(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
- 
Sets the table name to read from. 
- withTable(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
- 
Sets the table name to write to, the table should exist beforehand. 
- withTable(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadWithPartitions
- 
Name of the table in the external database. 
- withTable(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
- 
- withTable(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
-  
- withTable(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteWithResults
-  
- withTable(String) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Read
- 
Reads from the specified table. 
- withTable(String) - Method in class org.apache.beam.sdk.io.kudu.KuduIO.Write
- 
Writes to the specified table. 
- withTable(String) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.Read
-  
- withTable(String) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.ReadWithPartitions
-  
- withTable(String) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.Write
-  
- withTableDescription(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Specifies the table description. 
- withTableId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
- withTableId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
- 
- withTableId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.ReadChangeStream
- 
- withTableId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withTableId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withTableId(String) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
- 
Reads from the specified table. 
- withTableId(String) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
- 
Writes to the specified table. 
- withTableId(String) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.WriteRowMutations
- 
Writes to the specified table. 
- withTableProvider(TableProvider) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
- 
This method creates  BeamSqlEnv using empty
 Pipeline Options. 
- withTableProvider(String, TableProvider) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
-  
- withTableReference(TableReference) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
-  
- withTableSchema(TableSchema) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
- 
Set TableSchema. 
- withTableSchema(SnowflakeTableSchema) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
Table schema to be used during creating table. 
- withTempDirectory(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
Set the base directory used to generate temporary files. 
- withTempDirectory(ResourceId) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
Set the base directory used to generate temporary files. 
- withTempDirectory(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- withTempDirectory(ResourceId) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- withTempDirectory(ResourceId) - Method in class org.apache.beam.sdk.io.csv.CsvIO.Write
- 
Set the base directory used to generate temporary files. 
- withTempDirectory(String) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
Specifies a directory into which all temporary files will be placed. 
- withTempDirectory(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.FileIO.Write
- 
- withTempDirectory(ResourceId) - Method in class org.apache.beam.sdk.io.json.JsonIO.Write
- 
Set the base directory used to generate temporary files. 
- withTempDirectory(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Set the base directory used to generate temporary files. 
- withTempDirectory(ResourceId) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Set the base directory used to generate temporary files. 
- withTempDirectory(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withTempDirectory(ResourceId) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withTemplateCompatibility() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
- 
Use new template-compatible source implementation. 
- withTemplateCompatibility() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-  
- withTempLocation(String) - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
- 
Sets the path to a temporary location where the sorter writes intermediate files. 
- withTerminationCondition(Watch.Growth.TerminationCondition<HCatalogIO.Read, ?>) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
- 
If specified, the poll function will stop polling after the termination condition has been
 satisfied. 
- withTerminationPerInput(Watch.Growth.TerminationCondition<InputT, ?>) - Method in class org.apache.beam.sdk.transforms.Watch.Growth
- 
- withTestServices(BigQueryServices) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
-  
- withTestServices(BigQueryServices) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
-  
- withTestServices(BigQueryServices) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-  
- withThrottleDuration(Duration) - Method in class org.apache.beam.sdk.io.gcp.firestore.RpcQosOptions.Builder
- 
Configure the amount of time an attempt will be throttled if deemed necessary based on
 previous success rate. 
- withThrottlingReportTargetMs(int) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withThrottlingTargetMs(int) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withThrowWriteErrors(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BulkIO
- 
Whether to throw runtime exceptions when write (IO) errors occur. 
- withThrowWriteErrors(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withTikaConfigPath(String) - Method in class org.apache.beam.sdk.io.tika.TikaIO.ParseFiles
- 
- withTikaConfigPath(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.tika.TikaIO.ParseFiles
- 
Like with(tikaConfigPath).
 
- withTime(Long) - Method in class org.apache.beam.sdk.io.splunk.SplunkEvent.Builder
- 
Assigns time value to the event metadata. 
- withTimeDomain(TimeDomain) - Method in class org.apache.beam.sdk.transforms.Deduplicate.KeyedValues
- 
Returns a  KeyedValuesPTransform like this one but with the specified time
 domain. 
- withTimeDomain(TimeDomain) - Method in class org.apache.beam.sdk.transforms.Deduplicate.Values
- 
Returns a  ValuesPTransform like this one but with the specified time domain. 
- withTimeDomain(TimeDomain) - Method in class org.apache.beam.sdk.transforms.Deduplicate.WithRepresentativeValues
- 
Returns a  WithRepresentativeValuesPTransform like this one but with the
 specified time domain. 
- withTimeout(Duration) - Method in class org.apache.beam.io.requestresponse.RequestResponseIO
- 
- withTimeout(int) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
- 
Define the Redis connection timeout. 
- withTimeout(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
- 
- withTimeout(int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
-  
- withTimeout(int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadKeyPatterns
-  
- withTimeout(int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
-  
- withTimeout(int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.WriteStreams
- 
Set the connection timeout for the Redis server connection. 
- withTimePartitioning(TimePartitioning) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- withTimePartitioning(ValueProvider<TimePartitioning>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
- withTimestamp(Instant) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Document
- 
Sets the timestamp of the element in the PCollection, to be used in order to output
 WriteSummary to the same window from which the inputDoc originated. 
- withTimestamp(Instant) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StorageApiWritePayload
-  
- withTimestamp(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withTimestamp(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-  
- withTimestampAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
- 
When reading from Cloud Pub/Sub where record timestamps are provided as Pub/Sub message
 attributes, specifies the name of the attribute that contains the timestamp. 
- withTimestampAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
- 
Writes to Pub/Sub and adds each record's timestamp to the published messages in an attribute
 with the specified name. 
- withTimestampBound(TimestampBound) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
-  
- withTimestampBound(TimestampBound) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withTimestampBound(TimestampBound) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-  
- withTimestampCombiner(TimestampCombiner) - Method in class org.apache.beam.sdk.transforms.windowing.Window
- 
- withTimestampCombiner(TimestampCombiner) - Method in class org.apache.beam.sdk.values.WindowingStrategy
-  
- withTimestampFn(SerializableFunction<KinesisRecord, Instant>) - Method in class org.apache.beam.sdk.io.aws2.kinesis.WatermarkParameters
- 
Specify the SerializableFunctionto extract the event time from aKinesisRecord.
 
- withTimestampFn(SerializableFunction<Long, Instant>) - Method in class org.apache.beam.sdk.io.GenerateSequence
- 
Specifies the function to use to assign timestamps to the elements. 
- withTimestampFn(SerializableFunction<KV<K, V>, Instant>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
- withTimestampFn(SerializableFunction<KafkaRecord<K, V>, Instant>) - Static method in interface org.apache.beam.sdk.io.kafka.TimestampPolicyFactory
- 
- withTimestampFn(SerializableFunction<T, Instant>) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Read
- 
The timestamp function, used for estimating the watermark, mapping the record T to an  Instant
- withTimestampFn(SerializableFunction<V, Instant>) - Method in class org.apache.beam.sdk.io.sparkreceiver.SparkReceiverIO.Read
- 
A function to calculate timestamp for a record. 
- withTimestampFn2(SerializableFunction<KafkaRecord<K, V>, Instant>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
- withTimestampPolicyFactory(TimestampPolicyFactory<K, V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
- WithTimestamps<T> - Class in org.apache.beam.sdk.transforms
- 
- withTimeZone(DateTimeZone) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
-  
- withTimeZone(DateTimeZone) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
-  
- withTimeZone(DateTimeZone) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
-  
- withToDateTime(String) - Method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Read
- 
Read metric data till the toDateTime. 
- withTopic(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
-  
- withTopic(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
- 
Specify the JMS topic destination name where to receive messages from. 
- withTopic(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
- 
Specify the JMS topic destination name where to send messages to. 
- withTopic(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Sets the topic to read from. 
- withTopic(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
- 
- withTopic(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
- 
Sets the default Kafka topic to write to. 
- withTopic(String) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
- 
Set up the MQTT getTopic pattern. 
- withTopic(String) - Method in class org.apache.beam.sdk.io.pulsar.PulsarIO.Read
-  
- withTopic(String) - Method in class org.apache.beam.sdk.io.pulsar.PulsarIO.Write
-  
- withTopicArn(String) - Method in class org.apache.beam.sdk.io.aws2.sns.SnsIO.Write
- 
SNS topic ARN used for publishing to SNS. 
- withTopicFn(SerializableFunction<InputT, String>) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Write
-  
- withTopicNameMapper(SerializableFunction<EventT, String>) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
- 
Specify the JMS topic destination name where to send messages to dynamically. 
- withTopicPartitions(List<TopicPartition>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Sets a list of partitions to read from. 
- withTopicPattern(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Internally sets a Patternof topics to read from.
 
- withTopics(List<String>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Sets a list of topics to read from. 
- withTotalStreamTimeMillis(long) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.model.ChangeStreamRecordMetadata.Builder
- 
Sets the total streaming time (in millis) for this record. 
- withTraceSampleProbability(Double) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
- 
- withTransaction(PCollectionView<Transaction>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
-  
- withTransaction(PCollectionView<Transaction>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
-  
- withTransactionConfig(TransactionConfig) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.ReadAll
-  
- withTransactionConfig(TransactionConfig) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- withTransactionConfig(ValueProvider<TransactionConfig>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- withTrigger(Trigger) - Method in class org.apache.beam.sdk.values.WindowingStrategy
- 
Returns a  WindowingStrategy identical to  this but with the trigger set to
  wildcardTrigger. 
- withTriggeringFrequency(Duration) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Choose the frequency at which file writes are triggered. 
- withTriggeringFrequency(Duration) - Method in class org.apache.beam.sdk.io.iceberg.IcebergIO.WriteRows
- 
Sets the frequency at which data is written to files and a new Snapshotis produced.
 
- withTrustSelfSignedCerts(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
If Elasticsearch uses SSL/TLS then configure whether to trust self signed certs or not. 
- withType(Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Field
- 
- withType(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.transforms.Create.TimestampedValues
- 
Returns a  Create.TimestampedValues PTransform like this one that uses the given
  TypeDescriptor<T> to determine the  Coder to use to decode each of the objects
 into a value of type  T. 
- withType(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.transforms.Create.Values
- 
Returns a  Create.Values PTransform like this one that uses the given  TypeDescriptor<T> to determine the  Coder to use to decode each of the objects into a
 value of type  T. 
- withType(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.transforms.Create.WindowedValues
- 
Returns a  Create.WindowedValues PTransform like this one that uses the given  TypeDescriptor<T> to determine the  Coder to use to decode each of the objects into a
 value of type  T. 
- withTypeFn(ElasticsearchIO.Write.FieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.DocToBulk
- 
Provide a function to extract the target type from the document allowing for dynamic document
 routing. 
- withTypeFn(ElasticsearchIO.Write.FieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withTypeHint(Class<?>, Schema.FieldType) - Method in class org.apache.beam.sdk.extensions.python.PythonExternalTransform
- 
Specifies the field type of arguments. 
- withUnwindMapName(String) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- withUnwindMapName(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- withUpdateConfiguration(UpdateConfiguration) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
-  
- withUpdateFields(UpdateField...) - Method in class org.apache.beam.sdk.io.mongodb.UpdateConfiguration
- 
Sets the configurations for multiple updates. 
- withUpdateKey(String) - Method in class org.apache.beam.sdk.io.mongodb.UpdateConfiguration
-  
- withUpperBound(PartitionColumnT) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadWithPartitions
-  
- withUpsertScript(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.DocToBulk
- 
Whether to use scripted updates and what script to use. 
- withUpsertScript(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withUpToDateThreshold(Duration) - Method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Read
- 
Specifies how late records consumed by this source can be to still be considered on time. 
- withUri(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
-  
- withUri(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-  
- withUri(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
- 
Define the location of the MongoDB instances using an URI. 
- withUri(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
- 
Define the location of the MongoDB instances using an URI. 
- withUri(String) - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Read
-  
- withUri(String) - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Write
-  
- withUrl(String) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverConfiguration
-  
- withUrl(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverConfiguration
-  
- withUrl(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets URL of Snowflake server in following format:
 jdbc:snowflake:// .snowflakecomputing.com
- withUrls(List<String>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverConfiguration
-  
- withUrls(ValueProvider<List<String>>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverConfiguration
-  
- withUseCorrelationId(boolean) - Method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Read
- 
Toggles deduplication of messages based on the amqp correlation-id property on incoming
 messages. 
- withUsePartialUpdate(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.DocToBulk
- 
Provide an instruction to control whether partial updates or inserts (default) are issued to
 Elasticsearch. 
- withUsePartialUpdate(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withUserDataMapper(SingleStoreIO.UserDataMapper<T>) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.Write
-  
- withUserDataMapper(SnowflakeIO.UserDataMapper<T>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
User-defined function mapping user data into CSV lines. 
- withUsername(String) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Sets the username to connect to your database. 
- withUsername(ValueProvider<String>) - Method in class org.apache.beam.io.debezium.DebeziumIO.ConnectorConfiguration
- 
Sets the username to connect to your database. 
- withUsername(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the username for authentication. 
- withUsername(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
- 
Specify the username for authentication. 
- withUsername(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Specify the username used for authentication. 
- withUsername(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
- 
Specify the username for authentication. 
- withUsername(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
- 
If Elasticsearch authentication is enabled, provide the username. 
- withUsername(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
-  
- withUsername(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
-  
- withUsername(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
- 
Define the username to connect to the JMS broker (authenticated). 
- withUsername(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
- 
Define the username to connect to the JMS broker (authenticated). 
- withUsername(String) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
-  
- withUsername(String) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverConfiguration
-  
- withUsername(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.DriverConfiguration
-  
- withUsername(String) - Method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO.DataSourceConfiguration
-  
- withUsernamePasswordAuth(String, String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets username/password authentication. 
- withUsernamePasswordAuth(ValueProvider<String>, ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets username/password authentication. 
- withUsesReshuffle(boolean) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ParseFiles
- 
Specifies if a Reshuffle should run before file reads occur. 
- withUsesReshuffle(boolean) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.ReadFiles
- 
Specifies if a Reshuffle should run before file reads occur. 
- withUseStatefulBatches(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BulkIO
- 
Whether or not to use Stateful Processing to ensure bulk requests have the desired number of
 entities i.e. 
- withUseStatefulBatches(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
- 
- withValidate(boolean) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableConfig
-  
- withValidation() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
- 
Enable validation of the PubSub Read. 
- withValidation() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
- 
Enable validation of the PubSub Write. 
- withValidation() - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
- 
After creation we will validate that  PipelineOptions conforms to all the validation
 criteria from  <T>. 
- withValidation() - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory
- 
After creation we will validate that <T>conforms to all the validation criteria.
 
- withValidationEventHandler(ValidationEventHandler) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
- 
Sets the ValidationEventHandlerto use with JAXB.
 
- withValidationEventHandler(ValidationEventHandler) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
- 
- withValueClass(Class<V>) - Method in class org.apache.beam.sdk.io.cdap.CdapIO.Read
- 
Sets a value class. 
- withValueClass(Class<V>) - Method in class org.apache.beam.sdk.io.cdap.CdapIO.Write
- 
Sets a value class. 
- withValueDeserializer(Class<? extends Deserializer<V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Sets a Kafka  Deserializer to interpret value bytes read from Kafka. 
- withValueDeserializer(DeserializerProvider<V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- withValueDeserializer(Class<? extends Deserializer<V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Sets a Kafka  Deserializer to interpret value bytes read from Kafka. 
- withValueDeserializerAndCoder(Class<? extends Deserializer<V>>, Coder<V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
Sets a Kafka  Deserializer for interpreting value bytes read from Kafka along with a
  Coder for helping the Beam runner materialize value objects at runtime if necessary. 
- withValueDeserializerAndCoder(Class<? extends Deserializer<V>>, Coder<V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
Sets a Kafka  Deserializer for interpreting value bytes read from Kafka along with a
  Coder for helping the Beam runner materialize value objects at runtime if necessary. 
- withValueDeserializerProvider(DeserializerProvider<V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
-  
- withValueDeserializerProviderAndCoder(DeserializerProvider<V>, Coder<V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
-  
- withValueField(String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
-  
- withValueMapper(SerializableBiFunction<EventT, Session, Message>) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
- 
Map the EventTobject to aMessage.
 
- withValueSerializer(Class<? extends Serializer<V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
- 
- withValueSerializer(Class<? extends Serializer<V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
- 
- withValueTranslation(SimpleFunction<?, V>) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
- 
Transforms the values read from the source using the given value translation function. 
- withValueTranslation(SimpleFunction<?, V>, Coder<V>) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
- 
Transforms the values read from the source using the given value translation function. 
- withWallTimeWatermarkEstimator() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.ReadSourceDescriptors
- 
- withWarehouse(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets Snowflake Warehouse to use. 
- withWarehouse(String) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.DataSourceConfiguration
- 
Sets Snowflake Warehouse to use. 
- withWatermark(Instant) - Method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
- 
- withWatermarkFn(SerializableFunction<KV<K, V>, Instant>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
- withWatermarkFn2(SerializableFunction<KafkaRecord<K, V>, Instant>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
- 
- withWatermarkIdleDurationThreshold(Duration) - Method in class org.apache.beam.sdk.io.aws2.kinesis.WatermarkParameters
- 
Specify the watermark idle duration to consider before advancing the watermark. 
- withWatermarkIdleDurationThreshold(Duration) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Read
- 
Optional. 
- withWatermarkRefreshRate(Duration) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadChangeStream
-  
- withWindowCoder(Coder<? extends BoundedWindow>) - Method in class org.apache.beam.sdk.transforms.Create.WindowedValues
- 
Returns a  Create.WindowedValues PTransform like this one that uses the given  Coder<T> to decode each of the objects into a value of type  T. 
- withWindowedWrites() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.TypedWrite
- 
Preserves windowing of input elements and writes them to files based on the element's window. 
- withWindowedWrites() - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Write
- 
- withWindowedWrites() - Method in class org.apache.beam.sdk.io.csv.CsvIO.Write
- 
Preserves windowing of input elements and writes them to files based on the element's window. 
- withWindowedWrites() - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
- 
Specify that writes are windowed. 
- withWindowedWrites() - Method in class org.apache.beam.sdk.io.json.JsonIO.Write
- 
Preserves windowing of input elements and writes them to files based on the element's window. 
- withWindowedWrites() - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
Preserves windowing of input elements and writes them to files based on the element's window. 
- withWindowedWrites() - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
- withWindowedWrites() - Method in class org.apache.beam.sdk.io.WriteFiles
- 
Returns a new  WriteFiles that writes preserves windowing on it's input. 
- withWindowFn(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.values.WindowingStrategy
- 
Returns a  WindowingStrategy identical to  this but with the window function set
 to  wildcardWindowFn. 
- withWritableByteChannelFactory(FileBasedSink.WritableByteChannelFactory) - Method in class org.apache.beam.sdk.io.csv.CsvIO.Write
- 
Returns a transform for writing to text files like this one but that has the given  FileBasedSink.WritableByteChannelFactory to be used by the  FileBasedSink during
 output. 
- withWritableByteChannelFactory(FileBasedSink.WritableByteChannelFactory) - Method in class org.apache.beam.sdk.io.json.JsonIO.Write
- 
Returns a transform for writing to text files like this one but that has the given  FileBasedSink.WritableByteChannelFactory to be used by the  FileBasedSink during
 output. 
- withWritableByteChannelFactory(FileBasedSink.WritableByteChannelFactory) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
- 
- withWritableByteChannelFactory(FileBasedSink.WritableByteChannelFactory) - Method in class org.apache.beam.sdk.io.TextIO.Write
- 
See TypedWrite#withWritableByteChannelFactory(WritableByteChannelFactory).
 
- withWriteDisposition(BigQueryIO.Write.WriteDisposition) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Specifies what to do with existing data in the table, in case the table already exists. 
- withWriteDisposition(WriteDisposition) - Method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
- 
A disposition to be used during writing to table phase. 
- withWriteRequestMapperFn(SerializableFunction<T, KV<String, WriteRequest>>) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Write
-  
- withWriteResults() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
- 
- withWriteResults(JdbcIO.RowMapper<V>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
- 
- withWriterType(SolaceIO.WriterType) - Method in class org.apache.beam.sdk.io.solace.SolaceIO.Write
- 
Set the type of writer used by the connector. 
- withWriteTempDataset(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
- 
Temporary dataset. 
- withWriteTransaction() - Method in class org.apache.beam.sdk.io.neo4j.Neo4jIO.ReadAll
-  
- witValueField(String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
- 
Set the name of the value field in the resulting schema. 
- WordCount - Class in org.apache.beam.runners.spark.structuredstreaming.examples
- 
Duplicated from beam-examples-java to avoid dependency. 
- WordCount() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.examples.WordCount
-  
- WordCount.CountWords - Class in org.apache.beam.runners.spark.structuredstreaming.examples
- 
A PTransform that converts a PCollection containing lines of text into a PCollection of
 formatted word counts. 
- WordCount.FormatAsTextFn - Class in org.apache.beam.runners.spark.structuredstreaming.examples
- 
A SimpleFunction that converts a Word and Count into a printable string. 
- WordCount.WordCountOptions - Interface in org.apache.beam.runners.spark.structuredstreaming.examples
- 
- WorkerLogLevelOverrides() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
- 
Deprecated. 
- workerStatus(StreamObserver<BeamFnApi.WorkerStatusRequest>) - Method in class org.apache.beam.runners.fnexecution.status.BeamWorkerStatusGrpcService
-  
- wrap(Throwable) - Method in class org.apache.beam.sdk.testing.PAssert.PAssertionSite
-  
- wrap(String) - Method in class org.apache.beam.sdk.testing.PAssert.PAssertionSite
-  
- wrapDescriptorProto(DescriptorProtos.DescriptorProto) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProto
-  
- WrappedList(List<Object>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedList
-  
- WrappedMap(Map<Object, Object>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedMap
-  
- WrappedRow(Row) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedRow
-  
- WrappedSupervisor - Class in org.apache.beam.sdk.io.sparkreceiver
- 
Wrapper class for ReceiverSupervisorthat doesn't use Spark Environment.
 
- WrappedSupervisor(Receiver<?>, SparkConf, SerializableFunction<Object[], Void>) - Constructor for class org.apache.beam.sdk.io.sparkreceiver.WrappedSupervisor
-  
- wrapping(StreamObserver<V>) - Static method in class org.apache.beam.sdk.fn.stream.SynchronizedStreamObserver
- 
Create a new  SynchronizedStreamObserver which will delegate all calls to the underlying
  StreamObserver, synchronizing access to that observer. 
- wrapProcessContext(DoFn<InputT, ?>.ProcessContext) - Static method in class org.apache.beam.sdk.transforms.Contextful.Fn.Context
- 
- wrapSegment(String) - Static method in class org.apache.beam.sdk.metrics.Lineage
- 
Wrap segment to valid segment name. 
- WritableCoder<T extends org.apache.hadoop.io.Writable> - Class in org.apache.beam.sdk.io.hadoop
- 
A  WritableCoder is a  Coder for a Java class that implements  Writable. 
- WritableCoder(Class<T>) - Constructor for class org.apache.beam.sdk.io.hadoop.WritableCoder
-  
- WritableCoder.WritableCoderProviderRegistrar - Class in org.apache.beam.sdk.io.hadoop
- 
- WritableCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.hadoop.WritableCoder.WritableCoderProviderRegistrar
-  
- write(DataOutputView) - Method in class org.apache.beam.runners.flink.adapter.FlinkKey
-  
- write(Kryo, Output) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.SideInputValues.BaseSideInputValues
-  
- write(ElementT) - Method in class org.apache.beam.sdk.extensions.avro.io.AvroIO.Sink
-  
- write(Class<T>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
Writes a  PCollection to an Avro file (or multiple Avro files matching a sharding
 pattern). 
- write(ByteBuffer) - Method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingSeekableByteChannel
-  
- write(ByteBuffer) - Method in class org.apache.beam.sdk.extensions.gcp.util.channels.CountingWritableByteChannel
-  
- write(int) - Method in class org.apache.beam.sdk.fn.stream.DataStreams.ElementDelimitedOutputStream
-  
- write(byte[], int, int) - Method in class org.apache.beam.sdk.fn.stream.DataStreams.ElementDelimitedOutputStream
-  
- write() - Static method in class org.apache.beam.sdk.io.amqp.AmqpIO
-  
- Write() - Constructor for class org.apache.beam.sdk.io.amqp.AmqpIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO
-  
- Write() - Constructor for class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO
- 
- Write() - Constructor for class org.apache.beam.sdk.io.aws2.kinesis.KinesisIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.aws2.sns.SnsIO
-  
- Write() - Constructor for class org.apache.beam.sdk.io.aws2.sns.SnsIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO
- 
- Write() - Constructor for class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Write
- 
Deprecated. 
- write() - Static method in class org.apache.beam.sdk.io.cassandra.CassandraIO
- 
- Write() - Constructor for class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.cdap.CdapIO
-  
- Write() - Constructor for class org.apache.beam.sdk.io.cdap.CdapIO.Write
-  
- write(String, String) - Static method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO
-  
- Write() - Constructor for class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
-  
- write(String, CSVFormat) - Static method in class org.apache.beam.sdk.io.csv.CsvIO
- 
- Write() - Constructor for class org.apache.beam.sdk.io.csv.CsvIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO
-  
- Write() - Constructor for class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
-  
- write(OutputT) - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
- 
Called for each value in the bundle. 
- write(ElementT) - Method in interface org.apache.beam.sdk.io.FileIO.Sink
- 
Appends a single element to the file. 
- write() - Static method in class org.apache.beam.sdk.io.FileIO
- 
- Write() - Constructor for class org.apache.beam.sdk.io.FileIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
- 
- Write() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO
- 
- Write() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
-  
- write() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1
- 
- write() - Method in class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1
- 
The class returned by this method provides the ability to create  PTransforms
 for write operations available in the Firestore V1 API provided by  FirestoreStub. 
- Write() - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.FhirIO.Write
-  
- Write() - Constructor for class org.apache.beam.sdk.io.gcp.healthcare.HL7v2IO.Write
-  
- Write() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
-  
- write(PublisherOptions) - Static method in class org.apache.beam.sdk.io.gcp.pubsublite.PubsubLiteIO
- 
Write messages to Pub/Sub Lite. 
- write(Object, Encoder) - Method in class org.apache.beam.sdk.io.gcp.spanner.changestreams.encoder.TimestampEncoding
- 
Serializes a Timestampreceived as datum to the output encoder out.
 
- write() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO
- 
- Write() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
- 
Creates an Write.Builderfor creation of Write Transformation.
 
- write() - Static method in class org.apache.beam.sdk.io.hbase.HBaseIO
- 
- write() - Static method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO
- 
Write data to Hive. 
- Write() - Constructor for class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.influxdb.InfluxDbIO
-  
- Write() - Constructor for class org.apache.beam.sdk.io.influxdb.InfluxDbIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
- 
Write data to a JDBC datasource. 
- write() - Static method in class org.apache.beam.sdk.io.jms.JmsIO
-  
- Write() - Constructor for class org.apache.beam.sdk.io.jms.JmsIO.Write
-  
- write(String) - Static method in class org.apache.beam.sdk.io.json.JsonIO
- 
Instantiates a  JsonIO.Write for writing user types in  JSONFormat format. 
- Write() - Constructor for class org.apache.beam.sdk.io.json.JsonIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.kafka.KafkaIO
- 
- Write() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.kudu.KuduIO
-  
- Write() - Constructor for class org.apache.beam.sdk.io.kudu.KuduIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO
- 
Write data to GridFS. 
- write(MongoDbGridFSIO.WriteFn<T>) - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO
-  
- Write() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
-  
- write(T, OutputStream) - Method in interface org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.WriteFn
- 
Output the object to the given OutputStream. 
- write() - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbIO
- 
Write data to MongoDB. 
- Write() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.mqtt.MqttIO
-  
- Write() - Constructor for class org.apache.beam.sdk.io.mqtt.MqttIO.Write
-  
- write(GenericRecord) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Sink
-  
- write() - Static method in class org.apache.beam.sdk.io.pulsar.PulsarIO
- 
Write to Apache Pulsar. 
- Write() - Constructor for class org.apache.beam.sdk.io.pulsar.PulsarIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO
-  
- Write() - Constructor for class org.apache.beam.sdk.io.rabbitmq.RabbitMqIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.redis.RedisIO
- 
Write data to a Redis server. 
- Write() - Constructor for class org.apache.beam.sdk.io.redis.RedisIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO
- 
Write data to a SingleStoreDB datasource. 
- Write() - Constructor for class org.apache.beam.sdk.io.singlestore.SingleStoreIO.Write
-  
- write(SnowflakeBatchServiceConfig) - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeBatchServiceImpl
- 
Writing data to Snowflake in batch mode. 
- write(SnowflakeBatchServiceConfig) - Method in interface org.apache.beam.sdk.io.snowflake.services.SnowflakeServices.BatchService
-  
- write(SnowflakeStreamingServiceConfig) - Method in interface org.apache.beam.sdk.io.snowflake.services.SnowflakeServices.StreamingService
-  
- write(SnowflakeStreamingServiceConfig) - Method in class org.apache.beam.sdk.io.snowflake.services.SnowflakeStreamingServiceImpl
- 
Writing data to Snowflake in streaming mode. 
- write() - Static method in class org.apache.beam.sdk.io.snowflake.SnowflakeIO
- 
Write data to Snowflake via COPY statement. 
- Write() - Constructor for class org.apache.beam.sdk.io.snowflake.SnowflakeIO.Write
-  
- write(SerializableFunction<T, Solace.Record>) - Static method in class org.apache.beam.sdk.io.solace.SolaceIO
- 
Create a  SolaceIO.Write transform, to write to Solace with a custom type. 
- write() - Static method in class org.apache.beam.sdk.io.solace.SolaceIO
- 
Create a  SolaceIO.Write transform, to write to Solace using  Solace.Record objects. 
- Write() - Constructor for class org.apache.beam.sdk.io.solace.SolaceIO.Write
-  
- write() - Static method in class org.apache.beam.sdk.io.solr.SolrIO
-  
- Write() - Constructor for class org.apache.beam.sdk.io.solr.SolrIO.Write
-  
- write(String, String) - Static method in class org.apache.beam.sdk.io.splunk.SplunkIO
- 
Write to Splunk's Http Event Collector (HEC). 
- write(ValueProvider<String>, ValueProvider<String>) - Static method in class org.apache.beam.sdk.io.splunk.SplunkIO
- 
- Write() - Constructor for class org.apache.beam.sdk.io.splunk.SplunkIO.Write
-  
- write(String) - Method in class org.apache.beam.sdk.io.TextIO.Sink
-  
- write() - Static method in class org.apache.beam.sdk.io.TextIO
- 
A  PTransform that writes a  PCollection to a text file (or multiple text files
 matching a sharding pattern), with each element of the input collection encoded into its own
 line. 
- write(byte[]) - Method in class org.apache.beam.sdk.io.TFRecordIO.Sink
-  
- write() - Static method in class org.apache.beam.sdk.io.TFRecordIO
- 
A  PTransform that writes a  PCollection to TFRecord file (or multiple TFRecord
 files matching a sharding pattern), with each element of the input collection encoded into its
 own record. 
- Write() - Constructor for class org.apache.beam.sdk.io.TFRecordIO.Write
-  
- write(T) - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.Sink
-  
- write(T) - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.ThriftWriter
-  
- write(T) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Sink
-  
- write() - Static method in class org.apache.beam.sdk.io.xml.XmlIO
- 
- Write() - Constructor for class org.apache.beam.sdk.io.xml.XmlIO.Write
-  
- write(String) - Static method in class org.apache.beam.sdk.managed.Managed
- 
- write(T) - Method in interface org.apache.beam.sdk.state.ValueState
- 
Set the value. 
- WRITE_TRANSFORMS - Static variable in class org.apache.beam.sdk.managed.Managed
-  
- WRITE_URN - Static variable in class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar
-  
- WRITE_URN - Static variable in class org.apache.beam.sdk.io.gcp.pubsublite.internal.ExternalTransformRegistrarImpl
-  
- WRITE_URN - Static variable in class org.apache.beam.sdk.io.snowflake.crosslanguage.SnowflakeTransformRegistrar
-  
- writeArtifacts(RunnerApi.Pipeline, String) - Method in class org.apache.beam.runners.jobsubmission.PortablePipelineJarCreator
- 
Stages all dependencies in pipeline into the jar file at outputStream, returning a new pipeline
 that references these artifacts as classpath artifacts. 
- writeAvros(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that writes binary encoded Avro messages of a given type to a
 Google Cloud Pub/Sub stream. 
- writeAvros(Class<T>, SerializableFunction<ValueInSingleWindow<T>, Map<String, String>>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that writes binary encoded Avro messages of a given type to a
 Google Cloud Pub/Sub stream. 
- writeBatches() - Static method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO
-  
- WriteBatches() - Constructor for class org.apache.beam.sdk.io.aws2.sqs.SqsIO.WriteBatches
-  
- WriteBuilder() - Constructor for class org.apache.beam.sdk.io.aws2.kinesis.KinesisTransformRegistrar.WriteBuilder
-  
- WriteBuilder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite.WriteBuilder
-  
- WriteBuilder - Class in org.apache.beam.sdk.io.snowflake.crosslanguage
-  
- WriteBuilder() - Constructor for class org.apache.beam.sdk.io.snowflake.crosslanguage.WriteBuilder
-  
- WriteBuilder.Configuration - Class in org.apache.beam.sdk.io.snowflake.crosslanguage
- 
Parameters class to expose the transform to an external SDK. 
- writeCallMetric(TableReference) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
-  
- writeCompressed(WritableByteChannel) - Method in enum org.apache.beam.sdk.io.Compression
-  
- writeCustomType() - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- writeCustomType(Class<OutputT>) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
A  PTransform that writes a  PCollection to an avro file (or multiple avro files
 matching a sharding pattern), with each element of the input collection encoded into its own
 record of type OutputT. 
- writeCustomType() - Static method in class org.apache.beam.sdk.io.TextIO
- 
A  PTransform that writes a  PCollection to a text file (or multiple text files
 matching a sharding pattern), with each element of the input collection encoded into its own
 line. 
- writeCustomTypeToGenericRecords() - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
- writeDefaultJobName(JarOutputStream, String) - Static method in class org.apache.beam.runners.jobsubmission.PortablePipelineJarUtils
-  
- writeDetectNewPartitionMissingPartitions(HashMap<Range.ByteStringRange, Instant>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
- 
Write to metadata table serialized missing partitions and how long they have been missing. 
- writeDetectNewPartitionVersion() - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
- 
Set the version number for DetectNewPartition. 
- WriteDisposition - Enum in org.apache.beam.sdk.io.snowflake.enums
- 
Enum containing all supported dispositions during writing to table phase. 
- writeDynamic() - Static method in class org.apache.beam.sdk.io.FileIO
- 
Writes elements to files using a  FileIO.Sink and grouping the elements using "dynamic
 destinations". 
- WriteErrorMetrics(String) - Constructor for class org.apache.beam.sdk.transforms.errorhandling.ErrorHandler.PTransformErrorHandler.WriteErrorMetrics
-  
- writeExternal(ObjectOutput) - Method in class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
-  
- WriteFailure(Write, WriteResult, Status) - Constructor for class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.WriteFailure
-  
- WriteFiles<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io
- 
- WriteFiles() - Constructor for class org.apache.beam.sdk.io.WriteFiles
-  
- WriteFilesResult<DestinationT> - Class in org.apache.beam.sdk.io
- 
- writeFooter() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
- 
Writes footer at the end of output files. 
- writeGenericRecords(Schema) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
Writes Avro records of the specified schema. 
- writeGenericRecords(String) - Static method in class org.apache.beam.sdk.extensions.avro.io.AvroIO
- 
Writes Avro records of the specified schema. 
- writeGenericRecords() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
- 
- WriteGrouped(SpannerIO.Write) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.WriteGrouped
-  
- writeHeader() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
- 
Writes header at the beginning of output files. 
- WriteJmsResult<EventT> - Class in org.apache.beam.sdk.io.jms
- 
- WriteJmsResult(Pipeline, TupleTag<EventT>, PCollection<EventT>) - Constructor for class org.apache.beam.sdk.io.jms.WriteJmsResult
-  
- writeMessages() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that writes to a Google Cloud Pub/Sub stream. 
- writeMessagesDynamic() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Enables dynamic destination topics. 
- writeMetrics(MetricQueryResults) - Method in interface org.apache.beam.sdk.metrics.MetricsSink
-  
- writeNewPartition(NewPartition) - Method in class org.apache.beam.sdk.io.gcp.bigtable.changestreams.dao.MetadataTableDao
- 
After a split or merge from a close stream, write the new partition's information to the
 metadata table. 
- WriteOperation(FileBasedSink<?, DestinationT, OutputT>) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
- 
Constructs a WriteOperation using the default strategy for generating a temporary directory
 from the base output filename. 
- WriteOperation(FileBasedSink<?, DestinationT, OutputT>, ResourceId) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
- 
Create a new WriteOperation. 
- writeProtos(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
- 
- writeProtos(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that writes binary encoded protobuf messages of a given type to a
 Google Cloud Pub/Sub stream. 
- writeProtos(Class<T>, SerializableFunction<ValueInSingleWindow<T>, Map<String, String>>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that writes binary encoded protobuf messages of a given type to a
 Google Cloud Pub/Sub stream. 
- Writer(FileBasedSink.WriteOperation<DestinationT, OutputT>, String) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.Writer
- 
- writeRecords() - Static method in class org.apache.beam.sdk.io.kafka.KafkaIO
- 
- WriteRecords() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
-  
- WriteRegistrar() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOTranslation.WriteRegistrar
-  
- WriteRegistrar() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubSubPayloadTranslation.WriteRegistrar
-  
- WriteRegistrar() - Constructor for class org.apache.beam.sdk.io.iceberg.IcebergSchemaTransformTranslation.WriteRegistrar
-  
- WriteRegistrar() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaSchemaTransformTranslation.WriteRegistrar
-  
- WriteRegistrar() - Constructor for class org.apache.beam.sdk.io.kafka.upgrade.KafkaIOTranslation.WriteRegistrar
-  
- WriteResult - Class in org.apache.beam.sdk.io.gcp.bigquery
- 
- writeRowMutations() - Static method in class org.apache.beam.sdk.io.hbase.HBaseIO
-  
- writeRows(String, CSVFormat) - Static method in class org.apache.beam.sdk.io.csv.CsvIO
- 
- writeRows(IcebergCatalogConfig) - Static method in class org.apache.beam.sdk.io.iceberg.IcebergIO
-  
- WriteRows() - Constructor for class org.apache.beam.sdk.io.iceberg.IcebergIO.WriteRows
-  
- writeRows(String) - Static method in class org.apache.beam.sdk.io.json.JsonIO
- 
- writeRows() - Static method in class org.apache.beam.sdk.io.singlestore.SingleStoreIO
- 
Write Beam  Rows to a SingleStoreDB datasource. 
- writeStatement(Schema) - Method in class org.apache.beam.sdk.io.jdbc.JdbcWriteSchemaTransformProvider.JdbcWriteSchemaTransform
-  
- writeStreams() - Static method in class org.apache.beam.sdk.io.redis.RedisIO
- 
Write stream data to a Redis server. 
- WriteStreams() - Constructor for class org.apache.beam.sdk.io.redis.RedisIO.WriteStreams
-  
- WriteStreamServiceImpl(BigQueryOptions) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryServicesImpl.WriteStreamServiceImpl
-  
- writeStrings() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
- 
Returns A  PTransform that writes UTF-8 encoded strings to a Google Cloud Pub/Sub
 stream. 
- WriteSuccessSummary(int, long) - Constructor for class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.WriteSuccessSummary
-  
- WriteSuccessSummary(int, long) - Constructor for class org.apache.beam.sdk.io.gcp.firestore.FirestoreV1.WriteSuccessSummary
-  
- writeTableRows() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
- 
- writeTimeout(Integer) - Method in class org.apache.beam.sdk.io.aws2.common.HttpClientConfiguration.Builder
- 
Milliseconds to wait for a write on a socket before an exception is thrown. 
- writeTimeout() - Method in class org.apache.beam.sdk.io.aws2.common.HttpClientConfiguration
- 
Milliseconds to wait for a write on a socket before an exception is thrown. 
- writeTo(OutputStream, int, int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
- 
Writes lengthbytes starting atoffsetfrom the backing data store to the
 specified output stream.
 
- WriteToMySqlSchemaTransformProvider - Class in org.apache.beam.sdk.io.jdbc.providers
-  
- WriteToMySqlSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.jdbc.providers.WriteToMySqlSchemaTransformProvider
-  
- WriteToOracleSchemaTransformProvider - Class in org.apache.beam.sdk.io.jdbc.providers
-  
- WriteToOracleSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.jdbc.providers.WriteToOracleSchemaTransformProvider
-  
- writeToPort(String, BeamFnApi.RemoteGrpcPort) - Static method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortWrite
- 
Create a  RemoteGrpcPortWrite which writes the  RunnerApi.PCollection with the provided
 Pipeline id to the provided  BeamFnApi.RemoteGrpcPort. 
- WriteToPostgresSchemaTransformProvider - Class in org.apache.beam.sdk.io.jdbc.providers
-  
- WriteToPostgresSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.jdbc.providers.WriteToPostgresSchemaTransformProvider
-  
- WriteToPulsarDoFn - Class in org.apache.beam.sdk.io.pulsar
- 
Transform for writing to Apache Pulsar. 
- WriteToSqlServerSchemaTransformProvider - Class in org.apache.beam.sdk.io.jdbc.providers
-  
- WriteToSqlServerSchemaTransformProvider() - Constructor for class org.apache.beam.sdk.io.jdbc.providers.WriteToSqlServerSchemaTransformProvider
-  
- writeUnwind() - Static method in class org.apache.beam.sdk.io.neo4j.Neo4jIO
- 
Write all rows using a Neo4j Cypher UNWIND cypher statement. 
- WriteUnwind() - Constructor for class org.apache.beam.sdk.io.neo4j.Neo4jIO.WriteUnwind
-  
- writeUserEvent() - Static method in class org.apache.beam.sdk.extensions.ml.RecommendationAIIO
-  
- writeVoid() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
-  
- WriteVoid() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
-  
- WriteWithResults() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteWithResults
-  
- WS - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
-  
- WS - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
-