Skip navigation links
A B C D E F G H I J K L M N O P Q R S T U V W X Y Z _ 

A

abort() - Method in interface org.apache.beam.runners.fnexecution.state.StateDelegator.Registration
De-registers the handler for all future requests for state for the registered process bundle instruction id.
abort(Executor) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
absolute(String, String...) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData.Path
Construct a path from an absolute component path hierarchy.
AbstractBeamCalcRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
BeamRelNode to replace Project and Filter node.
AbstractBeamCalcRel(RelOptCluster, RelTraitSet, RelNode, RexProgram) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.AbstractBeamCalcRel
 
AbstractGetterTypeSupplier() - Constructor for class org.apache.beam.sdk.schemas.AutoValueSchema.AbstractGetterTypeSupplier
 
AbstractLegacyArtifactRetrievalService - Class in org.apache.beam.runners.fnexecution.artifact
An LegacyArtifactRetrievalService that handles everything aside from actually opening the backing resources.
AbstractLegacyArtifactRetrievalService() - Constructor for class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactRetrievalService
 
AbstractLegacyArtifactRetrievalService(Cache<String, ArtifactApi.ProxyManifest>) - Constructor for class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactRetrievalService
 
AbstractLegacyArtifactStagingService - Class in org.apache.beam.runners.fnexecution.artifact
An LegacyArtifactStagingServiceGrpc.LegacyArtifactStagingServiceImplBase that handles everything aside from actually opening the backing resources.
AbstractLegacyArtifactStagingService() - Constructor for class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactStagingService
 
accept(ByteString, Boolean) - Method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
 
accept(T) - Method in class org.apache.beam.sdk.fn.data.BeamFnDataSizeBasedBufferingOutboundObserver
 
accept(T) - Method in class org.apache.beam.sdk.fn.data.BeamFnDataTimeBasedBufferingOutboundObserver
 
accept(T) - Method in interface org.apache.beam.sdk.fn.data.FnDataReceiver
 
accept(T) - Method in class org.apache.beam.sdk.fn.stream.DataStreams.BlockingQueueIterator
 
accept(T1, T2) - Method in interface org.apache.beam.sdk.function.ThrowingBiConsumer
 
accept(T) - Method in interface org.apache.beam.sdk.function.ThrowingConsumer
 
accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierContext
 
accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierListContext
 
accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionContext
 
accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.FieldSpecifierContext
 
accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierContext
 
accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierListContext
 
accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifiedComponentContext
 
accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifyComponentContext
 
accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.SimpleIdentifierContext
 
accept(ParseTreeVisitor<? extends T>) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.WildcardContext
 
accept(SchemaZipFold.Context, Schema.FieldType, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold
Accepts two components, context.parent() is always ROW, MAP, ARRAY or absent.
accept(SchemaZipFold.Context, Optional<Schema.Field>, Optional<Schema.Field>) - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold
Accepts two fields, context.parent() is always ROW.
accessPattern() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.SideInputSpec
 
accumulate(T, T) - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold
Accumulate two results together.
AccumulatingCombineFn() - Constructor for class org.apache.beam.sdk.transforms.Combine.AccumulatingCombineFn
 
accumulatingFiredPanes() - Method in class org.apache.beam.sdk.transforms.windowing.Window
Returns a new Window PTransform that uses the registered WindowFn and Triggering behavior, and that accumulates elements in a pane after they are triggered.
ACCUMULATOR_NAME - Static variable in class org.apache.beam.runners.flink.metrics.FlinkMetricContainer
 
AccumulatorCheckpointingSparkListener() - Constructor for class org.apache.beam.runners.spark.aggregators.AggregatorsAccumulator.AccumulatorCheckpointingSparkListener
 
AccumulatorCheckpointingSparkListener() - Constructor for class org.apache.beam.runners.spark.metrics.MetricsAccumulator.AccumulatorCheckpointingSparkListener
 
ackDeadlineSeconds(PubsubClient.SubscriptionPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
Return the ack deadline, in seconds, for subscription.
ackDeadlineSeconds(PubsubClient.SubscriptionPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
 
ackDeadlineSeconds(PubsubClient.SubscriptionPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
 
ackDeadlineSeconds(PubsubClient.SubscriptionPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
acknowledge(PubsubClient.SubscriptionPath, List<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
Acknowldege messages from subscription with ackIds.
acknowledge(PubsubClient.SubscriptionPath, List<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
 
acknowledge(PubsubClient.SubscriptionPath, List<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
 
acknowledge(PubsubClient.SubscriptionPath, List<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
acquireTaskAttemptIdLock(Configuration, int) - Method in interface org.apache.beam.sdk.io.hadoop.format.ExternalSynchronization
Creates unique TaskAttemptID for given taskId.
acquireTaskAttemptIdLock(Configuration, int) - Method in class org.apache.beam.sdk.io.hadoop.format.HDFSSynchronization
 
acquireTaskIdLock(Configuration) - Method in interface org.apache.beam.sdk.io.hadoop.format.ExternalSynchronization
Creates TaskID with unique id among given job.
acquireTaskIdLock(Configuration) - Method in class org.apache.beam.sdk.io.hadoop.format.HDFSSynchronization
 
add(MetricsContainerStepMap) - Method in class org.apache.beam.runners.flink.metrics.MetricsAccumulator
 
add(NamedAggregators) - Method in class org.apache.beam.runners.spark.aggregators.NamedAggregatorsAccumulator
 
add(MetricsContainerStepMap) - Method in class org.apache.beam.runners.spark.metrics.MetricsContainerStepMapAccumulator
 
add(NamedAggregators) - Method in class org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregatorsAccumulator
 
add(MetricsContainerStepMap) - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsContainerStepMapAccumulator
 
add(String, Broadcast<?>, Coder<?>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.SideInputBroadcast
 
add(int, GlobalWatermarkHolder.SparkWatermarks) - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
 
add(T, long, Coder<T>) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.Sketch
 
add(T, Coder<T>) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.Sketch
 
add(KV<byte[], byte[]>) - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter
 
add(Class<?>, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase.ParameterListBuilder
 
add(Class<?>, String, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase.ParameterListBuilder
 
add(List<ValueInSingleWindow<T>>, TableDataInsertAllResponse.InsertErrors, TableReference, ValueInSingleWindow<TableRow>) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.ErrorContainer
 
add(InputT) - Method in interface org.apache.beam.sdk.state.GroupingState
Add a value to the buffer.
add(Iterable<TimestampedValue<T>>) - Static method in class org.apache.beam.sdk.testing.TestStream.ElementEvent
For internal use only: no backwards compatibility guarantees.
add(long) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn.LargestUnique
Adds a value to the heap, returning whether the value is (large enough to be) in the heap.
add(DisplayData.ItemSpec<?>) - Method in interface org.apache.beam.sdk.transforms.display.DisplayData.Builder
Register the given display item.
addAccum(AccumT) - Method in interface org.apache.beam.sdk.state.CombiningState
Add an accumulator to this state cell.
addAll(Map<Integer, Queue<GlobalWatermarkHolder.SparkWatermarks>>) - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
 
addArray(Collection<T>) - Method in class org.apache.beam.sdk.values.Row.Builder
 
addArray(Object...) - Method in class org.apache.beam.sdk.values.Row.Builder
 
addArrayField(String, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addAttempted(T, BiFunction<T, T, T>) - Method in class org.apache.beam.sdk.metrics.MetricResult
 
addBoolean(Map<String, Object>, String, boolean) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
addBooleanField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addByteArrayField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addByteField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addCollectionToSingletonOutput(PCollection<?>, String, PCollectionView<?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
Adds an output to this CollectionToSingleton Dataflow step, consuming the specified input PValue and producing the specified output PValue.
addCommitted(T, BiFunction<T, T, T>) - Method in class org.apache.beam.sdk.metrics.MetricResult
 
addDataSet(String, DataSet<T>) - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.BatchTranslationContext
 
addDataStream(String, DataStream<T>) - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext
 
addDateTimeField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addDecimalField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addDictionary(Map<String, Object>, String, Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
addDouble(Map<String, Object>, String, Double) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
addDoubleField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addElements(T, T...) - Method in class org.apache.beam.sdk.testing.TestStream.Builder
Adds the specified elements to the source with timestamp equal to the current watermark.
addElements(TimestampedValue<T>, TimestampedValue<T>...) - Method in class org.apache.beam.sdk.testing.TestStream.Builder
Adds the specified elements to the source with the provided timestamps.
addEncodingInput(Coder<?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
Sets the encoding for this Dataflow step.
addErrorForCode(int, String) - Method in class org.apache.beam.sdk.extensions.gcp.util.CustomHttpErrors.Builder
Adds a matcher to log the provided string if the error matches a particular status code.
addErrorForCodeAndUrlContains(int, String, String) - Method in class org.apache.beam.sdk.extensions.gcp.util.CustomHttpErrors.Builder
Adds a matcher to log the provided string if the error matches a particular status code and the url contains a certain string.
addExperiment(ExperimentalOptions, String) - Static method in interface org.apache.beam.sdk.options.ExperimentalOptions
Adds experiment to options if not already present.
addField(Schema.Field) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addField(String, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addFields(List<Schema.Field>) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addFields(Schema.Field...) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
AddFields - Class in org.apache.beam.sdk.schemas.transforms
A transform to add new nullable fields to a PCollection's schema.
AddFields() - Constructor for class org.apache.beam.sdk.schemas.transforms.AddFields
 
AddFields.Inner<T> - Class in org.apache.beam.sdk.schemas.transforms
Inner PTransform for AddFields.
addFloatField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addIfAbsent(T) - Method in interface org.apache.beam.sdk.state.SetState
Ensures a value is a member of the set, returning true if it was added and false otherwise.
addIfNotDefault(DisplayData.ItemSpec<T>, T) - Method in interface org.apache.beam.sdk.transforms.display.DisplayData.Builder
Register the given display item if the value is different than the specified default.
addIfNotNull(DisplayData.ItemSpec<?>) - Method in interface org.apache.beam.sdk.transforms.display.DisplayData.Builder
Register the given display item if the value is not null.
addInput(String, Boolean) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
Adds an input with the given name and value to this Dataflow step.
addInput(String, String) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
Adds an input with the given name and value to this Dataflow step.
addInput(String, Long) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
Adds an input with the given name and value to this Dataflow step.
addInput(String, PInput) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
Adds an input with the given name to this Dataflow step, coming from the specified input PValue.
addInput(String, Map<String, Object>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
Adds an input that is a dictionary of strings to objects.
addInput(String, List<? extends Map<String, Object>>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
Adds an input that is a list of objects.
addInput(HyperLogLogPlus, InputT) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.ApproximateDistinctFn
 
addInput(SketchFrequencies.Sketch<InputT>, InputT) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.CountMinSketchFn
 
addInput(MergingDigest, Double) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.TDigestQuantilesFn
 
addInput(CovarianceAccumulator, Row) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CovarianceFn
 
addInput(VarianceAccumulator, T) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.VarianceFn
 
addInput(ApproximateUnique.ApproximateUniqueCombineFn.LargestUnique, T) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn
 
addInput(InputT) - Method in interface org.apache.beam.sdk.transforms.Combine.AccumulatingCombineFn.Accumulator
Adds the given input value to this accumulator, modifying this accumulator.
addInput(AccumT, InputT) - Method in class org.apache.beam.sdk.transforms.Combine.AccumulatingCombineFn
 
addInput(double[], Double) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
 
addInput(Combine.Holder<V>, V) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
 
addInput(int[], Integer) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
 
addInput(long[], Long) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
 
addInput(AccumT, InputT) - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
Adds the given input value to the given accumulator, returning the new accumulator value.
addInput(List<V>, V) - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
 
addInput(Object[], DataT) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
 
addInput(Object[], DataT, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
 
addInput(AccumT, InputT, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
Adds the given input value to the given accumulator, returning the new accumulator value.
addInput(Top.BoundedHeap<KV<Integer, T>, SerializableComparator<KV<Integer, T>>>, T) - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
 
addInt16Field(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addInt32Field(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addInt64Field(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addIterable(Iterable<T>) - Method in class org.apache.beam.sdk.values.Row.Builder
 
addIterableField(String, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addLengthPrefixedCoder(String, RunnerApi.Components.Builder, boolean) - Static method in class org.apache.beam.runners.fnexecution.wire.LengthPrefixUnknownCoders
Recursively traverses the coder tree and wraps the first unknown coder in every branch with a LengthPrefixCoder unless an ancestor coder is itself a LengthPrefixCoder.
addList(Map<String, Object>, String, List<? extends Map<String, Object>>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
addList(Map<String, Object>, String, T[]) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
addLogicalTypeField(String, Schema.LogicalType<InputT, BaseT>) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addLong(Map<String, Object>, String, long) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
addLongs(Map<String, Object>, String, long...) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
addMapField(String, Schema.FieldType, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addMessageListener(Consumer<JobApi.JobMessage>) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation
Listen for job messages with a Consumer.
addMethodParameters(Method) - Method in class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase.ParameterListBuilder
 
addNameFilter(MetricNameFilter) - Method in class org.apache.beam.sdk.metrics.MetricsFilter.Builder
addNull(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
addNullableField(String, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addObject(Map<String, Object>, String, Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
addOptions(Schema.Options) - Method in class org.apache.beam.sdk.schemas.Schema.Options.Builder
 
addOutput(String, PCollection<?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.StepTranslationContext
Adds a primitive output to this Dataflow step with the given name as the local output name, producing the specified output PValue, including its Coder if a TypedPValue.
addOutputColumnList(List<ResolvedNodes.ResolvedOutputColumn>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
 
addOverrideForClass(Class<?>, DataflowWorkerLoggingOptions.Level) - Method in class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
Deprecated.
Overrides the default log level for the passed in class.
addOverrideForClass(Class<?>, SdkHarnessOptions.LogLevel) - Method in class org.apache.beam.sdk.options.SdkHarnessOptions.SdkHarnessLogLevelOverrides
Overrides the default log level for the passed in class.
addOverrideForName(String, DataflowWorkerLoggingOptions.Level) - Method in class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
Deprecated.
Overrides the default log level for the passed in name.
addOverrideForName(String, SdkHarnessOptions.LogLevel) - Method in class org.apache.beam.sdk.options.SdkHarnessOptions.SdkHarnessLogLevelOverrides
Overrides the default log logLevel for the passed in name.
addOverrideForPackage(Package, DataflowWorkerLoggingOptions.Level) - Method in class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
Deprecated.
Overrides the default log level for the passed in package.
addOverrideForPackage(Package, SdkHarnessOptions.LogLevel) - Method in class org.apache.beam.sdk.options.SdkHarnessOptions.SdkHarnessLogLevelOverrides
Overrides the default log level for the passed in package.
addProperties(Configuration, Properties) - Static method in class org.apache.beam.runners.apex.ApexYarnLauncher
Transfer the properties to the configuration object.
addResolvedTable(TableResolution.SimpleTableWithPath) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
Store a table together with its full path for repeated resolutions.
addRowField(String, Schema) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addRows(Object...) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestBoundedTable
Add rows to the builder.
addRows(String, Row...) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
 
addRows(Duration, Object...) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestUnboundedTable
Add rows to the builder.
addRunnerWireCoder(PipelineNode.PCollectionNode, RunnerApi.Components.Builder, RunnerApi.ExecutableStagePayload.WireCoderSetting) - Static method in class org.apache.beam.runners.fnexecution.wire.WireCoders
Creates a runner-side wire coder for a port read/write for the given PCollection.
addSchema(String, TableProvider) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
Add a top-level schema backed by the table provider.
addSdkWireCoder(PipelineNode.PCollectionNode, RunnerApi.Components.Builder, RunnerApi.ExecutableStagePayload.WireCoderSetting) - Static method in class org.apache.beam.runners.fnexecution.wire.WireCoders
Creates an SDK-side wire coder for a port read/write for the given PCollection.
addStateListener(Consumer<JobApi.JobStateEvent>) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation
Listen for job state changes with a Consumer.
addStep(PTransform<?, ?>, String) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
Adds a step to the Dataflow workflow for the given transform, with the given Dataflow step type.
addStep(String) - Method in class org.apache.beam.sdk.metrics.MetricsFilter.Builder
Add a step filter.
addString(Map<String, Object>, String, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
addStringField(String) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
addStringList(Map<String, Object>, String, List<String>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
addTraceFor(AbstractGoogleClient, String) - Method in class org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions.GoogleApiTracer
Creates a GoogleApiDebugOptions.GoogleApiTracer that sets the trace destination on all calls that match the given client type.
addTraceFor(AbstractGoogleClientRequest<?>, String) - Method in class org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions.GoogleApiTracer
Creates a GoogleApiDebugOptions.GoogleApiTracer that sets the trace traceDestination on all calls that match for the given request type.
addUdaf(String, Combine.CombineFn) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
Register a UDAF function which can be used in GROUP-BY expression.
addUdf(String, Class<?>, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
Register a UDF function which can be used in SQL expression.
addUdf(String, Class<? extends BeamSqlUdf>) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
Register a UDF function which can be used in SQL expression.
addUdf(String, SerializableFunction) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
Register a UDF function which can be used in SQL expression.
addValue(Object) - Method in class org.apache.beam.sdk.values.Row.Builder
 
addValues(List<Object>) - Method in class org.apache.beam.sdk.values.Row.Builder
 
addValues(Object...) - Method in class org.apache.beam.sdk.values.Row.Builder
 
advance() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
 
advance() - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
advance() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource.BigQueryStorageStreamReader
 
advance() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
For subscription mode only: Track progression of time according to the Clock passed .
advance() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
 
advance() - Method in class org.apache.beam.sdk.io.Source.Reader
Advances the reader to the next valid record.
advance() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
Advances the reader to the next valid record.
advanceBy(Duration) - Static method in class org.apache.beam.sdk.testing.TestStream.ProcessingTimeEvent
For internal use only: no backwards compatibility guarantees.
advanceImpl() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
 
advanceImpl() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
Advances to the next record and returns true, or returns false if there is no next record.
advanceNextBatchWatermarkToInfinity() - Method in class org.apache.beam.runners.spark.io.CreateStream
Advances the watermark in the next batch to the end-of-time.
advanceProcessingTime(Duration) - Method in class org.apache.beam.sdk.testing.TestStream.Builder
Advance the processing time by the specified amount.
advanceTo(Instant) - Static method in class org.apache.beam.sdk.testing.TestStream.WatermarkEvent
For internal use only: no backwards compatibility guarantees.
advanceWatermark() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
Advances the watermark.
advanceWatermarkForNextBatch(Instant) - Method in class org.apache.beam.runners.spark.io.CreateStream
Advances the watermark in the next batch.
advanceWatermarkTo(Instant) - Method in class org.apache.beam.sdk.testing.TestStream.Builder
Advance the watermark of this source to the specified instant.
advanceWatermarkToInfinity() - Method in class org.apache.beam.sdk.testing.TestStream.Builder
Advance the watermark to infinity, completing this TestStream.
AdvancingPhaser - Class in org.apache.beam.sdk.fn.stream
A Phaser which never terminates.
AdvancingPhaser(int) - Constructor for class org.apache.beam.sdk.fn.stream.AdvancingPhaser
 
AfterAll - Class in org.apache.beam.sdk.transforms.windowing
A composite Trigger that fires when all of its sub-triggers are ready.
afterBundleCommit(Instant, DoFn.BundleFinalizer.Callback) - Method in interface org.apache.beam.sdk.transforms.DoFn.BundleFinalizer
The provided function will be called after the runner successfully commits the output of a successful bundle.
AfterEach - Class in org.apache.beam.sdk.transforms.windowing
A composite Trigger that executes its sub-triggers in order.
AfterFirst - Class in org.apache.beam.sdk.transforms.windowing
A composite Trigger that fires once after at least one of its sub-triggers have fired.
AfterPane - Class in org.apache.beam.sdk.transforms.windowing
A Trigger that fires at some point after a specified number of input elements have arrived.
AfterProcessingTime - Class in org.apache.beam.sdk.transforms.windowing
A Trigger trigger that fires at a specified point in processing time, relative to when input first arrives.
AfterSynchronizedProcessingTime - Class in org.apache.beam.sdk.transforms.windowing
FOR INTERNAL USE ONLY.
afterTimeSinceNewOutput(ReadableDuration) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
Returns a Watch.Growth.TerminationCondition that holds after the given time has elapsed after the last time the Watch.Growth.PollResult for the current input contained a previously unseen output.
afterTimeSinceNewOutput(SerializableFunction<InputT, ReadableDuration>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
Like Watch.Growth.afterTimeSinceNewOutput(ReadableDuration), but the duration is input-dependent.
afterTotalOf(ReadableDuration) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
Returns a Watch.Growth.TerminationCondition that holds after the given time has elapsed after the current input was seen.
afterTotalOf(SerializableFunction<InputT, ReadableDuration>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
Like Watch.Growth.afterTotalOf(ReadableDuration), but the duration is input-dependent.
AfterWatermark - Class in org.apache.beam.sdk.transforms.windowing
AfterWatermark triggers fire based on progress of the system watermark.
AfterWatermark.AfterWatermarkEarlyAndLate - Class in org.apache.beam.sdk.transforms.windowing
 
AfterWatermark.FromEndOfWindow - Class in org.apache.beam.sdk.transforms.windowing
A watermark trigger targeted relative to the end of the window.
aggregate(Combine.CombineFn<InputT, ?, OutputT>) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
Aggregate the grouped data using the specified Combine.CombineFn.
aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
Build up an aggregation function over the input elements.
aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
 
aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
Build up an aggregation function over the input elements.
aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
 
aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
Build up an aggregation function over the input elements.
aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
 
aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
Build up an aggregation function over the input elements.
aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
 
aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
Build up an aggregation function over the input elements.
aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
 
aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
Build up an aggregation function over the input elements.
aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
 
aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
Build up an aggregation function over the input elements.
aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
aggregateField(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
Build up an aggregation function over the input elements.
aggregateField(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
 
aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
 
aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
 
aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
 
aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
 
aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
 
aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
 
aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
 
aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
 
aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
 
aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
 
aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
 
aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
 
aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
 
aggregateFieldBaseValue(String, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
 
aggregateFieldBaseValue(int, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
 
aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
Build up an aggregation function over the input elements.
aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
Build up an aggregation function over the input elements.
aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
Build up an aggregation function over the input elements.
aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
Build up an aggregation function over the input elements.
aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
Build up an aggregation function over the input elements.
aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
Build up an aggregation function over the input elements.
aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
Build up an aggregation function over the input elements.
aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
Build up an aggregation function over the input elements.
aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
Build up an aggregation function over the input elements.
aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
Build up an aggregation function over the input elements.
aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
Build up an aggregation function over the input elements.
aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
Build up an aggregation function over the input elements.
aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
Build up an aggregation function over the input elements.
aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
Build up an aggregation function over the input elements.
aggregateFields(List<String>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
Build up an aggregation function over the input elements.
aggregateFields(FieldAccessDescriptor, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
Build up an aggregation function over the input elements.
aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
 
aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
 
aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
 
aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
 
aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
 
aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
aggregateFieldsById(List<Integer>, Combine.CombineFn<CombineInputT, AccumT, CombineOutputT>, Schema.Field) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
 
AggregationCombineFnAdapter<T> - Class in org.apache.beam.sdk.extensions.sql.impl.transform.agg
Wrapper Combine.CombineFns for aggregation function calls.
AggregationCombineFnAdapter() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.agg.AggregationCombineFnAdapter
 
AggregationQuery - Class in org.apache.beam.sdk.io.mongodb
Builds a MongoDB AggregateIterable object.
AggregationQuery() - Constructor for class org.apache.beam.sdk.io.mongodb.AggregationQuery
 
AggregatorMetric - Class in org.apache.beam.runners.spark.metrics
An adapter between the NamedAggregators and Codahale's Metric interface.
AggregatorMetric - Class in org.apache.beam.runners.spark.structuredstreaming.metrics
An adapter between the NamedAggregators and Codahale's Metric interface.
AggregatorMetricSource - Class in org.apache.beam.runners.spark.metrics
A Spark Source that is tailored to expose an AggregatorMetric, wrapping an underlying NamedAggregators instance.
AggregatorMetricSource(String, NamedAggregators) - Constructor for class org.apache.beam.runners.spark.metrics.AggregatorMetricSource
 
AggregatorMetricSource - Class in org.apache.beam.runners.spark.structuredstreaming.metrics
A Spark Source that is tailored to expose an AggregatorMetric, wrapping an underlying NamedAggregators instance.
AggregatorMetricSource(String, NamedAggregators) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.metrics.AggregatorMetricSource
 
AggregatorsAccumulator - Class in org.apache.beam.runners.spark.aggregators
For resilience, Accumulators are required to be wrapped in a Singleton.
AggregatorsAccumulator() - Constructor for class org.apache.beam.runners.spark.aggregators.AggregatorsAccumulator
 
AggregatorsAccumulator - Class in org.apache.beam.runners.spark.structuredstreaming.aggregators
For resilience, Accumulators are required to be wrapped in a Singleton.
AggregatorsAccumulator() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.aggregators.AggregatorsAccumulator
 
AggregatorsAccumulator.AccumulatorCheckpointingSparkListener - Class in org.apache.beam.runners.spark.aggregators
Spark Listener which checkpoints NamedAggregators values for fault-tolerance.
align(Duration) - Method in interface org.apache.beam.sdk.state.Timer
Aligns the target timestamp used by Timer.setRelative() to the next boundary of period.
alignedTo(Duration, Instant) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
Aligns timestamps to the smallest multiple of period since the offset greater than the timestamp.
alignedTo(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
Aligns the time to be the smallest multiple of period greater than the epoch boundary (aka new Instant(0)).
alignTo(Duration, Instant) - Static method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform
For internal use only; no backwards-compatibility guarantees.
alignTo(Duration) - Static method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform
For internal use only; no backwards-compatibility guarantees.
AlignTo() - Constructor for class org.apache.beam.sdk.transforms.windowing.TimestampTransform.AlignTo
 
ALL_CONTEXTS - Static variable in class org.apache.beam.sdk.testing.CoderProperties
All the contexts, for use in test cases.
ALL_KEYS - Static variable in class org.apache.beam.sdk.io.range.ByteKeyRange
The range of all keys, with empty start and end keys.
allLeavesDescriptor(Schema, SerializableFunction<List<String>, String>) - Static method in class org.apache.beam.sdk.schemas.utils.SelectHelpers
 
allMatches(String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.AllMatches PTransform that checks if the entire line matches the Regex.
allMatches(Pattern) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.AllMatches PTransform that checks if the entire line matches the Regex.
AllMatches(Pattern) - Constructor for class org.apache.beam.sdk.transforms.Regex.AllMatches
 
allMetrics() - Method in class org.apache.beam.sdk.metrics.MetricResults
 
allocateAddressAndCreate(List<BindableService>, Endpoints.ApiServiceDescriptor.Builder) - Method in class org.apache.beam.runners.fnexecution.InProcessServerFactory
 
allocateAddressAndCreate(List<BindableService>, Endpoints.ApiServiceDescriptor.Builder) - Method in class org.apache.beam.runners.fnexecution.ServerFactory
Creates an instance of this server using an ephemeral address.
allocateAddressAndCreate(List<BindableService>, Endpoints.ApiServiceDescriptor.Builder) - Method in class org.apache.beam.runners.fnexecution.ServerFactory.InetSocketAddressServerFactory
 
allocatePortAndCreateFor(ServiceT, ServerFactory) - Static method in class org.apache.beam.runners.fnexecution.GrpcFnServer
Create a GrpcFnServer for the provided FnService running on an arbitrary port.
allOf(Iterable<SerializableMatcher<? super T>>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.allOf(Iterable).
allOf(SerializableMatcher<T>...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.allOf(Matcher[]).
allOf(Watch.Growth.TerminationCondition<InputT, FirstStateT>, Watch.Growth.TerminationCondition<InputT, SecondStateT>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
Returns a Watch.Growth.TerminationCondition that holds when both of the given two conditions hold.
allowsDynamicSplitting() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
 
allowsDynamicSplitting() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
 
allowsDynamicSplitting() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
Whether this reader should allow dynamic splitting of the offset ranges.
AlwaysPassMatcher() - Constructor for class org.apache.beam.sdk.testing.TestPipelineOptions.AlwaysPassMatcher
 
AlwaysPassMatcherFactory() - Constructor for class org.apache.beam.sdk.testing.TestPipelineOptions.AlwaysPassMatcherFactory
 
alwaysRetry() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy
Always retry all failures.
AmqpIO - Class in org.apache.beam.sdk.io.amqp
AmqpIO supports AMQP 1.0 protocol using the Apache QPid Proton-J library.
AmqpIO.Read - Class in org.apache.beam.sdk.io.amqp
A PTransform to read/receive messages using AMQP 1.0 protocol.
AmqpIO.Write - Class in org.apache.beam.sdk.io.amqp
A PTransform to send messages using AMQP 1.0 protocol.
AmqpMessageCoder - Class in org.apache.beam.sdk.io.amqp
A coder for AMQP message.
AmqpMessageCoder() - Constructor for class org.apache.beam.sdk.io.amqp.AmqpMessageCoder
 
AmqpMessageCoderProviderRegistrar - Class in org.apache.beam.sdk.io.amqp
A CoderProviderRegistrar for standard types used with AmqpIO.
AmqpMessageCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.amqp.AmqpMessageCoderProviderRegistrar
 
and(TupleTag<V>, List<V>) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
Returns a new CoGbkResult based on this, with the given tag and given data added to it.
and(TupleTag<V>, PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
Returns a new KeyedPCollectionTuple<K> that is the same as this, appended with the given PCollection.
and(String, PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
A version of KeyedPCollectionTuple.and(String, PCollection) that takes in a string instead of a TupleTag.
and(PCollection.IsBounded) - Method in enum org.apache.beam.sdk.values.PCollection.IsBounded
Returns the composed IsBounded property.
and(PCollection<T>) - Method in class org.apache.beam.sdk.values.PCollectionList
Returns a new PCollectionList that has all the PCollections of this PCollectionList plus the given PCollection appended to the end.
and(Iterable<PCollection<T>>) - Method in class org.apache.beam.sdk.values.PCollectionList
Returns a new PCollectionList that has all the PCollections of this PCollectionList plus the given PCollections appended to the end, in order.
and(TupleTag<T>, PCollection<T>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
Returns a new PCollectionTuple that has each PCollection and TupleTag of this PCollectionTuple plus the given PCollection associated with the given TupleTag.
and(String, PCollection<T>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
A version of PCollectionTuple.and(TupleTag, PCollection) that takes in a String instead of a TupleTag.
and(TupleTag<?>) - Method in class org.apache.beam.sdk.values.TupleTagList
Returns a new TupleTagList that has all the TupleTags of this TupleTagList plus the given TupleTag appended to the end.
and(List<TupleTag<?>>) - Method in class org.apache.beam.sdk.values.TupleTagList
Returns a new TupleTagList that has all the TupleTags of this TupleTagList plus the given TupleTags appended to the end, in order.
annotateFromBytes(PCollectionView<Map<ByteString, VideoContext>>, List<Feature>) - Static method in class org.apache.beam.sdk.extensions.ml.VideoIntelligence
Annotates videos from ByteStrings of their contents.
annotateFromBytesWithContext(List<Feature>) - Static method in class org.apache.beam.sdk.extensions.ml.VideoIntelligence
Annotates videos from key-value pairs of ByteStrings and VideoContext.
annotateFromURI(List<Feature>, PCollectionView<Map<String, VideoContext>>) - Static method in class org.apache.beam.sdk.extensions.ml.VideoIntelligence
Annotates videos from GCS URIs.
annotateFromUriWithContext(List<Feature>) - Static method in class org.apache.beam.sdk.extensions.ml.VideoIntelligence
Annotates videos from key-value pairs of GCS URI and VideoContext.
AnnotateVideo<T> - Class in org.apache.beam.sdk.extensions.ml
Base class for Video Intelligence transform.
AnnotateVideo(PCollectionView<Map<T, VideoContext>>, List<Feature>) - Constructor for class org.apache.beam.sdk.extensions.ml.AnnotateVideo
 
AnnotateVideo(List<Feature>) - Constructor for class org.apache.beam.sdk.extensions.ml.AnnotateVideo
 
AnnotateVideoBytesWithContext(List<Feature>) - Constructor for class org.apache.beam.sdk.extensions.ml.VideoIntelligence.AnnotateVideoBytesWithContext
 
AnnotateVideoFromBytes(PCollectionView<Map<ByteString, VideoContext>>, List<Feature>) - Constructor for class org.apache.beam.sdk.extensions.ml.VideoIntelligence.AnnotateVideoFromBytes
 
AnnotateVideoFromURI(PCollectionView<Map<String, VideoContext>>, List<Feature>) - Constructor for class org.apache.beam.sdk.extensions.ml.VideoIntelligence.AnnotateVideoFromURI
 
AnnotateVideoURIWithContext(List<Feature>) - Constructor for class org.apache.beam.sdk.extensions.ml.VideoIntelligence.AnnotateVideoURIWithContext
 
any(long) - Static method in class org.apache.beam.sdk.transforms.Sample
Sample#any(long) takes a PCollection<T> and a limit, and produces a new PCollection<T> containing up to limit elements of the input PCollection.
anyCombineFn(int) - Static method in class org.apache.beam.sdk.transforms.Sample
Returns a Combine.CombineFn that computes a fixed-sized potentially non-uniform sample of its inputs.
anyOf(Iterable<SerializableMatcher<? super T>>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.anyOf(Iterable).
anyOf(SerializableMatcher<T>...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.anyOf(Matcher[]).
anything() - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.anything().
ApexPipelineOptions - Interface in org.apache.beam.runners.apex
Options that configure the Apex pipeline.
ApexRunner - Class in org.apache.beam.runners.apex
A PipelineRunner that translates the pipeline to an Apex DAG and executes it on an Apex cluster.
ApexRunner(ApexPipelineOptions) - Constructor for class org.apache.beam.runners.apex.ApexRunner
 
ApexRunner.CreateApexPCollectionView<ElemT,ViewT> - Class in org.apache.beam.runners.apex
Creates a primitive PCollectionView.
ApexRunnerRegistrar - Class in org.apache.beam.runners.apex
Contains the PipelineRunnerRegistrar and PipelineOptionsRegistrar for the ApexRunner.
ApexRunnerRegistrar.Options - Class in org.apache.beam.runners.apex
Registers the ApexPipelineOptions.
ApexRunnerRegistrar.Runner - Class in org.apache.beam.runners.apex
Registers the ApexRunner.
ApexRunnerResult - Class in org.apache.beam.runners.apex
Result of executing a Pipeline with Apex in embedded mode.
ApexRunnerResult(DAG, Launcher.AppHandle) - Constructor for class org.apache.beam.runners.apex.ApexRunnerResult
 
ApexYarnLauncher - Class in org.apache.beam.runners.apex
Proxy to launch the YARN application through the hadoop script to run in the pre-configured environment (class path, configuration, native libraries etc.).
ApexYarnLauncher() - Constructor for class org.apache.beam.runners.apex.ApexYarnLauncher
 
ApexYarnLauncher.LaunchParams - Class in org.apache.beam.runners.apex
Launch parameters that will be serialized and passed to the child process.
ApexYarnLauncher.ProcessWatcher - Class in org.apache.beam.runners.apex
Starts a command and waits for it to complete.
append(K, W, Iterator<V>) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.BagUserStateHandler
Appends the values to the bag user state for the given key and window.
applicableTo(PCollection<?>) - Static method in class org.apache.beam.sdk.transforms.GroupByKey
 
ApplicationNameOptions - Interface in org.apache.beam.sdk.options
Options that allow setting the application name.
apply(WindowFunction.Context<T2>) - Method in class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator.GearpumpWindowFn
 
apply(Tuple2<ByteArray, byte[]>) - Method in class org.apache.beam.runners.spark.coders.CoderHelpers.FromByteFunction
 
apply(KV<String, Long>) - Method in class org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.FormatAsTextFn
 
apply(InputT) - Method in interface org.apache.beam.sdk.coders.DelegateCoder.CodingFunction
 
apply(Pipeline, String, RunnerApi.FunctionSpec, Map<String, PCollection<?>>) - Method in interface org.apache.beam.sdk.expansion.service.ExpansionService.TransformProvider
 
apply(Row) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSetOperatorsTransforms.BeamSqlRow2KvFn
 
apply(RexBuilder, List<RexNode>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.SqlCaseWithValueOperatorRewriter
 
apply(RexBuilder, List<RexNode>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.SqlCoalesceOperatorRewriter
 
apply(RexBuilder, List<RexNode>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.SqlExtractTimestampOperatorRewriter
 
apply(RexBuilder, List<RexNode>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.SqlIfNullOperatorRewriter
 
apply(RexBuilder, List<RexNode>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.SqlNullIfOperatorRewriter
 
apply(RexBuilder, List<RexNode>) - Method in interface org.apache.beam.sdk.extensions.sql.zetasql.SqlOperatorRewriter
Create and return a new RexNode that represents a call to this operator with the specified operands.
apply(T1, T2) - Method in interface org.apache.beam.sdk.function.ThrowingBiFunction
 
apply(T1) - Method in interface org.apache.beam.sdk.function.ThrowingFunction
 
apply(Statement, Description) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery
 
apply(Statement, Description) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
 
apply(Statement, Description) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal
 
apply(Void) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceProviderFromDataSourceConfiguration
 
apply(SQLException) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DefaultRetryStrategy
 
apply(Void) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.PoolableDataSourceProvider
 
apply(SQLException) - Method in interface org.apache.beam.sdk.io.jdbc.JdbcIO.RetryStrategy
 
apply(MongoCollection<Document>) - Method in class org.apache.beam.sdk.io.mongodb.AggregationQuery
 
apply(MongoCollection<Document>) - Method in class org.apache.beam.sdk.io.mongodb.FindQuery
 
apply(PTransform<? super PBegin, OutputT>) - Method in class org.apache.beam.sdk.Pipeline
Like Pipeline.apply(String, PTransform) but the transform node in the Pipeline graph will be named according to PTransform.getName().
apply(String, PTransform<? super PBegin, OutputT>) - Method in class org.apache.beam.sdk.Pipeline
Adds a root PTransform, such as Read or Create, to this Pipeline.
apply(Schema, Schema) - Method in class org.apache.beam.sdk.schemas.transforms.Cast.Narrowing
 
apply(Schema, Schema) - Method in interface org.apache.beam.sdk.schemas.transforms.Cast.Validator
 
apply(Schema, Schema) - Method in class org.apache.beam.sdk.schemas.transforms.Cast.Widening
 
apply(Schema, Schema) - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold
 
apply(T) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert.MatcherCheckerFn
 
apply(Statement, Description) - Method in class org.apache.beam.sdk.testing.TestPipeline
 
apply(double, double) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
Applies the binary operation to the two operands, returning the result.
apply(V, V) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
Applies the binary operation to the two operands, returning the result.
apply(int, int) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
Applies the binary operation to the two operands, returning the result.
apply(long, long) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
Applies the binary operation to the two operands, returning the result.
apply(Iterable<? extends InputT>) - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
Applies this CombineFn to a collection of input values to produce a combined output value.
apply(Iterable<? extends InputT>, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
Applies this CombineFnWithContext to a collection of input values to produce a combined output value.
apply(InputT, Contextful.Fn.Context) - Method in interface org.apache.beam.sdk.transforms.Contextful.Fn
Invokes the function on the given input with the given context.
apply(InputT) - Method in class org.apache.beam.sdk.transforms.InferableFunction
 
apply(PTransform<KeyedPCollectionTuple<K>, OutputT>) - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
Like KeyedPCollectionTuple.apply(String, PTransform) but defaulting to the name provided by the PTransform.
apply(String, PTransform<KeyedPCollectionTuple<K>, OutputT>) - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
Applies the given PTransform to this input KeyedPCollectionTuple and returns its OutputT.
apply(InputT) - Method in interface org.apache.beam.sdk.transforms.ProcessFunction
Returns the result of invoking this function on the given input.
apply(InputT) - Method in interface org.apache.beam.sdk.transforms.SerializableFunction
Returns the result of invoking this function on the given input.
apply(InputT) - Method in class org.apache.beam.sdk.transforms.SimpleFunction
 
apply(PrimitiveViewT) - Method in class org.apache.beam.sdk.transforms.ViewFn
A function to adapt a primitive view type to a desired view type.
apply(WithFailures.ExceptionElement<T>) - Method in class org.apache.beam.sdk.transforms.WithFailures.ExceptionAsMapHandler
 
apply(PTransform<? super PBegin, OutputT>) - Method in class org.apache.beam.sdk.values.PBegin
Like PBegin.apply(String, PTransform) but defaulting to the name of the PTransform.
apply(String, PTransform<? super PBegin, OutputT>) - Method in class org.apache.beam.sdk.values.PBegin
Applies the given PTransform to this PBegin, using name to identify this specific application of the transform.
apply(PTransform<? super PCollection<T>, OutputT>) - Method in class org.apache.beam.sdk.values.PCollection
of the PTransform.
apply(String, PTransform<? super PCollection<T>, OutputT>) - Method in class org.apache.beam.sdk.values.PCollection
Applies the given PTransform to this input PCollection, using name to identify this specific application of the transform.
apply(PTransform<PCollectionList<T>, OutputT>) - Method in class org.apache.beam.sdk.values.PCollectionList
Like PCollectionList.apply(String, PTransform) but defaulting to the name of the PTransform.
apply(String, PTransform<PCollectionList<T>, OutputT>) - Method in class org.apache.beam.sdk.values.PCollectionList
Applies the given PTransform to this input PCollectionList, using name to identify this specific application of the transform.
apply(PTransform<? super PCollectionTuple, OutputT>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
Like PCollectionTuple.apply(String, PTransform) but defaulting to the name of the PTransform.
apply(String, PTransform<? super PCollectionTuple, OutputT>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
Applies the given PTransform to this input PCollectionTuple, using name to identify this specific application of the transform.
apply(Materializations.MultimapView<Void, T>) - Method in class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn
 
apply(Materializations.MultimapView<Void, T>) - Method in class org.apache.beam.sdk.values.PCollectionViews.ListViewFn
 
apply(Materializations.MultimapView<Void, KV<K, V>>) - Method in class org.apache.beam.sdk.values.PCollectionViews.MapViewFn
 
apply(Materializations.MultimapView<Void, KV<K, V>>) - Method in class org.apache.beam.sdk.values.PCollectionViews.MultimapViewFn
 
apply(Materializations.MultimapView<Void, T>) - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn
 
applyTransform(InputT, PTransform<? super InputT, OutputT>) - Static method in class org.apache.beam.sdk.Pipeline
For internal use only; no backwards-compatibility guarantees.
applyTransform(String, InputT, PTransform<? super InputT, OutputT>) - Static method in class org.apache.beam.sdk.Pipeline
For internal use only; no backwards-compatibility guarantees.
ApproximateDistinct - Class in org.apache.beam.sdk.extensions.sketching
PTransforms for computing the approximate number of distinct elements in a stream.
ApproximateDistinct() - Constructor for class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct
 
ApproximateDistinct.ApproximateDistinctFn<InputT> - Class in org.apache.beam.sdk.extensions.sketching
Implements the Combine.CombineFn of ApproximateDistinct transforms.
ApproximateDistinct.GloballyDistinct<InputT> - Class in org.apache.beam.sdk.extensions.sketching
ApproximateDistinct.HyperLogLogPlusCoder - Class in org.apache.beam.sdk.extensions.sketching
Coder for HyperLogLogPlus class.
ApproximateDistinct.PerKeyDistinct<K,V> - Class in org.apache.beam.sdk.extensions.sketching
Implementation of ApproximateDistinct.perKey().
ApproximateQuantiles - Class in org.apache.beam.sdk.transforms
PTransforms for getting an idea of a PCollection's data distribution using approximate N-tiles (e.g.
ApproximateQuantiles.ApproximateQuantilesCombineFn<T,ComparatorT extends java.util.Comparator<T> & java.io.Serializable> - Class in org.apache.beam.sdk.transforms
The ApproximateQuantilesCombineFn combiner gives an idea of the distribution of a collection of values using approximate N-tiles.
ApproximateUnique - Class in org.apache.beam.sdk.transforms
PTransforms for estimating the number of distinct elements in a PCollection, or the number of distinct values associated with each key in a PCollection of KVs.
ApproximateUnique() - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique
 
ApproximateUnique.ApproximateUniqueCombineFn<T> - Class in org.apache.beam.sdk.transforms
CombineFn that computes an estimate of the number of distinct values that were combined.
ApproximateUnique.ApproximateUniqueCombineFn.LargestUnique - Class in org.apache.beam.sdk.transforms
A heap utility class to efficiently track the largest added elements.
ApproximateUnique.Globally<T> - Class in org.apache.beam.sdk.transforms
PTransform for estimating the number of distinct elements in a PCollection.
ApproximateUnique.PerKey<K,V> - Class in org.apache.beam.sdk.transforms
PTransform for estimating the number of distinct values associated with each key in a PCollection of KVs.
ApproximateUniqueCombineFn(long, Coder<T>) - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn
 
array() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
Returns the backing array.
array(TableSchema.ColumnType) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
array(Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
Create an array type for the given field type.
array(Schema.FieldType, boolean) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
Deprecated.
Set the nullability on the elementType instead
arrayContaining(T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
arrayContaining(Coder<T>, T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
arrayContaining(SerializableMatcher<? super T>...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
arrayContaining(List<SerializableMatcher<? super T>>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
arrayContainingInAnyOrder(T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
arrayContainingInAnyOrder(Coder<T>, T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
arrayContainingInAnyOrder(SerializableMatcher<? super T>...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
arrayContainingInAnyOrder(Collection<SerializableMatcher<? super T>>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
arrayElementType() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
arrayQualifier() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
arrayQualifier() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierListContext
 
ArrayQualifierContext(ParserRuleContext, int) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierContext
 
ArrayQualifierListContext(FieldSpecifierNotationParser.QualifierListContext) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierListContext
 
arrayWithSize(int) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.arrayWithSize(int).
arrayWithSize(SerializableMatcher<? super Integer>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
ARTIFACTS - Static variable in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService
 
ARTIFACTS - Static variable in class org.apache.beam.runners.fnexecution.artifact.JavaFilesystemLegacyArtifactStagingService
 
as(Class<T>) - Method in interface org.apache.beam.sdk.options.PipelineOptions
Transforms this object into an object of type <T> saving each property that has been manipulated.
as(Class<T>) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
Creates and returns an object that implements <T>.
as(Class<T>) - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
Creates and returns an object that implements <T> using the values configured on this builder during construction.
asCloudObject(Coder<?>, SdkComponents) - Static method in class org.apache.beam.runners.dataflow.util.CloudObjects
Convert the provided Coder into a CloudObject.
asInputStream(int, int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
Returns an InputStream wrapper which supplies the portion of this backing byte buffer starting at offset and up to length bytes.
asIterable() - Static method in class org.apache.beam.sdk.transforms.View
Returns a View.AsIterable transform that takes a PCollection as input and produces a PCollectionView mapping each window to an Iterable of the values in that window.
AsJsons<InputT> - Class in org.apache.beam.sdk.extensions.jackson
PTransform for serializing objects to JSON Strings.
AsJsons.AsJsonsWithFailures<FailureT> - Class in org.apache.beam.sdk.extensions.jackson
A PTransform that adds exception handling to AsJsons.
asList() - Static method in class org.apache.beam.sdk.transforms.View
Returns a View.AsList transform that takes a PCollection and returns a PCollectionView mapping each window to a List containing all of the elements in the window.
asMap() - Method in class org.apache.beam.sdk.transforms.display.DisplayData
 
asMap() - Static method in class org.apache.beam.sdk.transforms.View
Returns a View.AsMap transform that takes a PCollection<KV<K, V>> as input and produces a PCollectionView mapping each window to a Map<K, V>.
asMultimap() - Static method in class org.apache.beam.sdk.transforms.View
Returns a View.AsMultimap transform that takes a PCollection<KV<K, V>> as input and produces a PCollectionView mapping each window to its contents as a Map<K, Iterable<V>> for use as a side input.
asOutputReference(PValue, AppliedPTransform<?, ?, ?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
Encode a PValue reference as an output reference.
asOutputStream() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
Returns an output stream which writes to the backing buffer from the current position.
asPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription
Returns the string representation of this subscription as a path used in the Cloud Pub/Sub API.
asPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic
Returns the string representation of this topic as a path used in the Cloud Pub/Sub API.
asQueryable(QueryProvider, SchemaPlus, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteTable
 
asResponseObserver() - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClient
 
ASSERTION_ERROR - Static variable in class org.apache.beam.runners.apex.ApexRunner
TODO: this isn't thread safe and may cause issues when tests run in parallel Holds any most resent assertion error that was raised while processing elements.
assertionError() - Method in class org.apache.beam.sdk.testing.SuccessOrFailure
 
assertSourcesEqualReferenceSource(BoundedSource<T>, List<? extends BoundedSource<T>>, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
Given a reference Source and a list of Sources, assert that the union of the records read from the list of sources is equal to the records read from the reference source.
assertSplitAtFractionBehavior(BoundedSource<T>, int, double, SourceTestUtils.ExpectedSplitOutcome, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
Asserts that the source's reader either fails to splitAtFraction(fraction) after reading numItemsToReadBeforeSplit items, or succeeds in a way that is consistent according to SourceTestUtils.assertSplitAtFractionSucceedsAndConsistent(org.apache.beam.sdk.io.BoundedSource<T>, int, double, org.apache.beam.sdk.options.PipelineOptions).
assertSplitAtFractionExhaustive(BoundedSource<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
Asserts that for each possible start position, BoundedSource.BoundedReader#splitAtFraction at every interesting fraction (halfway between two fractions that differ by at least one item) can be called successfully and the results are consistent if a split succeeds.
assertSplitAtFractionFails(BoundedSource<T>, int, double, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
Asserts that the source's reader fails to splitAtFraction(fraction) after reading numItemsToReadBeforeSplit items.
assertSplitAtFractionSucceedsAndConsistent(BoundedSource<T>, int, double, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
Verifies some consistency properties of BoundedSource.BoundedReader#splitAtFraction on the given source.
assertThatAllRows(Schema) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery
 
assertThatTopicEventuallyReceives(Matcher<PubsubMessage>...) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
Repeatedly pull messages from TestPubsub.subscriptionPath() until receiving one for each matcher (or timeout is reached), then assert that the received messages match the expectations.
assertUnstartedReaderReadsSameAsItsSource(BoundedSource.BoundedReader<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
Assert that a Reader returns a Source that, when read from, produces the same records as the reader.
assign(BoundedWindow, Instant) - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
assignableTo(Schema) - Method in class org.apache.beam.sdk.schemas.Schema
Returns true if this Schema can be assigned to another Schema.
assignableToIgnoreNullable(Schema) - Method in class org.apache.beam.sdk.schemas.Schema
Returns true if this Schema can be assigned to another Schema, ignoring nullable.
AssignContext() - Constructor for class org.apache.beam.sdk.transforms.windowing.WindowFn.AssignContext
 
assignedWindows(WindowFn<T, W>, long) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
assignedWindowsWithValue(WindowFn<T, W>, TimestampedValue<T>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
assignShardKey(DestinationT, UserT, int) - Method in interface org.apache.beam.sdk.io.ShardingFunction
 
assignsToOneWindow() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
 
assignsToOneWindow() - Method in class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
 
assignsToOneWindow() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
 
assignsToOneWindow() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
Returns true if this WindowFn always assigns an element to exactly one window.
assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
 
assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
 
assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
 
assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
 
assignWindow(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
Returns the single window to which elements with this timestamp belong.
AssignWindowP<T> - Class in org.apache.beam.runners.jet.processors
/** * Jet Processor implementation for Beam's Windowing primitive.
assignWindows(WindowFn<Object, GlobalWindow>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
 
assignWindows(WindowFn<Object, W>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
 
assignWindows(WindowFn<T, W>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
 
assignWindows(WindowFn<Object, IntervalWindow>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
 
assignWindows(WindowFn<Object, IntervalWindow>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
 
assignWindows(Instant) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
 
assignWindows(WindowFn<T, W>.AssignContext) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
Given a timestamp and element, returns the set of windows into which it should be placed.
assignWindowsMapFunction(WindowFn<T, W>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.WindowingHelpers
 
asSingleton() - Static method in class org.apache.beam.sdk.transforms.View
Returns a View.AsSingleton transform that takes a PCollection with a single value per window as input and produces a PCollectionView that returns the value in the main input window when read as a side input.
asSingletonView() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
Returns a PTransform that produces a PCollectionView whose elements are the result of combining elements per-window in the input PCollection.
asV1Beta1Path() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription
Deprecated.
the v1beta1 API for Cloud Pub/Sub is deprecated.
asV1Beta1Path() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic
Deprecated.
the v1beta1 API for Cloud Pub/Sub is deprecated.
asV1Beta2Path() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription
Deprecated.
the v1beta2 API for Cloud Pub/Sub is deprecated.
asV1Beta2Path() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic
Deprecated.
the v1beta2 API for Cloud Pub/Sub is deprecated.
atMinimumTimestamp(V) - Static method in class org.apache.beam.sdk.values.TimestampedValue
Returns a new TimestampedValue with the minimum timestamp.
AtomicCoder<T> - Class in org.apache.beam.sdk.coders
A Coder that has no component Coders or other configuration.
AtomicCoder() - Constructor for class org.apache.beam.sdk.coders.AtomicCoder
 
AtomicLongFactory() - Constructor for class org.apache.beam.sdk.options.PipelineOptions.AtomicLongFactory
 
attached() - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
 
attachValues(List<Object>) - Method in class org.apache.beam.sdk.values.Row.Builder
 
attachValues(Object...) - Method in class org.apache.beam.sdk.values.Row.Builder
 
attempted(MetricKey, T) - Static method in class org.apache.beam.sdk.metrics.MetricResult
 
AttributeValueCoder - Class in org.apache.beam.sdk.io.aws.dynamodb
A Coder that serializes and deserializes the AttributeValue objects.
AttributeValueCoder - Class in org.apache.beam.sdk.io.aws2.dynamodb
A Coder that serializes and deserializes the AttributeValue objects.
AttributeValueCoderProviderRegistrar - Class in org.apache.beam.sdk.io.aws.dynamodb
A CoderProviderRegistrar for standard types used with DynamoDBIO.
AttributeValueCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.aws.dynamodb.AttributeValueCoderProviderRegistrar
 
AUTO - Static variable in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
autoCastField(Schema.Field, Object) - Static method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamTableUtils
Attempt to cast an object to a specified Schema.Field.Type.
autoLoadBuiltinFunctions() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
Load Beam SQL built-in functions defined in BeamBuiltinFunctionProvider.
autoLoadUserDefinedFunctions() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
Load UDF/UDAFs from UdfUdafProvider.
AutoValueSchema - Class in org.apache.beam.sdk.schemas
A SchemaProvider for AutoValue classes.
AutoValueSchema() - Constructor for class org.apache.beam.sdk.schemas.AutoValueSchema
 
AutoValueSchema.AbstractGetterTypeSupplier - Class in org.apache.beam.sdk.schemas
FieldValueTypeSupplier that's based on AutoValue getters.
AutoValueUtils - Class in org.apache.beam.sdk.schemas.utils
Utilities for managing AutoValue schemas.
AutoValueUtils() - Constructor for class org.apache.beam.sdk.schemas.utils.AutoValueUtils
 
AvailableParallelismFactory() - Constructor for class org.apache.beam.runners.direct.DirectOptions.AvailableParallelismFactory
 
AvroCoder<T> - Class in org.apache.beam.sdk.coders
A Coder using Avro binary format.
AvroCoder(Class<T>, Schema) - Constructor for class org.apache.beam.sdk.coders.AvroCoder
 
AvroCoder.JodaTimestampConversion - Class in org.apache.beam.sdk.coders
Conversion for DateTime.
AvroConvertType(boolean) - Constructor for class org.apache.beam.sdk.schemas.utils.AvroUtils.AvroConvertType
 
AvroGenericCoder - Class in org.apache.beam.sdk.coders
AvroCoder specialisation for GenericRecord.
AvroIO - Class in org.apache.beam.sdk.io
PTransforms for reading and writing Avro files.
AvroIO.Parse<T> - Class in org.apache.beam.sdk.io
AvroIO.ParseAll<T> - Class in org.apache.beam.sdk.io
AvroIO.ParseFiles<T> - Class in org.apache.beam.sdk.io
AvroIO.Read<T> - Class in org.apache.beam.sdk.io
AvroIO.ReadAll<T> - Class in org.apache.beam.sdk.io
Deprecated.
See AvroIO.readAll(Class) for details.
AvroIO.ReadFiles<T> - Class in org.apache.beam.sdk.io
AvroIO.RecordFormatter<ElementT> - Interface in org.apache.beam.sdk.io
Deprecated.
Users can achieve the same by providing this transform in a ParDo before using write in AvroIO AvroIO.write(Class).
AvroIO.Sink<ElementT> - Class in org.apache.beam.sdk.io
AvroIO.TypedWrite<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io
AvroIO.Write<T> - Class in org.apache.beam.sdk.io
This class is used as the default return value of AvroIO.write(java.lang.Class<T>)
AvroReader(AvroSource<T>) - Constructor for class org.apache.beam.sdk.io.AvroSource.AvroReader
Reads Avro records of type T from the specified source.
AvroRecordSchema - Class in org.apache.beam.sdk.schemas
A SchemaProvider for AVRO generated SpecificRecords and POJOs.
AvroRecordSchema() - Constructor for class org.apache.beam.sdk.schemas.AvroRecordSchema
 
AvroSource<T> - Class in org.apache.beam.sdk.io
Do not use in pipelines directly: most users should use AvroIO.Read.
AvroSource.AvroReader<T> - Class in org.apache.beam.sdk.io
A BlockBasedSource.BlockBasedReader for reading blocks from Avro files.
AvroTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.avro
AvroTable(String, Schema, String) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.avro.AvroTable
 
AvroTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.avro
AvroTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.avro.AvroTableProvider
 
AvroUtils - Class in org.apache.beam.sdk.schemas.utils
Utils to convert AVRO records to Beam rows.
AvroUtils.AvroConvertType - Class in org.apache.beam.sdk.schemas.utils
 
AvroUtils.AvroConvertValueForGetter - Class in org.apache.beam.sdk.schemas.utils
 
AvroUtils.AvroConvertValueForSetter - Class in org.apache.beam.sdk.schemas.utils
 
AvroUtils.FixedBytesField - Class in org.apache.beam.sdk.schemas.utils
Wrapper for fixed byte fields.
AvroWriteRequest<T> - Class in org.apache.beam.sdk.io.gcp.bigquery
 
AvroWriteRequest(T, Schema) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.AvroWriteRequest
 
awaitCompletion() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
 
awaitCompletion() - Method in class org.apache.beam.sdk.fn.data.CompletableFutureInboundDataClient
 
awaitCompletion() - Method in interface org.apache.beam.sdk.fn.data.InboundDataClient
Block until the client has completed reading from the inbound stream.
awaitTermination(Duration) - Method in class org.apache.beam.runners.spark.SparkPipelineResult
 
awaitTermination(Duration) - Method in class org.apache.beam.runners.spark.SparkRunnerDebugger.DebugSparkPipelineResult
 
AwsClientsProvider - Interface in org.apache.beam.sdk.io.aws.dynamodb
Provides instances of AWS clients.
AwsClientsProvider - Interface in org.apache.beam.sdk.io.aws.sns
Provides instances of AWS clients.
AWSClientsProvider - Interface in org.apache.beam.sdk.io.kinesis
Provides instances of AWS clients.
AwsCoders - Class in org.apache.beam.sdk.io.aws.coders
Coders for common AWS SDK objects.
AwsModule - Class in org.apache.beam.sdk.io.aws.options
A Jackson Module that registers a JsonSerializer and JsonDeserializer for AWSCredentialsProvider and some subclasses.
AwsModule() - Constructor for class org.apache.beam.sdk.io.aws.options.AwsModule
 
AwsModule - Class in org.apache.beam.sdk.io.aws2.options
A Jackson Module that registers a JsonSerializer and JsonDeserializer for AwsCredentialsProvider and some subclasses.
AwsModule() - Constructor for class org.apache.beam.sdk.io.aws2.options.AwsModule
 
AwsOptions - Interface in org.apache.beam.sdk.io.aws.options
Options used to configure Amazon Web Services specific options such as credentials and region.
AwsOptions - Interface in org.apache.beam.sdk.io.aws2.options
Options used to configure Amazon Web Services specific options such as credentials and region.
AwsOptions.AwsUserCredentialsFactory - Class in org.apache.beam.sdk.io.aws.options
Attempts to load AWS credentials.
AwsOptions.AwsUserCredentialsFactory - Class in org.apache.beam.sdk.io.aws2.options
Attempts to load AWS credentials.
AwsOptions.ClientConfigurationFactory - Class in org.apache.beam.sdk.io.aws.options
Default AWS client configuration.
AwsPipelineOptionsRegistrar - Class in org.apache.beam.sdk.io.aws.options
A registrar containing the default AWS options.
AwsPipelineOptionsRegistrar() - Constructor for class org.apache.beam.sdk.io.aws.options.AwsPipelineOptionsRegistrar
 
AwsPipelineOptionsRegistrar - Class in org.apache.beam.sdk.io.aws2.options
A registrar containing the default AWS options.
AwsPipelineOptionsRegistrar() - Constructor for class org.apache.beam.sdk.io.aws2.options.AwsPipelineOptionsRegistrar
 
AwsUserCredentialsFactory() - Constructor for class org.apache.beam.sdk.io.aws.options.AwsOptions.AwsUserCredentialsFactory
 
AwsUserCredentialsFactory() - Constructor for class org.apache.beam.sdk.io.aws2.options.AwsOptions.AwsUserCredentialsFactory
 

B

BACKLOG_UNKNOWN - Static variable in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
Constant representing an unknown amount of backlog.
backlogBytes() - Static method in class org.apache.beam.sdk.metrics.SourceMetrics
Gauge for source backlog in bytes.
backlogBytesOfSplit(String) - Static method in class org.apache.beam.sdk.metrics.SourceMetrics
Gauge for source split backlog in bytes.
backlogElements() - Static method in class org.apache.beam.sdk.metrics.SourceMetrics
Gauge for source backlog in elements.
backlogElementsOfSplit(String) - Static method in class org.apache.beam.sdk.metrics.SourceMetrics
Gauge for source split backlog in elements.
BackOffAdapter - Class in org.apache.beam.sdk.extensions.gcp.util
An adapter for converting between Apache Beam and Google API client representations of backoffs.
BackOffAdapter() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.BackOffAdapter
 
bag() - Static method in class org.apache.beam.sdk.state.StateSpecs
Create a StateSpec for a BagState, optimized for adding values frequently and occasionally retrieving all the values that have been added.
bag(Coder<T>) - Static method in class org.apache.beam.sdk.state.StateSpecs
Identical to StateSpecs.bag(), but with an element coder explicitly supplied.
BagState<T> - Interface in org.apache.beam.sdk.state
A ReadableState cell containing a bag of values.
BagUserStateSpec() - Constructor for class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.BagUserStateSpec
 
BaseBeamTable - Class in org.apache.beam.sdk.extensions.sql.meta
Basic implementation of BeamSqlTable methods used by predicate and filter push-down.
BaseBeamTable() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.BaseBeamTable
 
BasicDynamoDbClientProvider - Class in org.apache.beam.sdk.io.aws2.dynamodb
Basic implementation of DynamoDbClientProvider used by default in DynamoDBIO.
BasicDynamoDBProvider - Class in org.apache.beam.sdk.io.aws.dynamodb
Basic implementation of AwsClientsProvider used by default in DynamoDBIO.
BatchSideInputHandlerFactory - Class in org.apache.beam.runners.fnexecution.translation
BatchSideInputHandlerFactory.SideInputGetter - Interface in org.apache.beam.runners.fnexecution.translation
Returns the value for the side input with the given PCollection id from the runner.
BatchStatefulParDoOverrides - Class in org.apache.beam.runners.dataflow
PTransformOverrideFactories that expands to correctly implement stateful ParDo using window-unaware BatchViewOverrides.GroupByKeyAndSortValuesOnly to linearize processing per key.
BatchStatefulParDoOverrides() - Constructor for class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides
 
BatchStatefulParDoOverrides.BatchStatefulDoFn<K,V,OutputT> - Class in org.apache.beam.runners.dataflow
A key-preserving DoFn that explodes an iterable that has been grouped by key and window.
BEAM_FN_API_DATA_BUFFER_LIMIT - Static variable in interface org.apache.beam.sdk.fn.data.BeamFnDataBufferingOutboundObserver
Deprecated.
Use DATA_BUFFER_SIZE_LIMIT instead.
BEAM_FN_API_DATA_BUFFER_SIZE_LIMIT - Static variable in interface org.apache.beam.sdk.fn.data.BeamFnDataBufferingOutboundObserver
Deprecated.
Use DATA_BUFFER_SIZE_LIMIT instead.
BEAM_FN_API_DATA_BUFFER_TIME_LIMIT - Static variable in interface org.apache.beam.sdk.fn.data.BeamFnDataBufferingOutboundObserver
Deprecated.
Use DATA_BUFFER_TIME_LIMIT_MS instead.
BeamAggregateProjectMergeRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
This rule is essentially a wrapper around Calcite's AggregateProjectMergeRule.
BeamAggregateProjectMergeRule(Class<? extends Aggregate>, Class<? extends Project>, RelBuilderFactory) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rule.BeamAggregateProjectMergeRule
 
BeamAggregationRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
BeamRelNode to replace a Aggregate node.
BeamAggregationRel(RelOptCluster, RelTraitSet, RelNode, ImmutableBitSet, List<ImmutableBitSet>, List<AggregateCall>, WindowFn<Row, IntervalWindow>, int) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamAggregationRel
 
BeamAggregationRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
Rule to detect the window/trigger settings.
BeamAggregationRule(Class<? extends Aggregate>, Class<? extends Project>, RelBuilderFactory) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rule.BeamAggregationRule
 
BeamBasicAggregationRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
Aggregation rule that doesn't include projection.
BeamBasicAggregationRule(Class<? extends Aggregate>, RelBuilderFactory) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rule.BeamBasicAggregationRule
 
BeamBigQuerySqlDialect - Class in org.apache.beam.sdk.extensions.sql.meta.provider.bigquery
 
BeamBigQuerySqlDialect(SqlDialect.Context) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamBigQuerySqlDialect
 
BeamBuiltinAggregations - Class in org.apache.beam.sdk.extensions.sql.impl.transform
Built-in aggregations functions for COUNT/MAX/MIN/SUM/AVG/VAR_POP/VAR_SAMP.
BeamBuiltinAggregations() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamBuiltinAggregations
 
BeamBuiltinFunctionProvider - Class in org.apache.beam.sdk.extensions.sql.impl.udf
BeamBuiltinFunctionClass interface.
BeamBuiltinFunctionProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.udf.BeamBuiltinFunctionProvider
 
BeamBuiltinMethods - Class in org.apache.beam.sdk.extensions.sql.zetasql
BeamBuiltinMethods.
BeamBuiltinMethods() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.BeamBuiltinMethods
 
BeamCalciteSchema - Class in org.apache.beam.sdk.extensions.sql.impl
Adapter from TableProvider to Schema.
BeamCalciteTable - Class in org.apache.beam.sdk.extensions.sql.impl
Adapter from BeamSqlTable to a calcite Table.
BeamCalcRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
BeamRelNode to replace Project and Filter node.
BeamCalcRel(RelOptCluster, RelTraitSet, RelNode, RexProgram) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel
 
BeamCalcRel.WrappedList<T> - Class in org.apache.beam.sdk.extensions.sql.impl.rel
WrappedList translates List on access.
BeamCalcRel.WrappedMap<V> - Class in org.apache.beam.sdk.extensions.sql.impl.rel
WrappedMap translates Map on access.
BeamCalcRel.WrappedRow - Class in org.apache.beam.sdk.extensions.sql.impl.rel
WrappedRow translates Row on access.
BeamCalcRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
A ConverterRule to replace Calc with BeamCalcRel.
BeamCodegenUtils - Class in org.apache.beam.sdk.extensions.sql.zetasql
BeamCodegenUtils.
BeamCodegenUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.BeamCodegenUtils
 
BeamCoGBKJoinRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
A BeamJoinRel which does CoGBK Join
BeamCoGBKJoinRel(RelOptCluster, RelTraitSet, RelNode, RelNode, RexNode, Set<CorrelationId>, JoinRelType) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCoGBKJoinRel
 
BeamCoGBKJoinRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
Rule to convert LogicalJoin node to BeamCoGBKJoinRel node.
beamComputeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.AbstractBeamCalcRel
 
beamComputeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamAggregationRel
 
beamComputeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIntersectRel
 
beamComputeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
 
beamComputeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
 
beamComputeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
 
beamComputeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamMinusRel
 
beamComputeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamPushDownIOSourceRel
 
beamComputeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in interface org.apache.beam.sdk.extensions.sql.impl.rel.BeamRelNode
This method is called by org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner.NonCumulativeCostImpl.
beamComputeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSortRel
 
beamComputeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUncollectRel
 
beamComputeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnionRel
 
beamComputeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnnestRel
 
beamComputeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamValuesRel
 
BeamCostModel - Class in org.apache.beam.sdk.extensions.sql.impl.planner
VolcanoCost represents the cost of a plan node.
BeamCostModel.Factory - Class in org.apache.beam.sdk.extensions.sql.impl.planner
Implementation of RelOptCostFactory that creates BeamCostModels.
BeamEnumerableConverter - Class in org.apache.beam.sdk.extensions.sql.impl.rel
BeamRelNode to replace a Enumerable node.
BeamEnumerableConverter(RelOptCluster, RelTraitSet, RelNode) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter
 
BeamEnumerableConverterRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
A ConverterRule to Convert BeamRelNode to EnumerableConvention.
beamFieldTypeToZetaSqlType(Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlUtils
 
BeamFileSystemLegacyArtifactRetrievalService - Class in org.apache.beam.runners.fnexecution.artifact
An LegacyArtifactRetrievalService that uses FileSystems as its backing storage and uses the artifact layout and retrieval token format produced by BeamFileSystemLegacyArtifactStagingService.
BeamFileSystemLegacyArtifactRetrievalService() - Constructor for class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactRetrievalService
 
BeamFileSystemLegacyArtifactStagingService - Class in org.apache.beam.runners.fnexecution.artifact
This implementation is experimental.
BeamFileSystemLegacyArtifactStagingService() - Constructor for class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService
 
BeamFileSystemLegacyArtifactStagingService.StagingSessionToken - Class in org.apache.beam.runners.fnexecution.artifact
Serializable StagingSessionToken used to stage files with BeamFileSystemLegacyArtifactStagingService.
BeamFnDataBufferingOutboundObserver<T> - Interface in org.apache.beam.sdk.fn.data
A buffering outbound FnDataReceiver for the Beam Fn Data API.
BeamFnDataGrpcMultiplexer - Class in org.apache.beam.sdk.fn.data
A gRPC multiplexer for a specific Endpoints.ApiServiceDescriptor.
BeamFnDataGrpcMultiplexer(Endpoints.ApiServiceDescriptor, OutboundObserverFactory, OutboundObserverFactory.BasicFactory<BeamFnApi.Elements, BeamFnApi.Elements>) - Constructor for class org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer
 
BeamFnDataInboundObserver<T> - Class in org.apache.beam.sdk.fn.data
Decodes individually consumed ByteStrings with the provided Coder passing the individual decoded elements to the provided consumer.
BeamFnDataInboundObserver(LogicalEndpoint, Coder<T>, FnDataReceiver<T>, InboundDataClient) - Constructor for class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
 
BeamFnDataSizeBasedBufferingOutboundObserver<T> - Class in org.apache.beam.sdk.fn.data
A size-based buffering outbound FnDataReceiver for the Beam Fn Data API.
BeamFnDataTimeBasedBufferingOutboundObserver<T> - Class in org.apache.beam.sdk.fn.data
A buffering outbound FnDataReceiver with both size-based buffer and time-based buffer enabled for the Beam Fn Data API.
BeamIntersectRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
BeamRelNode to replace a Intersect node.
BeamIntersectRel(RelOptCluster, RelTraitSet, List<RelNode>, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIntersectRel
 
BeamIntersectRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
ConverterRule to replace Intersect with BeamIntersectRel.
BeamIOPushDownRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
 
BeamIOPushDownRule(RelBuilderFactory) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIOPushDownRule
 
BeamIOSinkRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
BeamRelNode to replace a TableModify node.
BeamIOSinkRel(RelOptCluster, RelOptTable, Prepare.CatalogReader, RelNode, TableModify.Operation, List<String>, List<RexNode>, boolean, BeamSqlTable, Map<String, String>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
 
BeamIOSinkRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
A ConverterRule to replace TableModify with BeamIOSinkRel.
BeamIOSourceRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
BeamRelNode to replace a TableScan node.
BeamIOSourceRel(RelOptCluster, RelTraitSet, RelOptTable, BeamSqlTable, Map<String, String>, BeamCalciteTable) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
 
BeamJavaTypeFactory - Class in org.apache.beam.sdk.extensions.sql.impl.planner
customized data type in Beam.
BeamJoinAssociateRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
This is very similar to JoinAssociateRule.
BeamJoinPushThroughJoinRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
This is exactly similar to JoinPushThroughJoinRule.
BeamJoinRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
An abstract BeamRelNode to implement Join Rels.
BeamJoinRel(RelOptCluster, RelTraitSet, RelNode, RelNode, RexNode, Set<CorrelationId>, JoinRelType) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
 
BeamJoinTransforms - Class in org.apache.beam.sdk.extensions.sql.impl.transform
Collections of PTransform and DoFn used to perform JOIN operation.
BeamJoinTransforms() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamJoinTransforms
 
BeamJoinTransforms.JoinAsLookup - Class in org.apache.beam.sdk.extensions.sql.impl.transform
Transform to execute Join as Lookup.
BeamKafkaCSVTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.kafka
A Kafka topic that saves records as CSV format.
BeamKafkaCSVTable(Schema, String, List<String>) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaCSVTable
 
BeamKafkaCSVTable(Schema, String, List<String>, CSVFormat) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaCSVTable
 
BeamKafkaCSVTable.CsvRecorderDecoder - Class in org.apache.beam.sdk.extensions.sql.meta.provider.kafka
A PTransform to convert KV<byte[], byte[]> to Row.
BeamKafkaCSVTable.CsvRecorderEncoder - Class in org.apache.beam.sdk.extensions.sql.meta.provider.kafka
A PTransform to convert Row to KV<byte[], byte[]>.
BeamKafkaTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.kafka
BeamKafkaTable represent a Kafka topic, as source or target.
BeamKafkaTable(Schema) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
 
BeamKafkaTable(Schema, String, List<String>) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
 
BeamKafkaTable(Schema, List<TopicPartition>, String) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
 
BeamLogicalConvention - Enum in org.apache.beam.sdk.extensions.sql.impl.rel
Convertion for Beam SQL.
BeamMinusRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
BeamRelNode to replace a Minus node.
BeamMinusRel(RelOptCluster, RelTraitSet, List<RelNode>, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamMinusRel
 
BeamMinusRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
ConverterRule to replace Minus with BeamMinusRel.
BeamPCollectionTable<InputT> - Class in org.apache.beam.sdk.extensions.sql.impl.schema
BeamPCollectionTable converts a PCollection<Row> as a virtual table, then a downstream query can query directly.
BeamPCollectionTable(PCollection<InputT>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.BeamPCollectionTable
 
BeamPushDownIOSourceRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
 
BeamPushDownIOSourceRel(RelOptCluster, RelTraitSet, RelOptTable, BeamSqlTable, List<String>, BeamSqlTableFilter, Map<String, String>, BeamCalciteTable) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamPushDownIOSourceRel
 
BeamRelDataTypeSystem - Class in org.apache.beam.sdk.extensions.sql.impl.planner
customized data type in Beam.
BeamRelNode - Interface in org.apache.beam.sdk.extensions.sql.impl.rel
A RelNode that can also give a PTransform that implements the expression.
beamRow2CsvLine(Row, CSVFormat) - Static method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamTableUtils
 
beamRowToZetaSqlStructValue(Row, Schema) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlUtils
 
BeamRuleSets - Class in org.apache.beam.sdk.extensions.sql.impl.planner
RuleSet used in BeamQueryPlanner.
BeamRuleSets() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRuleSets
 
beamSchema() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.avro.GenericRecordWriteConverter
 
beamSchema() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.parquet.GenericRecordReadConverter
 
BeamSetOperatorRelBase - Class in org.apache.beam.sdk.extensions.sql.impl.rel
Delegate for Set operators: BeamUnionRel, BeamIntersectRel and BeamMinusRel.
BeamSetOperatorRelBase(BeamRelNode, BeamSetOperatorRelBase.OpType, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSetOperatorRelBase
 
BeamSetOperatorRelBase.OpType - Enum in org.apache.beam.sdk.extensions.sql.impl.rel
Set operator type.
BeamSetOperatorsTransforms - Class in org.apache.beam.sdk.extensions.sql.impl.transform
Collections of PTransform and DoFn used to perform Set operations.
BeamSetOperatorsTransforms() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSetOperatorsTransforms
 
BeamSetOperatorsTransforms.BeamSqlRow2KvFn - Class in org.apache.beam.sdk.extensions.sql.impl.transform
Transform a BeamSqlRow to a KV<BeamSqlRow, BeamSqlRow>.
BeamSetOperatorsTransforms.SetOperatorFilteringDoFn - Class in org.apache.beam.sdk.extensions.sql.impl.transform
Filter function used for Set operators.
BeamSideInputJoinRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
A BeamJoinRel which does sideinput Join
BeamSideInputJoinRel(RelOptCluster, RelTraitSet, RelNode, RelNode, RexNode, Set<CorrelationId>, JoinRelType) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSideInputJoinRel
 
BeamSideInputJoinRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
Rule to convert LogicalJoin node to BeamSideInputJoinRel node.
BeamSideInputLookupJoinRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
A BeamJoinRel which does Lookup Join
BeamSideInputLookupJoinRel(RelOptCluster, RelTraitSet, RelNode, RelNode, RexNode, Set<CorrelationId>, JoinRelType) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSideInputLookupJoinRel
 
BeamSideInputLookupJoinRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
Rule to convert LogicalJoin node to BeamSideInputLookupJoinRel node.
BeamSideInputLookupJoinRule() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSideInputLookupJoinRule
 
BeamSortRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
BeamRelNode to replace a Sort node.
BeamSortRel(RelOptCluster, RelTraitSet, RelNode, RelCollation, RexNode, RexNode) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSortRel
 
BeamSortRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
ConverterRule to replace Sort with BeamSortRel.
BeamSqlCli - Class in org.apache.beam.sdk.extensions.sql
BeamSqlCli provides methods to execute Beam SQL with an interactive client.
BeamSqlCli() - Constructor for class org.apache.beam.sdk.extensions.sql.BeamSqlCli
 
BeamSqlDataCatalogExample - Class in org.apache.beam.sdk.extensions.sql.example
Example pipeline that uses Google Cloud Data Catalog to retrieve the table metadata.
BeamSqlDataCatalogExample() - Constructor for class org.apache.beam.sdk.extensions.sql.example.BeamSqlDataCatalogExample
 
BeamSqlDataCatalogExample.DCExamplePipelineOptions - Interface in org.apache.beam.sdk.extensions.sql.example
Pipeline options to specify the query and the output for the example.
BeamSqlEnv - Class in org.apache.beam.sdk.extensions.sql.impl
Contains the metadata of tables/UDF functions, and exposes APIs to query/validate/optimize/translate SQL statements.
BeamSqlEnv.BeamSqlEnvBuilder - Class in org.apache.beam.sdk.extensions.sql.impl
BeamSqlEnv's Builder.
BeamSqlOutputToConsoleFn - Class in org.apache.beam.sdk.extensions.sql.impl.transform
A test PTransform to display output in console.
BeamSqlOutputToConsoleFn(String) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSqlOutputToConsoleFn
 
BeamSqlPipelineOptions - Interface in org.apache.beam.sdk.extensions.sql.impl
Options used to configure BeamSQL.
BeamSqlPipelineOptionsRegistrar - Class in org.apache.beam.sdk.extensions.sql.impl
AutoService registrar for BeamSqlPipelineOptions.
BeamSqlPipelineOptionsRegistrar() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.BeamSqlPipelineOptionsRegistrar
 
BeamSqlRelUtils - Class in org.apache.beam.sdk.extensions.sql.impl.rel
Utilities for BeamRelNode.
BeamSqlRelUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils
 
BeamSqlRow2KvFn() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSetOperatorsTransforms.BeamSqlRow2KvFn
 
BeamSqlSeekableTable - Interface in org.apache.beam.sdk.extensions.sql
A seekable table converts a JOIN operator to an inline lookup.
BeamSqlTable - Interface in org.apache.beam.sdk.extensions.sql.meta
This interface defines a Beam Sql Table.
BeamSqlTableFilter - Interface in org.apache.beam.sdk.extensions.sql.meta
This interface defines Beam SQL Table Filter.
BeamSqlUdf - Interface in org.apache.beam.sdk.extensions.sql
Interface to create a UDF in Beam SQL.
BeamSqlUnparseContext - Class in org.apache.beam.sdk.extensions.sql.meta.provider.bigquery
 
BeamSqlUnparseContext(IntFunction<SqlNode>) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamSqlUnparseContext
 
BeamTableStatistics - Class in org.apache.beam.sdk.extensions.sql.impl
This class stores row count statistics.
BeamTableUtils - Class in org.apache.beam.sdk.extensions.sql.impl.schema
Utility methods for working with BeamTable.
BeamTableUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.schema.BeamTableUtils
 
BeamUncollectRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
BeamRelNode to implement an uncorrelated Uncollect, aka UNNEST.
BeamUncollectRel(RelOptCluster, RelTraitSet, RelNode, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUncollectRel
 
BeamUncollectRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
A ConverterRule to replace Uncollect with BeamUncollectRule.
BeamUnionRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
BeamRelNode to replace a Union.
BeamUnionRel(RelOptCluster, RelTraitSet, List<RelNode>, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnionRel
 
BeamUnionRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
A ConverterRule to replace Union with BeamUnionRule.
BeamUnnestRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
BeamRelNode to implement UNNEST, supporting specifically only Correlate with Uncollect.
BeamUnnestRel(RelOptCluster, RelTraitSet, RelNode, RelDataType, int) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnnestRel
 
BeamUnnestRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
A ConverterRule to replace Correlate Uncollect with BeamUnnestRule.
BeamValuesRel - Class in org.apache.beam.sdk.extensions.sql.impl.rel
BeamRelNode to replace a Values node.
BeamValuesRel(RelOptCluster, RelDataType, ImmutableList<ImmutableList<RexLiteral>>, RelTraitSet) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamValuesRel
 
BeamValuesRule - Class in org.apache.beam.sdk.extensions.sql.impl.rule
ConverterRule to replace Values with BeamValuesRel.
BeamWorkerStatusGrpcService - Class in org.apache.beam.runners.fnexecution.status
A Fn Status service which can collect run-time status information from SDK harnesses for debugging purpose.
BeamZetaSqlCalcRel - Class in org.apache.beam.sdk.extensions.sql.zetasql
BeamRelNode to replace Project and Filter node based on the ZetaSQL expression evaluator.
BeamZetaSqlCalcRel(RelOptCluster, RelTraitSet, RelNode, RexProgram) - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcRel
 
BeamZetaSqlCalcRule - Class in org.apache.beam.sdk.extensions.sql.zetasql
A ConverterRule to replace Calc with BeamZetaSqlCalcRel.
beforeProcessing(PipelineOptions) - Method in interface org.apache.beam.sdk.harness.JvmInitializer
Implement beforeProcessing to run some custom initialization after basic services such as logging, but before data processing begins.
beforeStart(ClientCallStreamObserver<RespT>) - Method in class org.apache.beam.sdk.fn.stream.ForwardingClientResponseObserver
 
begin() - Method in class org.apache.beam.sdk.Pipeline
Returns a PBegin owned by this Pipeline.
beginningOnDay(int) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
 
beginningOnDay(int, int) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
 
BIG_INT - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
BIG_QUERY_INSERT_ERROR_ERROR_CONTAINER - Static variable in interface org.apache.beam.sdk.io.gcp.bigquery.ErrorContainer
 
BigDecimalCoder - Class in org.apache.beam.sdk.coders
A BigDecimalCoder encodes a BigDecimal as an integer scale encoded with VarIntCoder and a BigInteger encoded using BigIntegerCoder.
BigDecimalConverter - Class in org.apache.beam.sdk.extensions.sql.impl.utils
Provides converters from BigDecimal to other numeric types based on the input Schema.TypeName.
BigDecimalConverter() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.BigDecimalConverter
 
bigdecimals() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for BigDecimal.
BigEndianIntegerCoder - Class in org.apache.beam.sdk.coders
A BigEndianIntegerCoder encodes Integers in 4 bytes, big-endian.
BigEndianLongCoder - Class in org.apache.beam.sdk.coders
A BigEndianLongCoder encodes Longs in 8 bytes, big-endian.
BigEndianShortCoder - Class in org.apache.beam.sdk.coders
A BigEndianShortCoder encodes Shorts in 2 bytes, big-endian.
BIGINT - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
BigIntegerCoder - Class in org.apache.beam.sdk.coders
A BigIntegerCoder encodes a BigInteger as a byte array containing the big endian two's-complement representation, encoded via ByteArrayCoder.
bigintegers() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for BigInteger.
BigqueryClient - Class in org.apache.beam.sdk.io.gcp.testing
A wrapper class to call Bigquery API calls.
BigqueryClient(String) - Constructor for class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
 
BigQueryCoderProviderRegistrar - Class in org.apache.beam.sdk.io.gcp.bigquery
A CoderProviderRegistrar for standard types used with BigQueryIO.
BigQueryCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryCoderProviderRegistrar
 
BigQueryFilter - Class in org.apache.beam.sdk.extensions.sql.meta.provider.bigquery
 
BigQueryFilter(List<RexNode>) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryFilter
 
BigQueryHelpers - Class in org.apache.beam.sdk.io.gcp.bigquery
A set of helper functions and classes used by BigQueryIO.
BigQueryHelpers() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
 
BigQueryInsertError - Class in org.apache.beam.sdk.io.gcp.bigquery
Model definition for BigQueryInsertError.
BigQueryInsertError(TableRow, TableDataInsertAllResponse.InsertErrors, TableReference) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertError
 
BigQueryInsertErrorCoder - Class in org.apache.beam.sdk.io.gcp.bigquery
A Coder that encodes BigQuery BigQueryInsertError objects.
BigQueryInsertErrorCoder() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertErrorCoder
 
BigQueryIO - Class in org.apache.beam.sdk.io.gcp.bigquery
PTransforms for reading and writing BigQuery tables.
BigQueryIO.Read - Class in org.apache.beam.sdk.io.gcp.bigquery
Implementation of BigQueryIO.read().
BigQueryIO.TypedRead<T> - Class in org.apache.beam.sdk.io.gcp.bigquery
BigQueryIO.TypedRead.Method - Enum in org.apache.beam.sdk.io.gcp.bigquery
Determines the method used to read data from BigQuery.
BigQueryIO.TypedRead.QueryPriority - Enum in org.apache.beam.sdk.io.gcp.bigquery
An enumeration type for the priority of a query.
BigQueryIO.Write<T> - Class in org.apache.beam.sdk.io.gcp.bigquery
Implementation of BigQueryIO.write().
BigQueryIO.Write.CreateDisposition - Enum in org.apache.beam.sdk.io.gcp.bigquery
An enumeration type for the BigQuery create disposition strings.
BigQueryIO.Write.Method - Enum in org.apache.beam.sdk.io.gcp.bigquery
Determines the method used to insert data in BigQuery.
BigQueryIO.Write.SchemaUpdateOption - Enum in org.apache.beam.sdk.io.gcp.bigquery
An enumeration type for the BigQuery schema update options strings.
BigQueryIO.Write.WriteDisposition - Enum in org.apache.beam.sdk.io.gcp.bigquery
An enumeration type for the BigQuery write disposition strings.
BigqueryMatcher - Class in org.apache.beam.sdk.io.gcp.testing
A matcher to verify data in BigQuery by processing given query and comparing with content's checksum.
BigqueryMatcher.TableAndQuery - Class in org.apache.beam.sdk.io.gcp.testing
 
BigQueryOptions - Interface in org.apache.beam.sdk.io.gcp.bigquery
Properties needed when using Google BigQuery with the Apache Beam SDK.
BigQuerySchemaRetrievalException - Exception in org.apache.beam.sdk.io.gcp.bigquery
Exception to signal that BigQuery schema retrieval failed.
BigQueryServices - Interface in org.apache.beam.sdk.io.gcp.bigquery
An interface for real, mock, or fake implementations of Cloud BigQuery services.
BigQueryServices.BigQueryServerStream<T> - Interface in org.apache.beam.sdk.io.gcp.bigquery
Container for reading data from streaming endpoints.
BigQueryServices.DatasetService - Interface in org.apache.beam.sdk.io.gcp.bigquery
An interface to get, create and delete Cloud BigQuery datasets and tables.
BigQueryServices.JobService - Interface in org.apache.beam.sdk.io.gcp.bigquery
An interface for the Cloud BigQuery load service.
BigQueryServices.StorageClient - Interface in org.apache.beam.sdk.io.gcp.bigquery
An interface representing a client object for making calls to the BigQuery Storage API.
BigQueryStorageQuerySource<T> - Class in org.apache.beam.sdk.io.gcp.bigquery
A Source representing reading the results of a query.
BigQueryStorageStreamSource<T> - Class in org.apache.beam.sdk.io.gcp.bigquery
A Source representing a single stream in a read session.
BigQueryStorageStreamSource.BigQueryStorageStreamReader<T> - Class in org.apache.beam.sdk.io.gcp.bigquery
A Source.Reader which reads records from a stream.
BigQueryStorageTableSource<T> - Class in org.apache.beam.sdk.io.gcp.bigquery
A Source representing reading from a table.
BigQueryTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.bigquery
BigQuery table provider.
BigQueryTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryTableProvider
 
BigQueryUtils - Class in org.apache.beam.sdk.io.gcp.bigquery
Utility methods for BigQuery related operations.
BigQueryUtils() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
 
BigQueryUtils.ConversionOptions - Class in org.apache.beam.sdk.io.gcp.bigquery
Options for how to convert BigQuery data to Beam data.
BigQueryUtils.ConversionOptions.Builder - Class in org.apache.beam.sdk.io.gcp.bigquery
BigQueryUtils.ConversionOptions.TruncateTimestamps - Enum in org.apache.beam.sdk.io.gcp.bigquery
Controls whether to truncate timestamps to millisecond precision lossily, or to crash when truncation would result.
BigtableIO - Class in org.apache.beam.sdk.io.gcp.bigtable
Transforms for reading from and writing to Google Cloud Bigtable.
BigtableIO.Read - Class in org.apache.beam.sdk.io.gcp.bigtable
A PTransform that reads from Google Cloud Bigtable.
BigtableIO.Write - Class in org.apache.beam.sdk.io.gcp.bigtable
A PTransform that writes to Google Cloud Bigtable.
BigtableIO.WriteWithResults - Class in org.apache.beam.sdk.io.gcp.bigtable
A PTransform that writes to Google Cloud Bigtable and emits a BigtableWriteResult for each batch written.
BigtableWriteResult - Class in org.apache.beam.sdk.io.gcp.bigtable
The result of writing a batch of rows to Bigtable.
BigtableWriteResult() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteResult
 
BigtableWriteResultCoder - Class in org.apache.beam.sdk.io.gcp.bigtable
A coder for BigtableWriteResult.
BigtableWriteResultCoder() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteResultCoder
 
BinaryCombineDoubleFn() - Constructor for class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
 
BinaryCombineFn() - Constructor for class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
 
BinaryCombineIntegerFn() - Constructor for class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
 
BinaryCombineLongFn() - Constructor for class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
 
binarySchema() - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.SchemaHelpers
 
bind(String, StateBinder) - Method in interface org.apache.beam.sdk.state.StateSpec
For internal use only; no backwards-compatibility guarantees.
bindBag(String, StateSpec<BagState<T>>, Coder<T>) - Method in interface org.apache.beam.sdk.state.StateBinder
 
bindCombining(String, StateSpec<CombiningState<InputT, AccumT, OutputT>>, Coder<AccumT>, Combine.CombineFn<InputT, AccumT, OutputT>) - Method in interface org.apache.beam.sdk.state.StateBinder
 
bindCombiningWithContext(String, StateSpec<CombiningState<InputT, AccumT, OutputT>>, Coder<AccumT>, CombineWithContext.CombineFnWithContext<InputT, AccumT, OutputT>) - Method in interface org.apache.beam.sdk.state.StateBinder
 
bindMap(String, StateSpec<MapState<KeyT, ValueT>>, Coder<KeyT>, Coder<ValueT>) - Method in interface org.apache.beam.sdk.state.StateBinder
 
bindSet(String, StateSpec<SetState<T>>, Coder<T>) - Method in interface org.apache.beam.sdk.state.StateBinder
 
bindValue(String, StateSpec<ValueState<T>>, Coder<T>) - Method in interface org.apache.beam.sdk.state.StateBinder
 
bindWatermark(String, StateSpec<WatermarkHoldState>, TimestampCombiner) - Method in interface org.apache.beam.sdk.state.StateBinder
Bind to a watermark StateSpec.
BitSetCoder - Class in org.apache.beam.sdk.coders
Coder for BitSet.
Block() - Constructor for class org.apache.beam.sdk.io.BlockBasedSource.Block
 
BlockBasedReader(BlockBasedSource<T>) - Constructor for class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
 
BlockBasedSource<T> - Class in org.apache.beam.sdk.io
A BlockBasedSource is a FileBasedSource where a file consists of blocks of records.
BlockBasedSource(String, EmptyMatchTreatment, long) - Constructor for class org.apache.beam.sdk.io.BlockBasedSource
Creates a BlockBasedSource based on a file name or pattern.
BlockBasedSource(String, long) - Constructor for class org.apache.beam.sdk.io.BlockBasedSource
BlockBasedSource(ValueProvider<String>, long) - Constructor for class org.apache.beam.sdk.io.BlockBasedSource
BlockBasedSource(ValueProvider<String>, EmptyMatchTreatment, long) - Constructor for class org.apache.beam.sdk.io.BlockBasedSource
BlockBasedSource(MatchResult.Metadata, long, long, long) - Constructor for class org.apache.beam.sdk.io.BlockBasedSource
Creates a BlockBasedSource for a single file.
BlockBasedSource.Block<T> - Class in org.apache.beam.sdk.io
A Block represents a block of records that can be read.
BlockBasedSource.BlockBasedReader<T> - Class in org.apache.beam.sdk.io
A Reader that reads records from a BlockBasedSource.
BlockingQueueIterator(BlockingQueue<T>) - Constructor for class org.apache.beam.sdk.fn.stream.DataStreams.BlockingQueueIterator
 
BOOLEAN - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
BOOLEAN - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
The type of boolean fields.
BooleanCoder - Class in org.apache.beam.sdk.coders
A Coder for Boolean.
BooleanCoder() - Constructor for class org.apache.beam.sdk.coders.BooleanCoder
 
booleans() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for Boolean.
Bounded(SparkContext, BoundedSource<T>, SerializablePipelineOptions, String) - Constructor for class org.apache.beam.runners.spark.io.SourceRDD.Bounded
 
BOUNDED_UNKNOWN - Static variable in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
 
BoundedReader() - Constructor for class org.apache.beam.sdk.io.BoundedSource.BoundedReader
 
BoundedReadFromUnboundedSource<T> - Class in org.apache.beam.sdk.io
PTransform that reads a bounded amount of data from an UnboundedSource, specified as one or both of a maximum number of elements or a maximum period of time to read.
BoundedSource<T> - Class in org.apache.beam.sdk.io
A Source that reads a finite amount of input and, because of that, supports some additional operations.
BoundedSource() - Constructor for class org.apache.beam.sdk.io.BoundedSource
 
BoundedSource.BoundedReader<T> - Class in org.apache.beam.sdk.io
A Reader that reads a bounded amount of input and supports some additional operations, such as progress estimation and dynamic work rebalancing.
BoundedSourceP<T> - Class in org.apache.beam.runners.jet.processors
Jet Processor implementation for reading from a bounded Beam source.
BoundedSourceWrapper<T> - Class in org.apache.beam.runners.gearpump.translators.io
wrapper over BoundedSource for Gearpump DataSource API.
BoundedSourceWrapper(BoundedSource<T>, PipelineOptions) - Constructor for class org.apache.beam.runners.gearpump.translators.io.BoundedSourceWrapper
 
BoundedWindow - Class in org.apache.beam.sdk.transforms.windowing
A BoundedWindow represents window information assigned to data elements.
BoundedWindow() - Constructor for class org.apache.beam.sdk.transforms.windowing.BoundedWindow
 
boundedWindowToGearpumpWindow(BoundedWindow) - Static method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils
 
boxIfPrimitive(TypeDescriptor) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
 
broadcast(JavaSparkContext) - Method in class org.apache.beam.runners.spark.util.SideInputBroadcast
 
bucketAccessible(GcsPath) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
Returns whether the GCS bucket exists and is accessible.
bucketOwner(GcsPath) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
Returns the project number of the project which owns this bucket.
BufferedExternalSorter - Class in org.apache.beam.sdk.extensions.sorter
Sorter that will use in memory sorting until the values can't fit into memory and will then fall back to external sorting.
BufferedExternalSorter.Options - Class in org.apache.beam.sdk.extensions.sorter
Contains configuration for the sorter.
BufferingStreamObserver<T> - Class in org.apache.beam.sdk.fn.stream
A thread safe StreamObserver which uses a bounded queue to pass elements to a processing thread responsible for interacting with the underlying CallStreamObserver.
BufferingStreamObserver(Phaser, CallStreamObserver<T>, ExecutorService, int) - Constructor for class org.apache.beam.sdk.fn.stream.BufferingStreamObserver
 
build() - Method in class org.apache.beam.runners.direct.WatermarkManager.TimerUpdate.TimerUpdateBuilder
Returns a new WatermarkManager.TimerUpdate with the most recently set completedTimers, setTimers, and deletedTimers.
build() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions.Builder
 
build() - Method in class org.apache.beam.sdk.extensions.gcp.util.CustomHttpErrors.Builder
 
build() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
Build function to create an instance of BeamSqlEnv based on preset fields.
build() - Method in class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase.ParameterListBuilder
 
build() - Method in class org.apache.beam.sdk.extensions.sql.impl.utils.SerializableRexNode.Builder
 
build() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table.Builder
 
build() - Method in class org.apache.beam.sdk.fn.test.TestStreams.Builder
 
build() - Method in class org.apache.beam.sdk.io.fs.CreateOptions.StandardCreateOptions.Builder
 
build() - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
 
build() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.ConversionOptions.Builder
 
build() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig.Builder
 
build() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction.Builder
 
build() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
 
build() - Method in class org.apache.beam.sdk.metrics.MetricsFilter.Builder
 
build() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Builder
 
build() - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
build() - Method in class org.apache.beam.sdk.schemas.Schema.Field.Builder
 
build() - Method in class org.apache.beam.sdk.schemas.Schema.Options.Builder
 
build() - Method in class org.apache.beam.sdk.values.Row.Builder
 
build() - Method in class org.apache.beam.sdk.values.Row.FieldValueBuilder
 
buildBeamSqlNullableSchema(Object...) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableUtils
 
buildBeamSqlSchema(Object...) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableUtils
Create a RowsBuilder with the specified row type info.
buildBeamSqlTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.avro.AvroTableProvider
 
buildBeamSqlTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryTableProvider
 
buildBeamSqlTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
 
buildBeamSqlTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datastore.DataStoreV1TableProvider
 
buildBeamSqlTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.KafkaTableProvider
 
buildBeamSqlTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTableProvider
 
buildBeamSqlTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.parquet.ParquetTableProvider
 
buildBeamSqlTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonTableProvider
 
buildBeamSqlTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.ReadOnlyTableProvider
 
buildBeamSqlTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.seqgen.GenerateSequenceTableProvider
 
buildBeamSqlTable(Table) - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.TableProvider
Build a BeamSqlTable using the given table meta info.
buildBeamSqlTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
 
buildBeamSqlTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTableProvider
 
buildBeamSqlTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.store.InMemoryMetaStore
 
builder(StructuralKey<?>) - Static method in class org.apache.beam.runners.direct.WatermarkManager.TimerUpdate
Creates a new WatermarkManager.TimerUpdate builder with the provided completed timers that needs the set and deleted timers to be added to it.
builder() - Static method in class org.apache.beam.runners.fnexecution.jobsubmission.JobPreparation
 
builder() - Static method in class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions
Builder() - Constructor for class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions.Builder
 
Builder() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.CustomHttpErrors.Builder
 
builder(TableProvider) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
Creates a builder with the default schema backed by the table provider.
builder() - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
 
builder() - Static method in class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase
Creates a ParameterListBuilder.
builder(RexNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.SerializableRexNode
 
Builder(RexNode) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.SerializableRexNode.Builder
 
builder() - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.avro.GenericRecordWriteConverter
 
builder() - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.parquet.GenericRecordReadConverter
 
builder() - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubMessageToRow
 
builder() - Static method in class org.apache.beam.sdk.extensions.sql.meta.Table
 
Builder() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.Table.Builder
 
builder() - Static method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.RetryConfiguration
 
builder() - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
Builder() - Constructor for class org.apache.beam.sdk.io.fs.CreateOptions.Builder
 
builder() - Static method in class org.apache.beam.sdk.io.fs.CreateOptions.StandardCreateOptions
Builder() - Constructor for class org.apache.beam.sdk.io.fs.CreateOptions.StandardCreateOptions.Builder
 
builder() - Static method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata
 
Builder() - Constructor for class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
 
builder() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.ConversionOptions
 
Builder() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.ConversionOptions.Builder
 
Builder() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig.Builder
 
Builder() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction.Builder
 
builder() - Static method in class org.apache.beam.sdk.io.kinesis.WatermarkParameters
 
builder() - Static method in class org.apache.beam.sdk.io.TextRowCountEstimator
 
Builder() - Constructor for class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
 
builder() - Static method in class org.apache.beam.sdk.metrics.MetricsFilter
 
Builder() - Constructor for class org.apache.beam.sdk.metrics.MetricsFilter.Builder
 
builder() - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor
 
Builder() - Constructor for class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Builder
 
Builder() - Constructor for class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
 
builder() - Static method in class org.apache.beam.sdk.schemas.Schema
 
Builder() - Constructor for class org.apache.beam.sdk.schemas.Schema.Builder
 
Builder() - Constructor for class org.apache.beam.sdk.schemas.Schema.Field.Builder
 
builder() - Static method in class org.apache.beam.sdk.schemas.Schema.Options
 
Builder() - Constructor for class org.apache.beam.sdk.transforms.DoFnSchemaInformation.Builder
 
builderFor(Endpoints.ApiServiceDescriptor) - Method in class org.apache.beam.sdk.fn.channel.ManagedChannelFactory
Create a ManagedChannelBuilder for the provided Endpoints.ApiServiceDescriptor.
builderFor(Endpoints.ApiServiceDescriptor) - Method in class org.apache.beam.sdk.fn.test.InProcessManagedChannelFactory
 
buildExternal(ExternalRead.Configuration) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.ReadBuilder
 
buildExternal(ExternalWrite.Configuration) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite.WriteBuilder
 
buildExternal(ConfigT) - Method in interface org.apache.beam.sdk.transforms.ExternalTransformBuilder
Builds the transform after it has been configured.
buildFrom(Descriptors.Descriptor) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
 
buildFrom(DescriptorProtos.FileDescriptorSet) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
 
buildFrom(Descriptors.FileDescriptor) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
 
buildFrom(InputStream) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
 
buildHCatRecords(int) - Static method in class org.apache.beam.sdk.io.hcatalog.test.HCatalogIOTestUtils
Returns a list of HCatRecords of passed size.
buildIOReader(PBegin) - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamPCollectionTable
 
buildIOReader(PBegin, BeamSqlTableFilter, List<String>) - Method in class org.apache.beam.sdk.extensions.sql.meta.BaseBeamTable
 
buildIOReader(PBegin) - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTable
create a PCollection<Row> from source.
buildIOReader(PBegin, BeamSqlTableFilter, List<String>) - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTable
create a PCollection<Row> from source with predicate and/or project pushed-down.
buildIOReader(PBegin) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.avro.AvroTable
 
buildIOReader(PBegin) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
 
buildIOReader(PBegin) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
 
buildIOReader(PBegin) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable
 
buildIOReader(PBegin, BeamSqlTableFilter, List<String>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable
 
buildIOReader(PBegin) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.parquet.ParquetTable
 
buildIOReader(PBegin) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestBoundedTable
 
buildIOReader(PBegin) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestUnboundedTable
 
buildIOReader(PBegin) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTable
 
buildIOWriter(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamPCollectionTable
 
buildIOWriter(PCollection<Row>) - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTable
create a IO.write() instance to write to target.
buildIOWriter(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.avro.AvroTable
 
buildIOWriter(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
 
buildIOWriter(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
 
buildIOWriter(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable
 
buildIOWriter(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.parquet.ParquetTable
 
buildIOWriter(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestBoundedTable
 
buildIOWriter(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTable
 
buildIOWriter(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTable
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamAggregationRel
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCoGBKJoinRel
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIntersectRel
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamMinusRel
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamPushDownIOSourceRel
 
buildPTransform() - Method in interface org.apache.beam.sdk.extensions.sql.impl.rel.BeamRelNode
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSideInputJoinRel
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSideInputLookupJoinRel
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSortRel
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUncollectRel
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnionRel
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnnestRel
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamValuesRel
 
buildPTransform() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcRel
 
buildRows(Schema, List<?>) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableUtils
Convenient way to build a BeamSqlRows.
buildTemporaryFilename(ResourceId, String) - Static method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
Constructs a temporary file resource given the temporary directory and a filename.
BUILTIN_AGGREGATOR_FACTORIES - Static variable in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamBuiltinAggregations
 
BuiltinStringFunctions - Class in org.apache.beam.sdk.extensions.sql.impl.udf
BuiltinStringFunctions.
BuiltinStringFunctions() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
BuiltinTrigonometricFunctions - Class in org.apache.beam.sdk.extensions.sql.impl.udf
TrigonometricFunctions.
BuiltinTrigonometricFunctions() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinTrigonometricFunctions
 
Bundle<T,CollectionT> - Interface in org.apache.beam.runners.local
An immutable collection of elements which are part of a PCollection.
BundleCheckpointHandler - Interface in org.apache.beam.runners.fnexecution.control
A handler which is invoked when the SDK returns BeamFnApi.DelayedBundleApplications as part of the bundle completion.
BundleFinalizationHandler - Interface in org.apache.beam.runners.fnexecution.control
A handler for the runner when a finalization request has been received.
BundleFinalizationHandlers - Class in org.apache.beam.runners.fnexecution.control
Utility methods for creating BundleFinalizationHandlers.
BundleFinalizationHandlers() - Constructor for class org.apache.beam.runners.fnexecution.control.BundleFinalizationHandlers
 
BundleFinalizationHandlers.InMemoryFinalizer - Class in org.apache.beam.runners.fnexecution.control
BundleProgressHandler - Interface in org.apache.beam.runners.fnexecution.control
A handler for bundle progress messages, both during bundle execution and on its completion.
BundleSplitHandler - Interface in org.apache.beam.runners.fnexecution.control
A handler which is invoked whenever an active bundle is split.
by(SerializableFunction<UserT, DestinationT>) - Method in class org.apache.beam.sdk.io.FileIO.Write
Specifies how to partition elements into groups ("destinations").
by(Contextful<Contextful.Fn<UserT, DestinationT>>) - Method in class org.apache.beam.sdk.io.FileIO.Write
By() - Constructor for class org.apache.beam.sdk.schemas.transforms.CoGroup.By
 
by(PredicateT) - Static method in class org.apache.beam.sdk.transforms.Filter
Returns a PTransform that takes an input PCollection<T> and returns a PCollection<T> with elements that satisfy the given predicate.
by(PredicateT) - Static method in class org.apache.beam.sdk.transforms.Filter
Binary compatibility adapter for Filter.by(ProcessFunction).
byFieldAccessDescriptor(FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.transforms.Group
Returns a transform that groups all elements in the input PCollection keyed by the fields specified.
byFieldIds(Integer...) - Static method in class org.apache.beam.sdk.schemas.transforms.Group
Returns a transform that groups all elements in the input PCollection keyed by the list of fields specified.
byFieldIds(Iterable<Integer>) - Static method in class org.apache.beam.sdk.schemas.transforms.Group
byFieldNames(String...) - Static method in class org.apache.beam.sdk.schemas.transforms.Group
Returns a transform that groups all elements in the input PCollection keyed by the list of fields specified.
byFieldNames(Iterable<String>) - Static method in class org.apache.beam.sdk.schemas.transforms.Group
ByFields() - Constructor for class org.apache.beam.sdk.schemas.transforms.Group.ByFields
 
BYTE - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
The type of byte fields.
ByteArray - Class in org.apache.beam.runners.spark.util
Serializable byte array.
ByteArray(byte[]) - Constructor for class org.apache.beam.runners.spark.util.ByteArray
 
ByteArrayCoder - Class in org.apache.beam.sdk.coders
A Coder for byte[].
ByteArrayKey(byte[]) - Constructor for class org.apache.beam.runners.jet.Utils.ByteArrayKey
 
ByteBuddyUtils - Class in org.apache.beam.sdk.schemas.utils
 
ByteBuddyUtils() - Constructor for class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils
 
ByteBuddyUtils.ConvertType - Class in org.apache.beam.sdk.schemas.utils
Give a Java type, returns the Java type expected for use with Row.
ByteBuddyUtils.ConvertValueForGetter - Class in org.apache.beam.sdk.schemas.utils
Takes a StackManipulation that returns a value.
ByteBuddyUtils.ConvertValueForSetter - Class in org.apache.beam.sdk.schemas.utils
Row is going to call the setter with its internal Java type, however the user object being set might have a different type internally.
ByteBuddyUtils.DefaultTypeConversionsFactory - Class in org.apache.beam.sdk.schemas.utils
 
ByteBuddyUtils.InjectPackageStrategy - Class in org.apache.beam.sdk.schemas.utils
A naming strategy for ByteBuddy classes.
ByteBuddyUtils.TransformingMap<K1,V1,K2,V2> - Class in org.apache.beam.sdk.schemas.utils
 
ByteBuddyUtils.TypeConversion<T> - Class in org.apache.beam.sdk.schemas.utils
 
ByteBuddyUtils.TypeConversionsFactory - Interface in org.apache.beam.sdk.schemas.utils
 
ByteCoder - Class in org.apache.beam.sdk.coders
A ByteCoder encodes Byte values in 1 byte using Java serialization.
ByteKey - Class in org.apache.beam.sdk.io.range
A class representing a key consisting of an array of bytes.
ByteKeyRange - Class in org.apache.beam.sdk.io.range
A class representing a range of ByteKeys.
ByteKeyRangeTracker - Class in org.apache.beam.sdk.io.range
ByteKeyRangeTracker - Class in org.apache.beam.sdk.transforms.splittabledofn
A RestrictionTracker for claiming ByteKeys in a ByteKeyRange in a monotonically increasing fashion.
BYTES - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
The type of bytes fields.
bytes() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for Byte.
bytesRead() - Static method in class org.apache.beam.sdk.metrics.SourceMetrics
Counter of bytes read by a source.
bytesReadBySplit(String) - Static method in class org.apache.beam.sdk.metrics.SourceMetrics
Counter of bytes read by a source split.
ByteStringCoder - Class in org.apache.beam.sdk.extensions.protobuf
A Coder for ByteString objects based on their encoded Protocol Buffer form.
bytesWritten() - Static method in class org.apache.beam.sdk.metrics.SinkMetrics
Counter of bytes written to a sink.

C

CACHED_CREATORS - Static variable in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
 
CACHED_CREATORS - Static variable in class org.apache.beam.sdk.schemas.utils.POJOUtils
 
CachedSideInputReader - Class in org.apache.beam.runners.spark.structuredstreaming.translation.utils
SideInputReader that caches materialized views.
CachedSideInputReader - Class in org.apache.beam.runners.spark.util
SideInputReader that caches materialized views.
CalciteConnectionWrapper - Class in org.apache.beam.sdk.extensions.sql.impl
Abstract wrapper for CalciteConnection to simplify extension.
CalciteConnectionWrapper(CalciteConnection) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
CalciteFactoryWrapper - Class in org.apache.beam.sdk.extensions.sql.impl
Wrapper for CalciteFactory.
CalciteQueryPlanner - Class in org.apache.beam.sdk.extensions.sql.impl
The core component to handle through a SQL statement, from explain execution plan, to generate a Beam pipeline.
CalciteQueryPlanner(JdbcConnection, RuleSet[]) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner
 
CalciteQueryPlanner.NonCumulativeCostImpl - Class in org.apache.beam.sdk.extensions.sql.impl
 
CalciteUtils - Class in org.apache.beam.sdk.extensions.sql.impl.utils
Utility methods for Calcite related operations.
CalciteUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
CalciteUtils.CharType - Class in org.apache.beam.sdk.extensions.sql.impl.utils
A LogicalType corresponding to CHAR.
CalciteUtils.DateType - Class in org.apache.beam.sdk.extensions.sql.impl.utils
A LogicalType corresponding to DATE.
CalciteUtils.TimestampWithLocalTzType - Class in org.apache.beam.sdk.extensions.sql.impl.utils
A LogicalType corresponding to TIMESTAMP_WITH_LOCAL_TIME_ZONE.
CalciteUtils.TimeType - Class in org.apache.beam.sdk.extensions.sql.impl.utils
A LogicalType corresponding to TIME.
CalciteUtils.TimeWithLocalTzType - Class in org.apache.beam.sdk.extensions.sql.impl.utils
A LogicalType corresponding to TIME_WITH_LOCAL_TIME_ZONE.
CalendarWindows - Class in org.apache.beam.sdk.transforms.windowing
A collection of WindowFns that windows values into calendar-based windows such as spans of days, months, or years.
CalendarWindows() - Constructor for class org.apache.beam.sdk.transforms.windowing.CalendarWindows
 
CalendarWindows.DaysWindows - Class in org.apache.beam.sdk.transforms.windowing
A WindowFn that windows elements into periods measured by days.
CalendarWindows.MonthsWindows - Class in org.apache.beam.sdk.transforms.windowing
A WindowFn that windows elements into periods measured by months.
CalendarWindows.YearsWindows - Class in org.apache.beam.sdk.transforms.windowing
A WindowFn that windows elements into periods measured by years.
call(Tuple2<ByteArray, byte[]>) - Method in class org.apache.beam.runners.spark.coders.CoderHelpers.FromByteFunction
 
call(Iterator<WindowedValue<InputT>>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.DoFnFunction
 
call(K, Iterator<WindowedValue<KV<K, InputT>>>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.GroupAlsoByWindowViaOutputBufferFn
 
cancel() - Method in class org.apache.beam.runners.apex.ApexRunnerResult
 
cancel() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
 
cancel() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
 
cancel() - Method in class org.apache.beam.runners.direct.DirectRunner.DirectPipelineResult
 
cancel() - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
 
cancel() - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
 
cancel(JobApi.CancelJobRequest, StreamObserver<JobApi.CancelJobResponse>) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService
 
cancel() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation
Cancel the job.
cancel() - Method in class org.apache.beam.runners.gearpump.GearpumpPipelineResult
 
cancel() - Method in class org.apache.beam.runners.jet.FailedRunningPipelineResults
 
cancel() - Method in class org.apache.beam.runners.jet.JetPipelineResult
 
cancel() - Method in class org.apache.beam.runners.spark.SparkPipelineResult
 
cancel() - Method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineResult
 
cancel() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
 
cancel() - Method in class org.apache.beam.sdk.fn.data.CompletableFutureInboundDataClient
 
cancel() - Method in interface org.apache.beam.sdk.fn.data.InboundDataClient
Cancels the client, causing it to drop any future inbound data.
cancel() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.BigQueryServerStream
Cancels the stream, releasing any client- and server-side resources.
cancel() - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices.FakeBigQueryServerStream
 
cancel() - Method in interface org.apache.beam.sdk.PipelineResult
Cancels the pipeline execution.
cancelled() - Method in interface org.apache.beam.runners.local.PipelineMessageReceiver
Report that the pipeline has been cancelled.
canConvertConvention(Convention) - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
 
canEqual(Object) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.DecodeUsingBeamCoder
 
canEqual(Object) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.EncodeUsingBeamCoder
 
CannotProvideCoderException - Exception in org.apache.beam.sdk.coders
The exception thrown when a CoderRegistry or CoderProvider cannot provide a Coder that has been requested.
CannotProvideCoderException(String) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
 
CannotProvideCoderException(String, CannotProvideCoderException.ReasonCode) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
 
CannotProvideCoderException(String, Throwable) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
 
CannotProvideCoderException(String, Throwable, CannotProvideCoderException.ReasonCode) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
 
CannotProvideCoderException(Throwable) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
 
CannotProvideCoderException(Throwable, CannotProvideCoderException.ReasonCode) - Constructor for exception org.apache.beam.sdk.coders.CannotProvideCoderException
 
CannotProvideCoderException.ReasonCode - Enum in org.apache.beam.sdk.coders
Indicates the reason that Coder inference failed.
canStopPolling(Instant, StateT) - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
Called by the Watch transform to determine whether the given termination state signals that Watch should stop calling Watch.Growth.PollFn for the current input, regardless of whether the last Watch.Growth.PollResult was complete or incomplete.
CassandraIO - Class in org.apache.beam.sdk.io.cassandra
An IO to read from Apache Cassandra.
CassandraIO.MutationType - Enum in org.apache.beam.sdk.io.cassandra
Specify the mutation type: either write or delete.
CassandraIO.Read<T> - Class in org.apache.beam.sdk.io.cassandra
A PTransform to read data from Apache Cassandra.
CassandraIO.Write<T> - Class in org.apache.beam.sdk.io.cassandra
A PTransform to mutate into Apache Cassandra.
Cast<T> - Class in org.apache.beam.sdk.schemas.transforms
Set of utilities for casting rows between schemas.
Cast() - Constructor for class org.apache.beam.sdk.schemas.transforms.Cast
 
Cast.CompatibilityError - Class in org.apache.beam.sdk.schemas.transforms
Describes compatibility errors during casting.
Cast.Narrowing - Class in org.apache.beam.sdk.schemas.transforms
Narrowing changes type without guarantee to preserve data.
Cast.Validator - Interface in org.apache.beam.sdk.schemas.transforms
Interface for statically validating casts.
Cast.Widening - Class in org.apache.beam.sdk.schemas.transforms
Widening changes to type that can represent any possible value of the original type.
castNumber(Number, Schema.TypeName, Schema.TypeName) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
 
castRow(Row, Schema, Schema) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
 
castValue(Object, Schema.FieldType, Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
 
channelNames - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
CHAR - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
CHAR_LENGTH - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
CHAR_LENGTH_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamBuiltinMethods
 
characters() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for Character.
charLength(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
CharType() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.CharType
 
check(RelNode) - Method in interface org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall.JoinChecker
 
checkDone() - Method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
 
checkDone() - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
 
checkDone() - Method in class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker
Called by the runner after DoFn.ProcessElement returns.
checkIfAnySubscriptionExists(String, Duration) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
Check if topics exist.
child() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.DecodeUsingBeamCoder
 
child() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.EncodeUsingBeamCoder
 
classesToTranslators() - Method in interface org.apache.beam.runners.dataflow.util.CoderCloudObjectTranslatorRegistrar
Gets a map from Coder to a CloudObjectTranslator that can translate that Coder.
classesToTranslators() - Method in class org.apache.beam.runners.dataflow.util.DefaultCoderCloudObjectTranslatorRegistrar
 
ClassLoaderLegacyArtifactRetrievalService - Class in org.apache.beam.runners.fnexecution.artifact
An LegacyArtifactRetrievalService that loads artifacts as ClassLoader resources.
ClassLoaderLegacyArtifactRetrievalService() - Constructor for class org.apache.beam.runners.fnexecution.artifact.ClassLoaderLegacyArtifactRetrievalService
 
ClassLoaderLegacyArtifactRetrievalService(ClassLoader) - Constructor for class org.apache.beam.runners.fnexecution.artifact.ClassLoaderLegacyArtifactRetrievalService
 
classNamesToTranslators() - Method in interface org.apache.beam.runners.dataflow.util.CoderCloudObjectTranslatorRegistrar
Gets a map from the name returned by CloudObject.getClassName() to a translator that can convert into the equivalent Coder.
classNamesToTranslators() - Method in class org.apache.beam.runners.dataflow.util.DefaultCoderCloudObjectTranslatorRegistrar
 
CLASSPATH_SCHEME - Static variable in class org.apache.beam.runners.apex.ApexRunner
 
ClassWithSchema() - Constructor for class org.apache.beam.sdk.schemas.utils.ReflectUtils.ClassWithSchema
 
cleanup() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
 
cleanupOnCancelOrFinish() - Method in class org.apache.beam.runners.apex.ApexRunnerResult
Opportunity for a subclass to perform cleanup, such as removing temporary files.
clear(K, W) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.BagUserStateHandler
Clears the bag user state for the given key and window.
clear() - Static method in class org.apache.beam.runners.spark.aggregators.AggregatorsAccumulator
 
clear() - Static method in class org.apache.beam.runners.spark.metrics.MetricsAccumulator
 
clear() - Static method in class org.apache.beam.runners.spark.structuredstreaming.aggregators.AggregatorsAccumulator
 
clear() - Static method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsAccumulator
 
clear() - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
 
clear() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
clear() - Method in interface org.apache.beam.sdk.state.State
Clear out the state location.
clearCache() - Static method in class org.apache.beam.runners.spark.io.MicrobatchSource
 
clearOutputElements() - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
clearOutputElements(TupleTag<T>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
clearWarnings() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
ClickHouseIO - Class in org.apache.beam.sdk.io.clickhouse
An IO to write to ClickHouse.
ClickHouseIO() - Constructor for class org.apache.beam.sdk.io.clickhouse.ClickHouseIO
 
ClickHouseIO.Write<T> - Class in org.apache.beam.sdk.io.clickhouse
A PTransform to write to ClickHouse.
ClickHouseWriter - Class in org.apache.beam.sdk.io.clickhouse
Writes Rows and field values using ClickHouseRowBinaryStream.
ClickHouseWriter() - Constructor for class org.apache.beam.sdk.io.clickhouse.ClickHouseWriter
 
clientBuffered(ExecutorService) - Static method in class org.apache.beam.sdk.fn.stream.OutboundObserverFactory
Create a buffering OutboundObserverFactory for client-side RPCs with the specified ExecutorService and the default buffer size.
clientBuffered(ExecutorService, int) - Static method in class org.apache.beam.sdk.fn.stream.OutboundObserverFactory
Create a buffering OutboundObserverFactory for client-side RPCs with the specified ExecutorService and buffer size.
ClientConfigurationFactory() - Constructor for class org.apache.beam.sdk.io.aws.options.AwsOptions.ClientConfigurationFactory
 
clientDirect() - Static method in class org.apache.beam.sdk.fn.stream.OutboundObserverFactory
Create the default OutboundObserverFactory for client-side RPCs, which uses basic unbuffered flow control.
Clock - Interface in org.apache.beam.runners.direct
Access to the current time.
clone() - Method in class org.apache.beam.runners.dataflow.util.CloudObject
 
clone() - Method in class org.apache.beam.runners.flink.metrics.MetricsAccumulator
 
clonesOf(OutT) - Static method in class org.apache.beam.sdk.transforms.SerializableFunctions
 
close() - Method in class org.apache.beam.runners.flink.metrics.FileReporter
 
close() - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactRetrievalService
 
close() - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactStagingService
 
close() - Method in class org.apache.beam.runners.fnexecution.control.DefaultExecutableStageContext
 
close() - Method in class org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory
 
close() - Method in class org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory.WrappedSdkHarnessClient
 
close() - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClient
 
close() - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService
 
close() - Method in interface org.apache.beam.runners.fnexecution.control.RemoteBundle
Closes this bundle.
close() - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor.ActiveBundle
Blocks until bundle processing is finished.
close() - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient
 
close() - Method in class org.apache.beam.runners.fnexecution.control.SingleEnvironmentInstanceJobBundleFactory
Deprecated.
 
close() - Method in class org.apache.beam.runners.fnexecution.data.GrpcDataService
 
close() - Method in class org.apache.beam.runners.fnexecution.environment.ProcessEnvironment
 
close() - Method in interface org.apache.beam.runners.fnexecution.environment.RemoteEnvironment
.
close() - Method in class org.apache.beam.runners.fnexecution.environment.StaticRemoteEnvironment
 
close() - Method in interface org.apache.beam.runners.fnexecution.FnService
.
close() - Method in class org.apache.beam.runners.fnexecution.GrpcFnServer
 
close() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService
 
close() - Method in class org.apache.beam.runners.fnexecution.logging.GrpcLoggingService
 
close() - Method in class org.apache.beam.runners.fnexecution.provisioning.StaticGrpcProvisionService
 
close() - Method in class org.apache.beam.runners.fnexecution.state.GrpcStateService
 
close() - Method in class org.apache.beam.runners.fnexecution.status.BeamWorkerStatusGrpcService
 
close() - Method in class org.apache.beam.runners.gearpump.translators.io.GearpumpSource
 
close() - Method in class org.apache.beam.runners.jet.processors.BoundedSourceP
 
close() - Method in class org.apache.beam.runners.jet.processors.UnboundedSourceP
 
close() - Method in class org.apache.beam.runners.portability.CloseableResource
Closes the underlying resource.
close(T) - Method in interface org.apache.beam.runners.portability.CloseableResource.Closer
 
close() - Method in class org.apache.beam.runners.portability.ExternalWorkerService
 
close() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
 
close() - Method in class org.apache.beam.sdk.expansion.service.ExpansionServer
 
close() - Method in class org.apache.beam.sdk.expansion.service.ExpansionService
 
close() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
close() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
 
close() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl
 
close() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer
 
close() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataSizeBasedBufferingOutboundObserver
 
close() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataTimeBasedBufferingOutboundObserver
 
close() - Method in interface org.apache.beam.sdk.fn.data.CloseableFnDataReceiver
.
close() - Method in class org.apache.beam.sdk.fn.stream.DataStreams.BlockingQueueIterator
 
close() - Method in class org.apache.beam.sdk.fn.stream.DataStreams.ElementDelimitedOutputStream
 
close() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
Closes the channel and returns the bundle result.
close() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
Closes any ReadableByteChannel created for the current reader.
close() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StorageClient
Close the client object.
close() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource.BigQueryStorageStreamReader
 
close() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
Gracefully close the underlying netty channel.
close() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
 
close() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
close() - Method in class org.apache.beam.sdk.io.hcatalog.test.EmbeddedMetastoreService
 
close() - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantDeserializer
 
close() - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantSerializer
 
close() - Method in class org.apache.beam.sdk.io.Source.Reader
Closes the reader.
close() - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.ThriftWriter
 
close() - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
CloseableFnDataReceiver<T> - Interface in org.apache.beam.sdk.fn.data
A receiver of streamed data that can be closed.
CloseableResource<T> - Class in org.apache.beam.runners.portability
An AutoCloseable that wraps a resource that needs to be cleaned up but does not implement AutoCloseable itself.
CloseableResource.CloseException - Exception in org.apache.beam.runners.portability
An exception that wraps errors thrown while a resource is being closed.
CloseableResource.Closer<T> - Interface in org.apache.beam.runners.portability
A function that knows how to clean up after a resource.
CloseableThrowingConsumer<ExceptionT extends java.lang.Exception,T> - Interface in org.apache.beam.sdk.function
A ThrowingConsumer that can be closed.
closeTo(double, double) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
CloudDebuggerOptions - Interface in org.apache.beam.runners.dataflow.options
Options for controlling Cloud Debugger.
CloudObject - Class in org.apache.beam.runners.dataflow.util
A representation of an arbitrary Java object to be instantiated by Dataflow workers.
cloudObjectClassName() - Method in interface org.apache.beam.runners.dataflow.util.CloudObjectTranslator
Gets the class name that will represent the CloudObject created by this CloudObjectTranslator.
cloudObjectClassName() - Method in class org.apache.beam.runners.dataflow.util.SchemaCoderCloudObjectTranslator
 
CloudObjects - Class in org.apache.beam.runners.dataflow.util
Utilities for converting an object to a CloudObject.
CloudObjectTranslator<T> - Interface in org.apache.beam.runners.dataflow.util
A translator that takes an object and creates a CloudObject which can be converted back to the original object.
CloudResourceManagerOptions - Interface in org.apache.beam.sdk.extensions.gcp.options
Properties needed when using Google CloudResourceManager with the Apache Beam SDK.
CO_GBK_RESULT_SCHEMA - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
CodahaleCsvSink - Class in org.apache.beam.runners.spark.structuredstreaming.metrics.sink
A Spark Sink that is tailored to report AggregatorMetric metrics to a CSV file.
CodahaleCsvSink(Properties, MetricRegistry, SecurityManager) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.metrics.sink.CodahaleCsvSink
 
CodahaleGraphiteSink - Class in org.apache.beam.runners.spark.structuredstreaming.metrics.sink
A Spark Sink that is tailored to report AggregatorMetric metrics to Graphite.
CodahaleGraphiteSink(Properties, MetricRegistry, SecurityManager) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.metrics.sink.CodahaleGraphiteSink
 
coder() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.TimerSpec
 
Coder<T> - Class in org.apache.beam.sdk.coders
A Coder<T> defines how to encode and decode values of type T into byte streams.
Coder() - Constructor for class org.apache.beam.sdk.coders.Coder
 
Coder() - Constructor for class org.apache.beam.sdk.io.range.OffsetRange.Coder
 
Coder.Context - Class in org.apache.beam.sdk.coders
Deprecated.
To implement a coder, do not use any Coder.Context. Just implement only those abstract methods which do not accept a Coder.Context and leave the default implementations for methods accepting a Coder.Context.
Coder.NonDeterministicException - Exception in org.apache.beam.sdk.coders
Exception thrown by Coder.verifyDeterministic() if the encoding is not deterministic, including details of why the encoding is not deterministic.
CoderCloudObjectTranslatorRegistrar - Interface in org.apache.beam.runners.dataflow.util
Coder authors have the ability to automatically have their Coder registered with the Dataflow Runner by creating a ServiceLoader entry and a concrete implementation of this interface.
coderConsistentWithEquals(Coder<T>, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<T> and values of type T, the values are equal if and only if the encoded bytes are equal.
coderConsistentWithEqualsInContext(Coder<T>, Coder.Context, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<T>, Coder.Context, and values of type T, the values are equal if and only if the encoded bytes are equal, in any Coder.Context.
coderDecodeEncodeContentsEqual(Coder<CollectionT>, CollectionT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<Collection<T>>, and value of type Collection<T>, encoding followed by decoding yields an equal value of type Collection<T>, in any Coder.Context.
coderDecodeEncodeContentsEqualInContext(Coder<CollectionT>, Coder.Context, CollectionT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<Collection<T>>, and value of type Collection<T>, encoding followed by decoding yields an equal value of type Collection<T>, in the given Coder.Context.
coderDecodeEncodeContentsInSameOrder(Coder<IterableT>, IterableT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<Collection<T>>, and value of type Collection<T>, encoding followed by decoding yields an equal value of type Collection<T>, in any Coder.Context.
coderDecodeEncodeContentsInSameOrderInContext(Coder<IterableT>, Coder.Context, IterableT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<Iterable<T>>, and value of type Iterable<T>, encoding followed by decoding yields an equal value of type Collection<T>, in the given Coder.Context.
coderDecodeEncodeEqual(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<T>, and value of type T, encoding followed by decoding yields an equal value of type T, in any Coder.Context.
coderDecodeEncodeEqualInContext(Coder<T>, Coder.Context, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<T>, Coder.Context, and value of type T, encoding followed by decoding yields an equal value of type T.
coderDecodeEncodeInContext(Coder<T>, Coder.Context, T, Matcher<T>) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<T>, Coder.Context, and value of type T, encoding followed by decoding yields a value of type T and tests that the matcher succeeds on the values.
coderDecodesBase64(Coder<T>, String, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
 
coderDecodesBase64(Coder<T>, List<String>, List<T>) - Static method in class org.apache.beam.sdk.testing.CoderProperties
 
coderDecodesBase64ContentsEqual(Coder<IterableT>, String, IterableT) - Static method in class org.apache.beam.sdk.testing.CoderProperties
 
coderDecodesBase64ContentsEqual(Coder<IterableT>, List<String>, List<IterableT>) - Static method in class org.apache.beam.sdk.testing.CoderProperties
 
coderDeterministic(Coder<T>, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<T>, and values of type T, if the values are equal then the encoded bytes are equal, in any Coder.Context.
coderDeterministicInContext(Coder<T>, Coder.Context, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<T>, Coder.Context, and values of type T, if the values are equal then the encoded bytes are equal.
coderEncodesBase64(Coder<T>, T, String) - Static method in class org.apache.beam.sdk.testing.CoderProperties
 
coderEncodesBase64(Coder<T>, List<T>, List<String>) - Static method in class org.apache.beam.sdk.testing.CoderProperties
 
CoderException - Exception in org.apache.beam.sdk.coders
An Exception thrown if there is a problem encoding or decoding a value.
CoderException(String) - Constructor for exception org.apache.beam.sdk.coders.CoderException
 
CoderException(String, Throwable) - Constructor for exception org.apache.beam.sdk.coders.CoderException
 
CoderException(Throwable) - Constructor for exception org.apache.beam.sdk.coders.CoderException
 
coderFor(TypeDescriptor<T>, List<? extends Coder<?>>) - Method in class org.apache.beam.sdk.coders.CoderProvider
Returns a Coder<T> to use for values of a particular type, given the Coders for each of the type's generic parameter types.
coderForFieldType(Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.SchemaCoder
 
coderFromCloudObject(CloudObject) - Static method in class org.apache.beam.runners.dataflow.util.CloudObjects
 
CoderHelpers - Class in org.apache.beam.runners.spark.coders
Serialization utility class.
CoderHelpers - Class in org.apache.beam.runners.spark.structuredstreaming.translation.helpers
Serialization utility class.
CoderHelpers.FromByteFunction<K,V> - Class in org.apache.beam.runners.spark.coders
A function for converting a byte array pair to a key-value pair.
CoderProperties - Class in org.apache.beam.sdk.testing
Properties for use in Coder tests.
CoderProperties() - Constructor for class org.apache.beam.sdk.testing.CoderProperties
 
CoderProperties.TestElementByteSizeObserver - Class in org.apache.beam.sdk.testing
An ElementByteSizeObserver that records the observed element sizes for testing purposes.
CoderProvider - Class in org.apache.beam.sdk.coders
A CoderProvider provides Coders.
CoderProvider() - Constructor for class org.apache.beam.sdk.coders.CoderProvider
 
CoderProviderRegistrar - Interface in org.apache.beam.sdk.coders
Coder creators have the ability to automatically have their coders registered with this SDK by creating a ServiceLoader entry and a concrete implementation of this interface.
CoderProviders - Class in org.apache.beam.sdk.coders
Static utility methods for creating and working with CoderProviders.
CoderRegistry - Class in org.apache.beam.sdk.coders
A CoderRegistry allows creating a Coder for a given Java class or type descriptor.
coderSerializable(Coder<T>) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that the given Coder<T> can be correctly serialized and deserialized.
CoGbkResult - Class in org.apache.beam.sdk.transforms.join
A row result of a CoGroupByKey.
CoGbkResult(CoGbkResultSchema, Iterable<RawUnionValue>) - Constructor for class org.apache.beam.sdk.transforms.join.CoGbkResult
A row in the PCollection resulting from a CoGroupByKey transform.
CoGbkResult(CoGbkResultSchema, Iterable<RawUnionValue>, int) - Constructor for class org.apache.beam.sdk.transforms.join.CoGbkResult
 
CoGbkResult.CoGbkResultCoder - Class in org.apache.beam.sdk.transforms.join
CoGbkResultSchema - Class in org.apache.beam.sdk.transforms.join
A schema for the results of a CoGroupByKey.
CoGbkResultSchema(TupleTagList) - Constructor for class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
Builds a schema from a tuple of TupleTag<?>s.
CoGroup - Class in org.apache.beam.sdk.schemas.transforms
A transform that performs equijoins across multiple schema PCollections.
CoGroup() - Constructor for class org.apache.beam.sdk.schemas.transforms.CoGroup
 
CoGroup.By - Class in org.apache.beam.sdk.schemas.transforms
Defines the set of fields to extract for the join key, as well as other per-input join options.
CoGroup.ExpandCrossProduct - Class in org.apache.beam.sdk.schemas.transforms
A PTransform that calculates the cross-product join.
CoGroup.Impl - Class in org.apache.beam.sdk.schemas.transforms
The implementing PTransform.
CoGroup.Result - Class in org.apache.beam.sdk.schemas.transforms
 
CoGroupByKey<K> - Class in org.apache.beam.sdk.transforms.join
A PTransform that performs a CoGroupByKey on a tuple of tables.
COLLECTION_TYPES - Static variable in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
CollectionCoder<T> - Class in org.apache.beam.sdk.coders
A CollectionCoder encodes Collections in the format of IterableLikeCoder.
CollectionCoder(Coder<T>) - Constructor for class org.apache.beam.sdk.coders.CollectionCoder
 
column(SqlParserPos, SqlIdentifier, SqlDataTypeSpec, SqlNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlDdlNodes
Creates a column declaration.
Column() - Constructor for class org.apache.beam.sdk.io.clickhouse.TableSchema.Column
 
columns() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema
 
columnType() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.Column
 
ColumnType() - Constructor for class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
Combine - Class in org.apache.beam.sdk.transforms
PTransforms for combining PCollection elements globally and per-key.
combine(Iterable<? extends Instant>) - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
Combines the given times, which must be from the same window and must have been passed through TimestampCombiner.merge(org.apache.beam.sdk.transforms.windowing.BoundedWindow, java.lang.Iterable<? extends org.joda.time.Instant>).
combine(Instant...) - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
Combine.AccumulatingCombineFn<InputT,AccumT extends Combine.AccumulatingCombineFn.Accumulator<InputT,AccumT,OutputT>,OutputT> - Class in org.apache.beam.sdk.transforms
A CombineFn that uses a subclass of Combine.AccumulatingCombineFn.Accumulator as its accumulator type.
Combine.AccumulatingCombineFn.Accumulator<InputT,AccumT,OutputT> - Interface in org.apache.beam.sdk.transforms
The type of mutable accumulator values used by this AccumulatingCombineFn.
Combine.BinaryCombineDoubleFn - Class in org.apache.beam.sdk.transforms
An abstract subclass of Combine.CombineFn for implementing combiners that are more easily and efficiently expressed as binary operations on doubles.
Combine.BinaryCombineFn<V> - Class in org.apache.beam.sdk.transforms
An abstract subclass of Combine.CombineFn for implementing combiners that are more easily expressed as binary operations.
Combine.BinaryCombineIntegerFn - Class in org.apache.beam.sdk.transforms
An abstract subclass of Combine.CombineFn for implementing combiners that are more easily and efficiently expressed as binary operations on ints
Combine.BinaryCombineLongFn - Class in org.apache.beam.sdk.transforms
An abstract subclass of Combine.CombineFn for implementing combiners that are more easily and efficiently expressed as binary operations on longs.
Combine.CombineFn<InputT,AccumT,OutputT> - Class in org.apache.beam.sdk.transforms
A CombineFn<InputT, AccumT, OutputT> specifies how to combine a collection of input values of type InputT into a single output value of type OutputT.
Combine.Globally<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
Combine.Globally<InputT, OutputT> takes a PCollection<InputT> and returns a PCollection<OutputT> whose elements are the result of combining all the elements in each window of the input PCollection, using a specified CombineFn<InputT, AccumT, OutputT>.
Combine.GloballyAsSingletonView<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
Combine.GloballyAsSingletonView<InputT, OutputT> takes a PCollection<InputT> and returns a PCollectionView<OutputT> whose elements are the result of combining all the elements in each window of the input PCollection, using a specified CombineFn<InputT, AccumT, OutputT>.
Combine.GroupedValues<K,InputT,OutputT> - Class in org.apache.beam.sdk.transforms
GroupedValues<K, InputT, OutputT> takes a PCollection<KV<K, Iterable<InputT>>>, such as the result of GroupByKey, applies a specified CombineFn<InputT, AccumT, OutputT> to each of the input KV<K, Iterable<InputT>> elements to produce a combined output KV<K, OutputT> element, and returns a PCollection<KV<K, OutputT>> containing all the combined output elements.
Combine.Holder<V> - Class in org.apache.beam.sdk.transforms
Holds a single value value of type V which may or may not be present.
Combine.IterableCombineFn<V> - Class in org.apache.beam.sdk.transforms
Converts a SerializableFunction from Iterable<V>s to Vs into a simple Combine.CombineFn over Vs.
Combine.PerKey<K,InputT,OutputT> - Class in org.apache.beam.sdk.transforms
PerKey<K, InputT, OutputT> takes a PCollection<KV<K, InputT>>, groups it by key, applies a combining function to the InputT values associated with each key to produce a combined OutputT value, and returns a PCollection<KV<K, OutputT>> representing a map from each distinct key of the input PCollection to the corresponding combined value.
Combine.PerKeyWithHotKeyFanout<K,InputT,OutputT> - Class in org.apache.beam.sdk.transforms
Like Combine.PerKey, but sharding the combining of hot keys.
Combine.SimpleCombineFn<V> - Class in org.apache.beam.sdk.transforms
Deprecated.
CombineFieldsByFields() - Constructor for class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
 
CombineFn() - Constructor for class org.apache.beam.sdk.transforms.Combine.CombineFn
 
combineFn() - Static method in class org.apache.beam.sdk.transforms.Count
Returns a Combine.CombineFn that counts the number of its inputs.
combineFn() - Static method in class org.apache.beam.sdk.transforms.Latest
Returns a Combine.CombineFn that selects the latest element among its inputs.
combineFn(int) - Static method in class org.apache.beam.sdk.transforms.Sample
Returns a Combine.CombineFn that computes a fixed-sized uniform sample of its inputs.
CombineFnBase - Class in org.apache.beam.sdk.transforms
For internal use only; no backwards-compatibility guarantees.
CombineFnBase() - Constructor for class org.apache.beam.sdk.transforms.CombineFnBase
 
CombineFnBase.GlobalCombineFn<InputT,AccumT,OutputT> - Interface in org.apache.beam.sdk.transforms
For internal use only; no backwards-compatibility guarantees.
CombineFns - Class in org.apache.beam.sdk.transforms
Static utility methods that create combine function instances.
CombineFns() - Constructor for class org.apache.beam.sdk.transforms.CombineFns
 
CombineFns.CoCombineResult - Class in org.apache.beam.sdk.transforms
A tuple of outputs produced by a composed combine functions.
CombineFns.ComposeCombineFnBuilder - Class in org.apache.beam.sdk.transforms
A builder class to construct a composed CombineFnBase.GlobalCombineFn.
CombineFns.ComposedCombineFn<DataT> - Class in org.apache.beam.sdk.transforms
A composed Combine.CombineFn that applies multiple CombineFns.
CombineFns.ComposedCombineFnWithContext<DataT> - Class in org.apache.beam.sdk.transforms
CombineFnTester - Class in org.apache.beam.sdk.testing
Utilities for testing CombineFns.
CombineFnTester() - Constructor for class org.apache.beam.sdk.testing.CombineFnTester
 
CombineFnWithContext() - Constructor for class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
 
CombineWithContext - Class in org.apache.beam.sdk.transforms
This class contains combine functions that have access to PipelineOptions and side inputs through CombineWithContext.Context.
CombineWithContext() - Constructor for class org.apache.beam.sdk.transforms.CombineWithContext
 
CombineWithContext.CombineFnWithContext<InputT,AccumT,OutputT> - Class in org.apache.beam.sdk.transforms
A combine function that has access to PipelineOptions and side inputs through CombineWithContext.Context.
CombineWithContext.Context - Class in org.apache.beam.sdk.transforms
Information accessible to all methods in CombineFnWithContext and KeyedCombineFnWithContext.
CombineWithContext.RequiresContextInternal - Interface in org.apache.beam.sdk.transforms
An internal interface for signaling that a GloballyCombineFn or a PerKeyCombineFn needs to access CombineWithContext.Context.
combining(Combine.CombineFn<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
Create a StateSpec for a CombiningState which uses a Combine.CombineFn to automatically merge multiple values of type InputT into a single resulting OutputT.
combining(CombineWithContext.CombineFnWithContext<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
For internal use only; no backwards compatibility guarantees
combining(Coder<AccumT>, Combine.CombineFn<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
Identical to #combining(CombineFn), but with an accumulator coder explicitly supplied.
combining(Coder<AccumT>, CombineWithContext.CombineFnWithContext<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
For internal use only; no backwards compatibility guarantees
combiningFromInputInternal(Coder<InputT>, Combine.CombineFn<InputT, AccumT, OutputT>) - Static method in class org.apache.beam.sdk.state.StateSpecs
For internal use only; no backwards-compatibility guarantees.
CombiningState<InputT,AccumT,OutputT> - Interface in org.apache.beam.sdk.state
A ReadableState cell defined by a Combine.CombineFn, accepting multiple input values, combining them as specified into accumulators, and producing a single output value.
comment(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.Table.Builder
 
commit() - Method in class org.apache.beam.runners.fnexecution.splittabledofn.SDFFeederViaStateAndTimers
Commits the state and timers: clears both if no checkpoint happened, or adjusts the restriction and sets a wake-up timer if a checkpoint happened.
commit() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
commitManifest(ArtifactApi.CommitManifestRequest, StreamObserver<ArtifactApi.CommitManifestResponse>) - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactStagingService
 
commitOffsetsInFinalize() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Finalized offsets are committed to Kafka.
commonPrefixLength(RandomAccessData, RandomAccessData) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.UnsignedLexicographicalComparator
Compute the length of the common prefix of the two provided sets of bytes.
compact(AccumT) - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
Returns an accumulator that represents the same logical value as the input accumulator, but may have a more compact representation.
compact(List<V>) - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
 
compact(Object[]) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
 
compact(Object[], CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
 
compact(AccumT, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
Returns an accumulator that represents the same logical value as the input accumulator, but may have a more compact representation.
compare(JobMessage, JobMessage) - Method in class org.apache.beam.runners.dataflow.util.MonitoringUtil.TimeStampComparator
 
compare(RandomAccessData, RandomAccessData) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.UnsignedLexicographicalComparator
 
compare(RandomAccessData, RandomAccessData, int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.UnsignedLexicographicalComparator
Compare the two sets of bytes starting at the given offset.
compare(T, T) - Method in class org.apache.beam.sdk.transforms.Top.Largest
Deprecated.
 
compare(T, T) - Method in class org.apache.beam.sdk.transforms.Top.Natural
 
compare(T, T) - Method in class org.apache.beam.sdk.transforms.Top.Reversed
 
compare(T, T) - Method in class org.apache.beam.sdk.transforms.Top.Smallest
Deprecated.
 
compare(KV<K, V>, KV<K, V>) - Method in class org.apache.beam.sdk.values.KV.OrderByKey
 
compare(KV<K, V>, KV<K, V>) - Method in class org.apache.beam.sdk.values.KV.OrderByValue
 
compareTo(ByteArray) - Method in class org.apache.beam.runners.spark.util.ByteArray
 
compareTo(Path) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
compareTo(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKey
ByteKey implements Comparable<ByteKey> by comparing the arrays in lexicographic order.
compareTo(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
 
CompatibilityError() - Constructor for class org.apache.beam.sdk.schemas.transforms.Cast.CompatibilityError
 
CompletableFutureInboundDataClient - Class in org.apache.beam.sdk.fn.data
An InboundDataClient backed by a CompletableFuture.
complete() - Method in class org.apache.beam.runners.jet.processors.BoundedSourceP
 
complete() - Method in class org.apache.beam.runners.jet.processors.ImpulseP
 
complete() - Method in class org.apache.beam.runners.jet.processors.StatefulParDoP
 
complete() - Method in class org.apache.beam.runners.jet.processors.UnboundedSourceP
 
complete() - Method in class org.apache.beam.runners.jet.processors.ViewP
 
complete() - Method in class org.apache.beam.runners.jet.processors.WindowGroupP
 
complete() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
 
complete() - Method in class org.apache.beam.sdk.fn.data.CompletableFutureInboundDataClient
 
complete() - Method in interface org.apache.beam.sdk.fn.data.InboundDataClient
Mark the client as completed.
complete(List<TimestampedValue<OutputT>>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
Constructs a Watch.Growth.PollResult with the given outputs and declares that there will be no new outputs for the current input.
complete(Instant, List<OutputT>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
Like Watch.Growth.PollResult.complete(List), but assigns the same timestamp to all new outputs.
completed() - Method in interface org.apache.beam.runners.local.PipelineMessageReceiver
Report that the pipeline has successfully completed.
COMPONENT_ENCODINGS - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
compose() - Static method in class org.apache.beam.sdk.transforms.CombineFns
compose(SerializableFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.PTransform
For a SerializableFunction<InputT, OutputT> fn, returns a PTransform given by applying fn.apply(v) to the input PCollection<InputT>.
compose(String, SerializableFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.PTransform
Like PTransform.compose(SerializableFunction), but with a custom name.
ComposeCombineFnBuilder() - Constructor for class org.apache.beam.sdk.transforms.CombineFns.ComposeCombineFnBuilder
 
COMPOSITE_TYPES - Static variable in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
CompositeSource - Class in org.apache.beam.runners.spark.metrics
Composite source made up of several MetricRegistry instances.
CompositeSource(String, MetricRegistry...) - Constructor for class org.apache.beam.runners.spark.metrics.CompositeSource
 
CompositeSource - Class in org.apache.beam.runners.spark.structuredstreaming.metrics
Composite source made up of several MetricRegistry instances.
CompositeSource(String, MetricRegistry...) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.metrics.CompositeSource
 
CompressedReader(CompressedSource<T>, FileBasedSource.FileBasedReader<T>) - Constructor for class org.apache.beam.sdk.io.CompressedSource.CompressedReader
Create a CompressedReader from a CompressedSource and delegate reader.
CompressedSource<T> - Class in org.apache.beam.sdk.io
A Source that reads from compressed files.
CompressedSource.CompressedReader<T> - Class in org.apache.beam.sdk.io
Reader for a CompressedSource.
CompressedSource.CompressionMode - Enum in org.apache.beam.sdk.io
Deprecated.
Use Compression instead
CompressedSource.DecompressingChannelFactory - Interface in org.apache.beam.sdk.io
Factory interface for creating channels that decompress the content of an underlying channel.
Compression - Enum in org.apache.beam.sdk.io
Various compression types for reading/writing files.
compute(Partition, TaskContext) - Method in class org.apache.beam.runners.spark.io.SourceRDD.Bounded
 
compute(Partition, TaskContext) - Method in class org.apache.beam.runners.spark.io.SourceRDD.Unbounded
 
computeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter
 
computeSelfCost(RelOptPlanner, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
 
CONCAT - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
concat(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
concat(String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
concat(String, String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
concat(String, String, String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
concat(String, String, String, String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
CONCAT_FIELD_NAMES - Static variable in class org.apache.beam.sdk.schemas.utils.SelectHelpers
This policy keeps all levels of a name.
CONCAT_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamBuiltinMethods
 
concatFieldNames() - Method in class org.apache.beam.sdk.schemas.transforms.Select.Flattened
For nested fields, concatenate all the names separated by a _ character in the flattened schema.
CONCRETE_CLASS - Static variable in class org.apache.beam.sdk.io.WriteFiles
For internal use by runners.
config() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
config() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
 
configuration - Variable in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver
 
Configuration() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.Configuration
 
Configuration() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite.Configuration
 
Configuration() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
 
Configuration() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.Write.External.Configuration
 
ConfigurationLocator() - Constructor for class org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptions.ConfigurationLocator
 
configure(Map<String, ?>, boolean) - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantDeserializer
 
configure(Map<String, ?>, boolean) - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantSerializer
 
configure() - Static method in class org.apache.beam.sdk.transforms.windowing.Window
Returns a new builder for a Window transform for setting windowing parameters other than the windowing function.
ConfluentSchemaRegistryDeserializerProvider<T> - Class in org.apache.beam.sdk.io.kafka
A DeserializerProvider that uses Confluent Schema Registry to resolve a Deserializers and Coder given a subject.
connect(String, Properties) - Method in class org.apache.beam.sdk.extensions.sql.impl.JdbcDriver
Configures Beam-specific options and opens a JDBC connection to Calcite.
connect(TableProvider, PipelineOptions) - Static method in class org.apache.beam.sdk.extensions.sql.impl.JdbcDriver
Connects to the driver using standard JdbcDriver.connect(String, Properties) call, but overrides the initial schema factory.
connect() - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
Connect to the Redis instance.
CONNECT_STRING_PREFIX - Static variable in class org.apache.beam.sdk.extensions.sql.impl.JdbcDriver
 
connection() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
ConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
 
ConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.ConnectionConfiguration
 
ConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
 
ConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.solr.SolrIO.ConnectionConfiguration
 
consistentWithEquals() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BigEndianShortCoder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BitSetCoder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.BooleanCoder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.ByteCoder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.Coder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.DoubleCoder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.DurationCoder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.FloatCoder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.InstantCoder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.KvCoder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
LengthPrefixCoder is consistent with equals if the nested Coder is.
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.ListCoder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.MapCoder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.NullableCoder
NullableCoder is consistent with equals if the nested Coder is.
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.SerializableCoder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.VarIntCoder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.coders.VarLongCoder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
Returns true if this Coder is injective with respect to Object.equals(java.lang.Object).
consistentWithEquals() - Method in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow.Coder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.io.fs.MetadataCoder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.io.fs.MetadataCoderV2
 
consistentWithEquals() - Method in class org.apache.beam.sdk.io.fs.ResourceIdCoder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.io.range.OffsetRange.Coder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
 
consistentWithEquals() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
 
ConsoleIO - Class in org.apache.beam.runners.spark.io
Print to console.
ConsoleIO.Write - Class in org.apache.beam.runners.spark.io
Write to console.
ConsoleIO.Write.Unbound<T> - Class in org.apache.beam.runners.spark.io
PTransform writing PCollection to the console.
constant(FileBasedSink.FilenamePolicy, SerializableFunction<UserT, OutputT>) - Static method in class org.apache.beam.sdk.io.DynamicFileDestinations
Returns a FileBasedSink.DynamicDestinations that always returns the same FileBasedSink.FilenamePolicy.
constant(FileBasedSink.FilenamePolicy) - Static method in class org.apache.beam.sdk.io.DynamicFileDestinations
A specialization of #constant(FilenamePolicy, SerializableFunction) for the case where UserT and OutputT are the same type and the format function is the identity.
constant(OutT) - Static method in class org.apache.beam.sdk.transforms.SerializableFunctions
 
CONSTANT_WINDOW_SIZE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
 
constantDestinations(FileBasedSink.FilenamePolicy, Schema, Map<String, Object>, CodecFactory, SerializableFunction<UserT, OutputT>) - Static method in class org.apache.beam.sdk.io.AvroIO
Returns a DynamicAvroDestinations that always returns the same FileBasedSink.FilenamePolicy, schema, metadata, and codec.
constructFilter(List<RexNode>) - Method in class org.apache.beam.sdk.extensions.sql.meta.BaseBeamTable
 
constructFilter(List<RexNode>) - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTable
Generate an IO implementation of BeamSqlTableFilter for predicate push-down.
constructFilter(List<RexNode>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable
 
contains(PCollectionView<T>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.SparkSideInputReader
 
contains(PCollectionView<T>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.utils.CachedSideInputReader
 
contains(PCollectionView<T>) - Method in class org.apache.beam.runners.spark.util.CachedSideInputReader
 
contains(PCollectionView<T>) - Method in class org.apache.beam.runners.spark.util.SparkSideInputReader
 
contains(Descriptors.Descriptor) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
 
contains(T) - Method in interface org.apache.beam.sdk.state.SetState
Returns true if this set contains the specified element.
contains(T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.contains(Object[]).
contains(Coder<T>, T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.contains(Object[]).
contains(SerializableMatcher<? super T>...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
contains(List<SerializableMatcher<? super T>>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.contains(List).
contains(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
Returns whether this window contains the given window.
containsInAnyOrder(T...) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
Asserts that the iterable in question contains the provided elements.
containsInAnyOrder(Iterable<T>) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
Asserts that the iterable in question contains the provided elements.
containsInAnyOrder(T...) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
Checks that the Iterable contains the expected elements, in any order.
containsInAnyOrder(Iterable<T>) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
Checks that the Iterable contains the expected elements, in any order.
containsInAnyOrder(T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
containsInAnyOrder(Coder<T>, T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
containsInAnyOrder(SerializableMatcher<? super T>...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
containsInAnyOrder(Collection<SerializableMatcher<? super T>>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
containsKey(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
Returns true if the specified ByteKey is contained within this range.
containsKey(Object) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
containsSeekableInput(RelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
This method returns whether any of the children of the relNode are Seekable.
containsString(String) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
containsValue(Object) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
Context(boolean) - Constructor for class org.apache.beam.sdk.coders.Coder.Context
Deprecated.
 
Context(TableDataInsertAllResponse.InsertErrors) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy.Context
 
Context() - Constructor for class org.apache.beam.sdk.schemas.utils.SchemaZipFold.Context
 
Context() - Constructor for class org.apache.beam.sdk.transforms.CombineWithContext.Context
 
Context() - Constructor for class org.apache.beam.sdk.transforms.Contextful.Fn.Context
 
Contextful<ClosureT> - Class in org.apache.beam.sdk.transforms
Pair of a bit of user code (a "closure") and the Requirements needed to run it.
Contextful.Fn<InputT,OutputT> - Interface in org.apache.beam.sdk.transforms
A function from an input to an output that may additionally access Contextful.Fn.Context when computing the result.
Contextful.Fn.Context - Class in org.apache.beam.sdk.transforms
An accessor for additional capabilities available in Contextful.Fn.apply(InputT, org.apache.beam.sdk.transforms.Contextful.Fn.Context).
contextSideInput - Variable in class org.apache.beam.sdk.extensions.ml.AnnotateVideo
 
continuously(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.FileIO.Match
continuously(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.FileIO.MatchAll
continuously(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
Continuously watches for new files at the given interval until the given termination condition is reached, where the input to the condition is the filepattern.
control(StreamObserver<BeamFnApi.InstructionRequest>) - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService
Called by gRPC for each incoming connection from an SDK harness, and enqueue an available SDK harness client.
ControlClientPool - Interface in org.apache.beam.runners.fnexecution.control
A pool of control clients that brokers incoming SDK harness connections (in the form of InstructionRequestHandlers.
ControlClientPool.Sink - Interface in org.apache.beam.runners.fnexecution.control
A sink for InstructionRequestHandlers keyed by worker id.
ControlClientPool.Source - Interface in org.apache.beam.runners.fnexecution.control
ConversionContext - Class in org.apache.beam.sdk.extensions.sql.zetasql.translation
Conversion context, some rules need this data to convert the nodes.
ConversionOptions() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.ConversionOptions
 
convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamCalcRule
 
convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamEnumerableConverterRule
 
convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIntersectRule
 
convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIOSinkRule
 
convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamMinusRule
 
convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSideInputLookupJoinRule
 
convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSortRule
 
convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamUncollectRule
 
convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamUnionRule
 
convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamValuesRule
 
convert() - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable.RowToDocument
 
convert(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcRule
 
convert(ResolvedNodes.ResolvedQueryStmt, List<RelNode>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter
 
convert(SqlNode) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl
 
Convert - Class in org.apache.beam.sdk.schemas.transforms
A set of utilities for converting between different objects supporting schemas.
Convert() - Constructor for class org.apache.beam.sdk.schemas.transforms.Convert
 
convert(TypeDescriptor) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
 
convertArray(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
 
convertArray(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
 
convertArray(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
 
convertArray(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
 
convertAvroFieldStrict(Object, Schema, Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Strict conversion from AVRO to Beam, strict because it doesn't do widening or narrowing during conversion.
convertAvroFormat(Schema.FieldType, Object, BigQueryUtils.ConversionOptions) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
Tries to convert an Avro decoded value to a Beam field value based on the target type of the Beam field.
convertByteBuffer(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
 
convertByteBuffer(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
 
convertByteBuffer(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
 
convertByteBuffer(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
 
convertCharSequence(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
 
convertCharSequence(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
 
convertCharSequence(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
 
convertCharSequence(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
 
convertCollection(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
 
convertCollection(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
 
convertCollection(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
 
convertCollection(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
 
convertDateTime(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
 
convertDateTime(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
 
convertDateTime(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
 
convertDateTime(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
 
convertDateValueToDateString(Value) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
convertDefault(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.AvroUtils.AvroConvertType
 
convertDefault(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.AvroUtils.AvroConvertValueForGetter
 
convertDefault(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.AvroUtils.AvroConvertValueForSetter
 
convertDefault(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
 
convertDefault(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
 
convertDefault(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
 
convertDefault(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
 
ConvertedSchemaInformation(SchemaCoder<T>, Schema.FieldType) - Constructor for class org.apache.beam.sdk.schemas.utils.ConvertHelpers.ConvertedSchemaInformation
 
convertEnum(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
 
convertEnum(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
 
convertEnum(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
 
convertEnum(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
 
convertGenericRecordToTableRow(GenericRecord, TableSchema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
 
ConvertHelpers - Class in org.apache.beam.sdk.schemas.utils
Helper functions for converting between equivalent schema types.
ConvertHelpers() - Constructor for class org.apache.beam.sdk.schemas.utils.ConvertHelpers
 
ConvertHelpers.ConvertedSchemaInformation<T> - Class in org.apache.beam.sdk.schemas.utils
Return value after converting a schema.
convertIterable(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
 
convertIterable(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
 
convertIterable(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
 
convertIterable(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
 
convertList(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
 
convertList(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
 
convertList(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
 
convertList(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
 
convertMap(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
 
convertMap(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
 
convertMap(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
 
convertMap(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
 
convertNumbers(TableRow) - Static method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
 
convertPrimitive(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
 
convertPrimitive(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
 
convertPrimitive(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
 
convertPrimitive(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
 
convertRelOptCost(RelOptCost) - Static method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
convertResolvedLiteral(ResolvedNodes.ResolvedLiteral) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
Convert a resolved literal to a RexNode.
convertRexNodeFromResolvedExpr(ResolvedNodes.ResolvedExpr, List<ResolvedColumn>, List<RelDataTypeField>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
Create a RexNode for a corresponding resolved expression node.
convertRexNodeFromResolvedExpr(ResolvedNodes.ResolvedExpr) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
Create a RexNode for a corresponding resolved expression.
convertRexNodeFromResolvedExprWithRefScan(ResolvedNodes.ResolvedExpr, List<ResolvedColumn>, List<RelDataTypeField>, List<ResolvedColumn>, List<ResolvedColumn>, List<RelDataTypeField>, List<ResolvedColumn>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
Extract the RexNode from expression with ref scan.
convertRootQuery(ConversionContext, ResolvedNodes.ResolvedQueryStmt) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.QueryStatementConverter
 
convertTimeValueToTimeString(Value) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
convertToBagSpecInternal(StateSpec<CombiningState<InputT, AccumT, OutputT>>) - Static method in class org.apache.beam.sdk.state.StateSpecs
For internal use only; no backwards-compatibility guarantees.
convertToBeamRel(String, QueryPlanner.QueryParameters) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner
It parses and validate the input query, then convert into a BeamRelNode tree.
convertToBeamRel(String, QueryPlanner.QueryParameters) - Method in interface org.apache.beam.sdk.extensions.sql.impl.QueryPlanner
It parses and validate the input query, then convert into a BeamRelNode tree.
convertToBeamRel(String) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
 
convertToBeamRel(String, Map<String, Value>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
 
convertToBeamRel(String, List<Value>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
 
convertToBeamRel(String, QueryPlanner.QueryParameters) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
 
convertToFileResourceIfPossible(String) - Static method in class org.apache.beam.sdk.io.FileBasedSink
This is a helper function for turning a user-provided output filename prefix and converting it into a ResourceId for writing output files.
ConvertType(boolean) - Constructor for class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertType
 
ConvertValueForGetter(StackManipulation) - Constructor for class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
 
ConvertValueForSetter(StackManipulation) - Constructor for class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
 
copy() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
Returns a copy of this RandomAccessData.
copy() - Method in class org.apache.beam.runners.spark.aggregators.NamedAggregatorsAccumulator
 
copy() - Method in class org.apache.beam.runners.spark.metrics.MetricsContainerStepMapAccumulator
 
copy() - Method in class org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregatorsAccumulator
 
copy() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsContainerStepMapAccumulator
 
copy(Iterable<String>, Iterable<String>) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
 
copy(RelTraitSet, RelNode, ImmutableBitSet, List<ImmutableBitSet>, List<AggregateCall>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamAggregationRel
 
copy(RelTraitSet, RelNode, RexProgram) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel
 
copy(RelTraitSet, RexNode, RelNode, RelNode, JoinRelType, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCoGBKJoinRel
 
copy(RelTraitSet, List<RelNode>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter
 
copy(RelTraitSet, List<RelNode>, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIntersectRel
 
copy(RelTraitSet, List<RelNode>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
 
copy(RelTraitSet, List<RelNode>, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamMinusRel
 
copy(RelTraitSet, RexNode, RelNode, RelNode, JoinRelType, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSideInputJoinRel
 
copy(RelTraitSet, RexNode, RelNode, RelNode, JoinRelType, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSideInputLookupJoinRel
 
copy(RelTraitSet, RelNode, RelCollation, RexNode, RexNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSortRel
 
copy(RelTraitSet, RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUncollectRel
 
copy(RelTraitSet, List<RelNode>, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnionRel
 
copy(RelTraitSet, RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnnestRel
 
copy(RelTraitSet, RelNode, RexProgram) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcRel
 
copy(List<ResourceIdT>, List<ResourceIdT>) - Method in class org.apache.beam.sdk.io.FileSystem
Copies a List of file-like resources from one location to another.
copy(List<ResourceId>, List<ResourceId>, MoveOptions...) - Static method in class org.apache.beam.sdk.io.FileSystems
Copies a List of file-like resources from one location to another.
copyFrom(ByteBuffer) - Static method in class org.apache.beam.sdk.io.range.ByteKey
Creates a new ByteKey backed by a copy of the data remaining in the specified ByteBuffer.
copyFrom(byte[]) - Static method in class org.apache.beam.sdk.io.range.ByteKey
Creates a new ByteKey backed by a copy of the specified byte[].
copyFrom(FieldSpecifierNotationParser.DotExpressionComponentContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionComponentContext
 
copyFrom(FieldSpecifierNotationParser.QualifierListContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifierListContext
 
copyResourcesFromJar(JarFile) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.PortablePipelineJarCreator
Copy resources from inputJar to PortablePipelineJarCreator.outputStream.
cosh(Double) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinTrigonometricFunctions
COSH(X)
Count - Class in org.apache.beam.sdk.transforms
PTransforms to count the elements in a PCollection.
countAsserts(Pipeline) - Static method in class org.apache.beam.sdk.testing.PAssert
 
COUNTER - Static variable in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTable
 
Counter - Interface in org.apache.beam.sdk.metrics
A metric that reports a single long value and can be incremented or decremented.
counter(String, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
Create a metric that can be incremented and decremented, and is aggregated by taking the sum.
counter(Class<?>, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
Create a metric that can be incremented and decremented, and is aggregated by taking the sum.
CounterImpl - Class in org.apache.beam.runners.jet.metrics
Implementation of Counter.
CounterMark(long, Instant) - Constructor for class org.apache.beam.sdk.io.CountingSource.CounterMark
Creates a checkpoint mark reflecting the last emitted value.
CountingSource - Class in org.apache.beam.sdk.io
Most users should use GenerateSequence instead.
CountingSource.CounterMark - Class in org.apache.beam.sdk.io
The checkpoint for an unbounded CountingSource is simply the last value produced.
CountWords() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.CountWords
 
CovarianceFn<T extends java.lang.Number> - Class in org.apache.beam.sdk.extensions.sql.impl.transform.agg
Combine.CombineFn for Covariance on Number types.
CrashingRunner - Class in org.apache.beam.sdk.testing
A PipelineRunner that applies no overrides and throws an exception on calls to Pipeline.run().
CrashingRunner() - Constructor for class org.apache.beam.sdk.testing.CrashingRunner
 
create(DataflowPipelineOptions) - Static method in class org.apache.beam.runners.dataflow.DataflowClient
 
create(PCollectionView<?>, Coder<T>) - Static method in class org.apache.beam.runners.dataflow.DataflowRunner.StreamingPCollectionViewWriterFn
 
create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions.DataflowClientFactory
 
create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions.StagerFactory
 
create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowPipelineOptions.DefaultGcpRegionFactory
 
create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowPipelineOptions.StagingLocationFactory
 
create(PipelineOptions) - Method in class org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions.WorkerHarnessContainerImageFactory
 
create() - Static method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.PayloadTranslator
 
create(PipelineOptions) - Method in class org.apache.beam.runners.direct.DirectOptions.AvailableParallelismFactory
 
create(Clock, ExecutableGraph<ExecutableT, ? super CollectionT>, Function<ExecutableT, String>) - Static method in class org.apache.beam.runners.direct.WatermarkManager
Creates a new WatermarkManager.
create(FlinkJobServerDriver.FlinkServerConfiguration) - Static method in class org.apache.beam.runners.flink.FlinkJobInvoker
 
create(boolean) - Static method in class org.apache.beam.runners.flink.TestFlinkRunner
 
create() - Static method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactRetrievalService
 
create(JobInfo) - Static method in class org.apache.beam.runners.fnexecution.control.DefaultExecutableStageContext
 
create(JobInfo) - Static method in class org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory
 
create(JobInfo, Map<String, EnvironmentFactory.Provider>) - Static method in class org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory
 
create() - Static method in class org.apache.beam.runners.fnexecution.control.MapControlClientPool
create(String) - Method in interface org.apache.beam.runners.fnexecution.control.OutputReceiverFactory
Get a new FnDataReceiver for an output PCollection.
create(ReferenceCountingExecutableStageContextFactory.Creator, SerializableFunction<Object, Boolean>) - Static method in class org.apache.beam.runners.fnexecution.control.ReferenceCountingExecutableStageContextFactory
 
create(EnvironmentFactory, GrpcFnServer<GrpcDataService>, GrpcFnServer<GrpcStateService>, IdGenerator) - Static method in class org.apache.beam.runners.fnexecution.control.SingleEnvironmentInstanceJobBundleFactory
Deprecated.
 
create(String, String) - Method in class org.apache.beam.runners.fnexecution.control.TimerReceiverFactory
 
create(PipelineOptions, ExecutorService, OutboundObserverFactory) - Static method in class org.apache.beam.runners.fnexecution.data.GrpcDataService
 
create(PipelineOptions, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<FnApiControlClientPoolService>, ControlClientPool.Source) - Static method in class org.apache.beam.runners.fnexecution.environment.EmbeddedEnvironmentFactory
 
create(GrpcFnServer<FnApiControlClientPoolService>, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<LegacyArtifactRetrievalService>, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool.Source, IdGenerator) - Static method in class org.apache.beam.runners.fnexecution.environment.ExternalEnvironmentFactory
 
create(ProcessManager, RunnerApi.Environment, String, InstructionRequestHandler) - Static method in class org.apache.beam.runners.fnexecution.environment.ProcessEnvironment
 
create(ProcessManager, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool.Source, IdGenerator, PipelineOptions) - Static method in class org.apache.beam.runners.fnexecution.environment.ProcessEnvironmentFactory
 
create() - Static method in class org.apache.beam.runners.fnexecution.environment.ProcessManager
 
create(ServiceT, Endpoints.ApiServiceDescriptor, ServerFactory) - Static method in class org.apache.beam.runners.fnexecution.GrpcFnServer
Create a GrpcFnServer for the provided FnService which will run at the endpoint specified in the Endpoints.ApiServiceDescriptor.
create(ServiceT, Endpoints.ApiServiceDescriptor) - Static method in class org.apache.beam.runners.fnexecution.GrpcFnServer
Deprecated.
This create function is used for Dataflow migration purpose only.
create() - Static method in class org.apache.beam.runners.fnexecution.InProcessServerFactory
 
create(List<BindableService>, Endpoints.ApiServiceDescriptor) - Method in class org.apache.beam.runners.fnexecution.InProcessServerFactory
 
create(Endpoints.ApiServiceDescriptor, Function<String, String>, ThrowingConsumer<Exception, String>, JobInvoker) - Static method in class org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService
Creates an InMemoryJobService.
create(Endpoints.ApiServiceDescriptor, Function<String, String>, ThrowingConsumer<Exception, String>, JobInvoker, int) - Static method in class org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService
Creates an InMemoryJobService.
create() - Method in interface org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver.JobInvokerFactory
 
create(String, String, String, Struct) - Static method in class org.apache.beam.runners.fnexecution.provisioning.JobInfo
 
create(ProvisionApi.ProvisionInfo) - Static method in class org.apache.beam.runners.fnexecution.provisioning.StaticGrpcProvisionService
 
create(List<BindableService>, Endpoints.ApiServiceDescriptor) - Method in class org.apache.beam.runners.fnexecution.ServerFactory
Creates an instance of this server at the address specified by the given service descriptor and bound to multiple services.
create(List<BindableService>, Endpoints.ApiServiceDescriptor) - Method in class org.apache.beam.runners.fnexecution.ServerFactory.InetSocketAddressServerFactory
 
create() - Static method in class org.apache.beam.runners.fnexecution.state.GrpcStateService
Create a new GrpcStateService.
create(Endpoints.ApiServiceDescriptor, HeaderAccessor) - Static method in class org.apache.beam.runners.fnexecution.status.BeamWorkerStatusGrpcService
Create new instance of BeamWorkerStatusGrpcService.
create(PipelineOptions) - Method in class org.apache.beam.runners.portability.testing.TestPortablePipelineOptions.DefaultJobServerConfigFactory
 
create(PipelineOptions) - Method in class org.apache.beam.runners.spark.SparkCommonPipelineOptions.TmpCheckpointDirFactory
 
create(PipelineOptions) - Method in class org.apache.beam.runners.spark.SparkContextOptions.EmptyListenersList
 
create(SparkJobServerDriver.SparkServerConfiguration) - Static method in class org.apache.beam.runners.spark.SparkJobInvoker
 
create() - Static method in class org.apache.beam.runners.spark.SparkRunner
Creates and returns a new SparkRunner with default options.
create(SparkPipelineOptions) - Static method in class org.apache.beam.runners.spark.SparkRunner
Creates and returns a new SparkRunner with specified options.
create() - Static method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingRunner
Creates and returns a new SparkStructuredStreamingRunner with default options.
create(SparkStructuredStreamingPipelineOptions) - Static method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingRunner
Creates and returns a new SparkStructuredStreamingRunner with specified options.
create(PipelineOptions) - Method in class org.apache.beam.runners.spark.TestSparkPipelineOptions.DefaultStopPipelineWatermarkFactory
 
create(byte[], Coder<T>) - Static method in class org.apache.beam.runners.spark.util.SideInputBroadcast
 
create(ExpansionService, String, int) - Static method in class org.apache.beam.sdk.expansion.service.ExpansionServer
Create a ExpansionServer for the provided ExpansionService running on an arbitrary port.
create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.DefaultProjectFactory
 
create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.EnableStreamingEngineFactory
 
create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpTempLocationFactory
 
create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpUserCredentialsFactory
 
create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcsOptions.ExecutorServiceFactory
 
create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.options.GcsOptions.PathValidatorFactory
 
create(GcsPath, String) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
Creates an object in GCS.
create(GcsPath, String, Integer) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
Same as GcsUtil.create(GcsPath, String) but allows overriding {code uploadBufferSizeBytes}.
create(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.GcsUtilFactory
Returns an instance of GcsUtil based on the PipelineOptions.
create(PipelineOptions, Storage, HttpRequestInitializer, ExecutorService, Integer) - Static method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.GcsUtilFactory
Returns an instance of GcsUtil based on the given parameters.
create(StorageObject) - Static method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.StorageObjectOrIOException
 
create(IOException) - Static method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.StorageObjectOrIOException
 
create(Coder<InputT>) - Static method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.ApproximateDistinctFn
Returns an ApproximateDistinct.ApproximateDistinctFn combiner with the given input coder.
create(Coder<InputT>) - Static method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.CountMinSketchFn
Returns a SketchFrequencies.CountMinSketchFn combiner with the given input coder.
create(double) - Static method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.TDigestQuantilesFn
Returns TDigestQuantiles.TDigestQuantilesFn combiner with the given compression factor.
create(BufferedExternalSorter.Options) - Static method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter
 
create(ExternalSorter.Options) - Static method in class org.apache.beam.sdk.extensions.sorter.ExternalSorter
Returns a Sorter configured with the given ExternalSorter.Options.
create(BufferedExternalSorter.Options) - Static method in class org.apache.beam.sdk.extensions.sorter.SortValues
Returns a SortValues<PrimaryKeyT, SecondaryKeyT, ValueT> PTransform.
create(double, double, double) - Static method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
 
create(double) - Static method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
It creates an instance with rate=0 and window=rowCount for bounded sources.
create(Class<?>, String) - Static method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
Creates Function from given class.
create(Method) - Static method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
Creates Function from given method.
create(List<String>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.TableName
Full table name with path.
create(List<String>, String) - Static method in class org.apache.beam.sdk.extensions.sql.impl.TableName
Table name plus the path up to but not including table name.
create(String, Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamBuiltinAggregations
 
create(DataCatalogPipelineOptions) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
 
create() - Static method in class org.apache.beam.sdk.fn.data.CompletableFutureInboundDataClient
Create a new CompletableFutureInboundDataClient using a new CompletableFuture.
create(StreamObserver<ReqT>, Runnable) - Static method in class org.apache.beam.sdk.fn.stream.ForwardingClientResponseObserver
 
create() - Static method in class org.apache.beam.sdk.fn.test.InProcessManagedChannelFactory
 
create(int, Duration) - Static method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.RetryConfiguration
 
create(PipelineOptions) - Method in class org.apache.beam.sdk.io.aws.options.AwsOptions.AwsUserCredentialsFactory
 
create(PipelineOptions) - Method in class org.apache.beam.sdk.io.aws.options.AwsOptions.ClientConfigurationFactory
 
create(PipelineOptions) - Method in class org.apache.beam.sdk.io.aws.options.S3Options.S3UploadBufferSizeBytesFactory
 
create(int, Duration) - Static method in class org.apache.beam.sdk.io.aws.sns.SnsIO.RetryConfiguration
 
create(PipelineOptions) - Method in class org.apache.beam.sdk.io.aws2.options.AwsOptions.AwsUserCredentialsFactory
 
create(int, Duration) - Static method in class org.apache.beam.sdk.io.aws2.sns.SnsIO.RetryConfiguration
 
create(String[], String, String) - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
Creates a new Elasticsearch connection configuration.
create(int, Duration) - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.RetryConfiguration
Creates RetryConfiguration for ElasticsearchIO with provided maxAttempts, maxDurations and exponential backoff based retries.
create(WritableByteChannel) - Method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
Deprecated.
 
create(WritableByteChannel) - Method in interface org.apache.beam.sdk.io.FileBasedSink.WritableByteChannelFactory
 
create(EmptyMatchTreatment) - Static method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
create(ResourceIdT, CreateOptions) - Method in class org.apache.beam.sdk.io.FileSystem
Returns a write channel for the given ResourceIdT.
create(ResourceId, String) - Static method in class org.apache.beam.sdk.io.FileSystems
Returns a write channel for the given ResourceId.
create(ResourceId, CreateOptions) - Static method in class org.apache.beam.sdk.io.FileSystems
Returns a write channel for the given ResourceId with CreateOptions.
create(MatchResult.Status, List<MatchResult.Metadata>) - Static method in class org.apache.beam.sdk.io.fs.MatchResult
create(MatchResult.Status, IOException) - Static method in class org.apache.beam.sdk.io.fs.MatchResult
Returns a MatchResult given the MatchResult.Status and IOException.
create(String, ValueProvider<String>, Boolean, Boolean, BigQueryIO.TypedRead.QueryPriority, String, String, String, SerializableFunction<SchemaAndRecord, T>, Coder<T>, BigQueryServices) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageQuerySource
 
create(Storage.ReadSession, Storage.Stream, TableSchema, SerializableFunction<SchemaAndRecord, T>, Coder<T>, BigQueryServices) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource
 
create(ValueProvider<TableReference>, ReadOptions.TableReadOptions, ValueProvider<List<String>>, ValueProvider<String>, SerializableFunction<SchemaAndRecord, T>, Coder<T>, BigQueryServices) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageTableSource
 
create(Schema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery
Creates an instance of this rule.
create(long) - Static method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteResult
 
create() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
Creates an instance of this rule.
create() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal
Creates an instance of this rule.
create(Mutation, Mutation...) - Static method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
Creates a new group.
create(Mutation, Iterable<Mutation>) - Static method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
 
create() - Static method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
create() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
create(BatchTransactionId) - Static method in class org.apache.beam.sdk.io.gcp.spanner.Transaction
 
create(String, String, String, Boolean) - Static method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher.TableAndQuery
 
create(Map<String, String>) - Static method in class org.apache.beam.sdk.io.hcatalog.HCatalogBeamSchema
Create the schema adapter.
create(PipelineOptions) - Method in class org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptions.ConfigurationLocator
 
create(DataSource) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
 
create(String, String) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
 
create(ValueProvider<String>, ValueProvider<String>) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
 
create() - Static method in class org.apache.beam.sdk.io.kinesis.WatermarkParameters
 
create() - Static method in class org.apache.beam.sdk.io.mongodb.AggregationQuery
 
create() - Static method in class org.apache.beam.sdk.io.mongodb.FindQuery
 
create(String, String) - Static method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
Describe a connection configuration to the MQTT broker.
create(String, String, String) - Static method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
Deprecated.
This constructor will be removed in a future version of Beam, please use #create(String, String)} and MqttIO.ConnectionConfiguration.withClientId(String) instead.
create() - Static method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
 
create(String, int) - Static method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
 
create(String) - Static method in class org.apache.beam.sdk.io.solr.SolrIO.ConnectionConfiguration
Creates a new Solr connection configuration.
create(int, Duration) - Static method in class org.apache.beam.sdk.io.solr.SolrIO.RetryConfiguration
 
create(long, long, long, long) - Static method in class org.apache.beam.sdk.metrics.DistributionResult
 
create(long, Instant) - Static method in class org.apache.beam.sdk.metrics.GaugeResult
 
create(String, MetricName) - Static method in class org.apache.beam.sdk.metrics.MetricKey
 
create(Iterable<MetricResult<Long>>, Iterable<MetricResult<DistributionResult>>, Iterable<MetricResult<GaugeResult>>) - Static method in class org.apache.beam.sdk.metrics.MetricQueryResults
 
create(MetricKey, Boolean, T) - Static method in class org.apache.beam.sdk.metrics.MetricResult
 
create(MetricKey, T, T) - Static method in class org.apache.beam.sdk.metrics.MetricResult
 
create(PipelineOptions) - Method in class org.apache.beam.sdk.metrics.MetricsOptions.NoOpMetricsSink
 
create(PipelineOptions) - Method in interface org.apache.beam.sdk.options.DefaultValueFactory
Creates a default value for a getter marked with Default.InstanceFactory.
create(PipelineOptions) - Method in class org.apache.beam.sdk.options.PipelineOptions.AtomicLongFactory
 
create(PipelineOptions) - Method in class org.apache.beam.sdk.options.PipelineOptions.DirectRunner
 
create(PipelineOptions) - Method in class org.apache.beam.sdk.options.PipelineOptions.JobNameFactory
 
create(PipelineOptions) - Method in class org.apache.beam.sdk.options.PipelineOptions.UserAgentFactory
 
create() - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
Creates and returns an object that implements PipelineOptions using the values configured on this builder during construction.
create() - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
Creates and returns an object that implements PipelineOptions.
create() - Static method in class org.apache.beam.sdk.Pipeline
Constructs a pipeline from default PipelineOptions.
create(PipelineOptions) - Static method in class org.apache.beam.sdk.Pipeline
Constructs a pipeline from the provided PipelineOptions.
create() - Static method in class org.apache.beam.sdk.PipelineRunner
Creates a runner from the default app PipelineOptions.
create(Class<?>, Schema) - Method in interface org.apache.beam.sdk.schemas.Factory
 
create() - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return an empty FieldAccessDescriptor.
create(Map<String, Integer>) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
Create an enumeration type over a set of String->Integer values.
create(List<String>) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
Create an enumeration type from a fixed set of String values; integer values will be automatically chosen.
create(String...) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
Create an enumeration type from a fixed set of String values; integer values will be automatically chosen.
create(Schema.Field...) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
Create an OneOfType logical type.
create(List<Schema.Field>) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
Create an OneOfType logical type.
create(List<Schema.Field>, Map<String, Integer>) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
Create an OneOfType logical type.
create(Object...) - Method in interface org.apache.beam.sdk.schemas.SchemaUserTypeCreator
 
create() - Static method in class org.apache.beam.sdk.schemas.transforms.AddFields
 
create(List<String>, String) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast.CompatibilityError
 
create() - Static method in class org.apache.beam.sdk.schemas.transforms.Filter
 
create() - Static method in class org.apache.beam.sdk.schemas.transforms.RenameFields
Create an instance of this transform.
create() - Static method in class org.apache.beam.sdk.schemas.transforms.Select
 
create(Class, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils.ClassWithSchema
 
create(List<String>, Optional<Schema.TypeName>) - Static method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold.Context
 
create() - Static method in class org.apache.beam.sdk.testing.TestPipeline
Creates and returns a new test pipeline.
create(PipelineOptions) - Method in class org.apache.beam.sdk.testing.TestPipelineOptions.AlwaysPassMatcherFactory
 
create(Coder<T>) - Static method in class org.apache.beam.sdk.testing.TestStream
Create a new TestStream.Builder with no elements and watermark equal to BoundedWindow.TIMESTAMP_MIN_VALUE.
create(Schema) - Static method in class org.apache.beam.sdk.testing.TestStream
 
create(Schema, TypeDescriptor<T>, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Static method in class org.apache.beam.sdk.testing.TestStream
 
create(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
Returns an approximate quantiles combiner with the given compareFn and desired number of quantiles.
create(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
Like ApproximateQuantiles.ApproximateQuantilesCombineFn.create(int, Comparator), but sorts values using their natural ordering.
create(int, ComparatorT, long, double) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
Creates an approximate quantiles combiner with the given compareFn and desired number of quantiles.
Create<T> - Class in org.apache.beam.sdk.transforms
Create<T> takes a collection of elements of type T known when the pipeline is constructed and returns a PCollection<T> containing the elements.
Create() - Constructor for class org.apache.beam.sdk.transforms.Create
 
create() - Static method in class org.apache.beam.sdk.transforms.Distinct
Returns a Distinct<T> PTransform.
create() - Static method in class org.apache.beam.sdk.transforms.DoFnSchemaInformation
Create an instance.
create() - Static method in class org.apache.beam.sdk.transforms.GroupByKey
Returns a GroupByKey<K, V> PTransform.
create() - Static method in class org.apache.beam.sdk.transforms.Impulse
Create a new Impulse PTransform.
create() - Static method in class org.apache.beam.sdk.transforms.join.CoGroupByKey
Returns a CoGroupByKey<K> PTransform.
create() - Static method in class org.apache.beam.sdk.transforms.Keys
Returns a Keys<K> PTransform.
create() - Static method in class org.apache.beam.sdk.transforms.KvSwap
Returns a KvSwap<K, V> PTransform.
create() - Static method in class org.apache.beam.sdk.transforms.Values
Returns a Values<V> PTransform.
Create.OfValueProvider<T> - Class in org.apache.beam.sdk.transforms
Create.TimestampedValues<T> - Class in org.apache.beam.sdk.transforms
A PTransform that creates a PCollection whose elements have associated timestamps.
Create.Values<T> - Class in org.apache.beam.sdk.transforms
A PTransform that creates a PCollection from a set of in-memory objects.
createAccumulator() - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.ApproximateDistinctFn
 
createAccumulator() - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.CountMinSketchFn
 
createAccumulator() - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.TDigestQuantilesFn
 
createAccumulator() - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CovarianceFn
 
createAccumulator() - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.VarianceFn
 
createAccumulator() - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
 
createAccumulator() - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn
 
createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
 
createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
 
createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
 
createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
 
createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
Returns a new, mutable accumulator value, representing the accumulation of zero input values.
createAccumulator() - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
 
createAccumulator() - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
 
createAccumulator(CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
 
createAccumulator(CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
Returns a new, mutable accumulator value, representing the accumulation of zero input values.
createAccumulator() - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
 
createAccumulator() - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
 
createAll(Class<?>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
Creates Function for each method in a given class.
createArrayOf(String, Object[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
createArtifactServerFactory(JobServerDriver.ServerConfiguration) - Static method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver
 
createBatchExecutionEnvironment(FlinkPipelineOptions, List<String>) - Static method in class org.apache.beam.runners.flink.FlinkExecutionEnvironments
If the submitted job is a batch processing job, this method creates the adequate Flink ExecutionEnvironment depending on the user-specified options.
createBlob() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
createBoundedTableStatistics(Double) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
 
createBucket(String, Bucket) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
Creates a Bucket under the specified project in Cloud Storage or propagates an exception.
createBuilder(S3Options) - Method in interface org.apache.beam.sdk.io.aws.options.S3ClientBuilderFactory
 
createBuilder(S3Options) - Method in class org.apache.beam.sdk.io.aws.s3.DefaultS3ClientBuilderFactory
 
createClob() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
createCombineFn(AggregateCall, Schema.Field, String) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.AggregationCombineFnAdapter
Creates either a UDAF or a built-in Combine.CombineFn.
createConstantCombineFn() - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.AggregationCombineFnAdapter
 
createConstructorCreator(Class<T>, Constructor<T>, Schema, List<FieldValueTypeInformation>, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
 
createConstructorCreator(Class<T>, Constructor<T>, Schema, List<FieldValueTypeInformation>, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
 
createContextual(DeserializationContext, BeanProperty) - Method in class org.apache.beam.sdk.options.ValueProvider.Deserializer
 
CreateDataflowView<ElemT,ViewT> - Class in org.apache.beam.runners.dataflow
A DataflowRunner marker class for creating a PCollectionView.
createDataset(String, String, String, String, Long) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
Create a Dataset with the given location, description and default expiration time for tables in the dataset (if null, tables don't expire).
createDataset(String, String, String, String, Long) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
createDecompressingChannel(ReadableByteChannel) - Method in enum org.apache.beam.sdk.io.CompressedSource.CompressionMode
Deprecated.
 
createDecompressingChannel(ReadableByteChannel) - Method in interface org.apache.beam.sdk.io.CompressedSource.DecompressingChannelFactory
Given a channel, create a channel that decompresses the content read from the channel.
createDefault() - Static method in class org.apache.beam.runners.fnexecution.ServerFactory
createDefault() - Static method in interface org.apache.beam.runners.fnexecution.ServerFactory.UrlFactory
 
createDefault() - Static method in class org.apache.beam.sdk.coders.CoderRegistry
Creates a CoderRegistry containing registrations for all standard coders part of the core Java Apache Beam SDK and also any registrations provided by coder registrars.
createDefault() - Static method in class org.apache.beam.sdk.fn.channel.ManagedChannelFactory
 
createDefault() - Static method in class org.apache.beam.sdk.schemas.SchemaRegistry
 
createDynamoDB() - Method in interface org.apache.beam.sdk.io.aws.dynamodb.AwsClientsProvider
 
createDynamoDB() - Method in class org.apache.beam.sdk.io.aws.dynamodb.BasicDynamoDBProvider
 
createEnvironment(RunnerApi.Environment) - Method in class org.apache.beam.runners.fnexecution.environment.DockerEnvironmentFactory
Creates a new, active RemoteEnvironment backed by a local Docker container.
createEnvironment(RunnerApi.Environment) - Method in class org.apache.beam.runners.fnexecution.environment.EmbeddedEnvironmentFactory
 
createEnvironment(RunnerApi.Environment) - Method in interface org.apache.beam.runners.fnexecution.environment.EnvironmentFactory
Creates an active RunnerApi.Environment and returns a handle to it.
createEnvironment(RunnerApi.Environment) - Method in class org.apache.beam.runners.fnexecution.environment.ExternalEnvironmentFactory
Creates a new, active RemoteEnvironment backed by an unmanaged worker.
createEnvironment(RunnerApi.Environment) - Method in class org.apache.beam.runners.fnexecution.environment.ProcessEnvironmentFactory
Creates a new, active RemoteEnvironment backed by a forked process.
createEnvironment(RunnerApi.Environment) - Method in class org.apache.beam.runners.fnexecution.environment.StaticRemoteEnvironmentFactory
 
createEnvironmentFactory(GrpcFnServer<FnApiControlClientPoolService>, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<LegacyArtifactRetrievalService>, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool, IdGenerator) - Method in class org.apache.beam.runners.fnexecution.environment.DockerEnvironmentFactory.Provider
 
createEnvironmentFactory(GrpcFnServer<FnApiControlClientPoolService>, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<LegacyArtifactRetrievalService>, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool, IdGenerator) - Method in class org.apache.beam.runners.fnexecution.environment.EmbeddedEnvironmentFactory.Provider
 
createEnvironmentFactory(GrpcFnServer<FnApiControlClientPoolService>, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<LegacyArtifactRetrievalService>, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool, IdGenerator) - Method in interface org.apache.beam.runners.fnexecution.environment.EnvironmentFactory.Provider
Creates EnvironmentFactory for the provided GrpcServices.
createEnvironmentFactory(GrpcFnServer<FnApiControlClientPoolService>, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<LegacyArtifactRetrievalService>, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool, IdGenerator) - Method in class org.apache.beam.runners.fnexecution.environment.ExternalEnvironmentFactory.Provider
 
createEnvironmentFactory(GrpcFnServer<FnApiControlClientPoolService>, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<LegacyArtifactRetrievalService>, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool, IdGenerator) - Method in class org.apache.beam.runners.fnexecution.environment.ProcessEnvironmentFactory.Provider
 
createEnvironmentFactory(GrpcFnServer<FnApiControlClientPoolService>, GrpcFnServer<GrpcLoggingService>, GrpcFnServer<LegacyArtifactRetrievalService>, GrpcFnServer<StaticGrpcProvisionService>, ControlClientPool, IdGenerator) - Method in class org.apache.beam.runners.fnexecution.environment.StaticRemoteEnvironmentFactory.Provider
 
createEpoll() - Static method in class org.apache.beam.sdk.fn.channel.ManagedChannelFactory
 
createEpollDomainSocket() - Static method in class org.apache.beam.runners.fnexecution.ServerFactory
Create a EpollDomainSocket.
createEpollSocket() - Static method in class org.apache.beam.runners.fnexecution.ServerFactory
Create a EpollSocket.
createFactory() - Method in class org.apache.beam.sdk.extensions.sql.impl.JdbcDriver
 
createFactoryForCreateSubscription() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
createFactoryForPublish(PubsubClient.TopicPath, Iterable<PubsubClient.OutgoingMessage>, Iterable<PubsubClient.OutgoingMessage>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
Return a factory for testing publishers.
createFactoryForPull(Clock, PubsubClient.SubscriptionPath, int, Iterable<PubsubClient.IncomingMessage>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
Return a factory for testing subscribers.
createForSubrangeOfFile(String, long, long) - Method in class org.apache.beam.sdk.io.AvroSource
Deprecated.
Used by Dataflow worker
createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.AvroSource
 
createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.BlockBasedSource
Creates a BlockBasedSource for the specified range in a single file.
createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.CompressedSource
Creates a CompressedSource for a subrange of a file.
createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.FileBasedSource
Creates and returns a new FileBasedSource of the same type as the current FileBasedSource backed by a given file and an offset range.
createForSubrangeOfFile(MatchResult.Metadata, long, long) - Method in class org.apache.beam.sdk.io.xml.XmlSource
 
createFrom(String) - Static method in class org.apache.beam.sdk.fn.channel.SocketAddressFactory
Parse a SocketAddress from the given string.
CreateGearpumpPCollectionViewTranslator<ElemT,ViewT> - Class in org.apache.beam.runners.gearpump.translators
CreateGearpumpPCollectionView bridges input stream to down stream transforms.
CreateGearpumpPCollectionViewTranslator() - Constructor for class org.apache.beam.runners.gearpump.translators.CreateGearpumpPCollectionViewTranslator
 
createGetter(FieldValueTypeInformation, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
 
createGetterConversions(StackManipulation) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.DefaultTypeConversionsFactory
 
createGetterConversions(StackManipulation) - Method in interface org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversionsFactory
 
createInput(Pipeline, Map<String, PCollection<?>>) - Method in interface org.apache.beam.sdk.expansion.service.ExpansionService.TransformProvider
 
createInputFormatInstance() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
Creates instance of InputFormat class.
createJar(File, File) - Static method in class org.apache.beam.runners.apex.ApexYarnLauncher
Create a jar file from the given directory.
createJob(Job) - Method in class org.apache.beam.runners.dataflow.DataflowClient
Creates the Dataflow Job.
createJobInvocation(String, String, ListeningExecutorService, RunnerApi.Pipeline, FlinkPipelineOptions, PortablePipelineRunner) - Method in class org.apache.beam.runners.flink.FlinkJobInvoker
 
createJobServerFactory(JobServerDriver.ServerConfiguration) - Static method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver
 
createJobService() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver
 
createKinesisProducer(KinesisProducerConfiguration) - Method in interface org.apache.beam.sdk.io.kinesis.AWSClientsProvider
 
createNClob() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
createNewDataset(String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
Creates a new dataset.
createNewTable(String, String, Table) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
 
CreateOptions - Class in org.apache.beam.sdk.io.fs
An abstract class that contains common configuration options for creating resources.
CreateOptions() - Constructor for class org.apache.beam.sdk.io.fs.CreateOptions
 
CreateOptions.Builder<BuilderT extends CreateOptions.Builder<BuilderT>> - Class in org.apache.beam.sdk.io.fs
An abstract builder for CreateOptions.
CreateOptions.StandardCreateOptions - Class in org.apache.beam.sdk.io.fs
A standard configuration options with builder.
CreateOptions.StandardCreateOptions.Builder - Class in org.apache.beam.sdk.io.fs
createOutputMap(Iterable<String>) - Static method in class org.apache.beam.runners.fnexecution.translation.PipelineTranslatorUtils
Creates a mapping from PCollection id to output tag integer.
createPane(boolean, boolean, PaneInfo.Timing) - Static method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
 
createPane(boolean, boolean, PaneInfo.Timing, long, long) - Static method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
Factory method to create a PaneInfo with the specified parameters.
createPipelineOptions(Map<String, String>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter
 
createPrepareContext() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
createPrimitiveOutputInternal(Pipeline, WindowingStrategy<?, ?>, PCollection.IsBounded, Coder<T>) - Static method in class org.apache.beam.sdk.values.PCollection
For internal use only; no backwards-compatibility guarantees.
createPrimitiveOutputInternal(Pipeline, WindowingStrategy<?, ?>, PCollection.IsBounded, Coder<T>, TupleTag<?>) - Static method in class org.apache.beam.sdk.values.PCollection
For internal use only; no backwards-compatibility guarantees.
createProcessContext(ValueInSingleWindow<InputT>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
createPushDownRel(RelDataType, List<String>, BeamSqlTableFilter) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
 
createQuery(Expression, Class<T>) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
createQuery(Expression, Type) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
createQuery(String, String, String) - Static method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher
 
createQueryUsingStandardSql(String, String, String) - Static method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher
 
createRandomSubscription(PubsubClient.ProjectPath, PubsubClient.TopicPath, int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
Create a random subscription for topic.
createReader(PipelineOptions) - Method in class org.apache.beam.runners.gearpump.translators.io.BoundedSourceWrapper
 
createReader(PipelineOptions) - Method in class org.apache.beam.runners.gearpump.translators.io.GearpumpSource
 
createReader(PipelineOptions) - Method in class org.apache.beam.runners.gearpump.translators.io.UnboundedSourceWrapper
 
createReader(PipelineOptions, UnboundedSource.CheckpointMark) - Method in class org.apache.beam.runners.gearpump.translators.io.ValuesSource
 
createReader(DataSourceOptions) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.DatasetSourceBatch
 
createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.BoundedSource
Returns a new BoundedSource.BoundedReader that reads from this source.
createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
 
createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
 
createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource
 
createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
 
createReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
 
createReader(PipelineOptions, CheckpointMarkT) - Method in class org.apache.beam.sdk.io.UnboundedSource
Create a new UnboundedSource.UnboundedReader to read from this source, resuming from the given checkpoint if present.
createReadSession(Storage.CreateReadSessionRequest) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StorageClient
Create a new read session against an existing table.
createRunner(ReadyCheckingSideInputReader) - Method in class org.apache.beam.runners.gearpump.translators.utils.DoFnRunnerFactory
 
createSessionToken(String) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver
 
createSetter(FieldValueTypeInformation, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
 
createSetterConversions(StackManipulation) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.DefaultTypeConversionsFactory
 
createSetterConversions(StackManipulation) - Method in interface org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversionsFactory
 
createSimpleSqlFunction(String, SqlTypeName) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.AvroSource
 
createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.BlockBasedSource
Creates a BlockBasedReader.
createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.CompressedSource
Creates a FileBasedReader to read a single file.
createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
Creates and returns an instance of a FileBasedReader implementation for the current source assuming the source represents a single file.
createSingleFileReader(PipelineOptions) - Method in class org.apache.beam.sdk.io.xml.XmlSource
 
createSnsPublisher() - Method in interface org.apache.beam.sdk.io.aws.sns.AwsClientsProvider
 
createSourceForSubrange(long, long) - Method in class org.apache.beam.sdk.io.FileBasedSource
 
createSourceForSubrange(long, long) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
Returns an OffsetBasedSource for a subrange of the current source.
createSQLXML() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
createStateBackend(FlinkPipelineOptions) - Method in interface org.apache.beam.runners.flink.FlinkStateBackendFactory
 
createStatement() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
createStatement(int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
createStatement(int, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
createStaticCreator(Class<T>, Method, Schema, List<FieldValueTypeInformation>, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
 
createStaticCreator(Class<T>, Method, Schema, List<FieldValueTypeInformation>, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
 
CreateStream<T> - Class in org.apache.beam.runners.spark.io
Create an input stream from Queue.
createStreamExecutionEnvironment(FlinkPipelineOptions, List<String>) - Static method in class org.apache.beam.runners.flink.FlinkExecutionEnvironments
If the submitted job is a stream processing job, this method creates the adequate Flink StreamExecutionEnvironment depending on the user-specified options.
createStruct(String, Object[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
createStructuralValues(Coder<T>, List<T>) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
Testing utilities below depend on standard assertions and matchers to compare elements read by sources.
createSubscription(PubsubClient.TopicPath, PubsubClient.SubscriptionPath, int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
Create subscription to topic.
createSubscription(PubsubClient.TopicPath, PubsubClient.SubscriptionPath, int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
 
createSubscription(PubsubClient.TopicPath, PubsubClient.SubscriptionPath, int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
 
createSubscription(PubsubClient.TopicPath, PubsubClient.SubscriptionPath, int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
createTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
 
createTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.InMemoryMetaTableProvider
 
createTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.ReadOnlyTableProvider
 
createTable(Table) - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.TableProvider
Creates a table.
createTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
 
createTable(Table) - Method in class org.apache.beam.sdk.extensions.sql.meta.store.InMemoryMetaStore
 
createTable(Table) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
Creates the specified table if it does not exist.
createTable(Table) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
CreateTables<DestinationT,ElementT> - Class in org.apache.beam.sdk.io.gcp.bigquery
Creates any tables needed before performing streaming writes to the tables.
CreateTables(BigQueryIO.Write.CreateDisposition, DynamicDestinations<?, DestinationT>) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.CreateTables
 
createTimestampPolicy(TopicPartition, Optional<Instant>) - Method in interface org.apache.beam.sdk.io.kafka.TimestampPolicyFactory
Creates a TimestampPolicy for a partition.
createTopic(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
Create topic.
createTopic(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
 
createTopic(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
 
createTopic(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
createTransaction() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO
Returns a transform that creates a batch transaction.
CreateTransaction() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
 
createTranslationContext(JobInfo, FlinkPipelineOptions, String, List<String>) - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator
Creates a batch translation context.
createTranslationContext(JobInfo, FlinkPipelineOptions, String, List<String>) - Method in interface org.apache.beam.runners.flink.FlinkPortablePipelineTranslator
 
createTranslationContext(JobInfo, FlinkPipelineOptions, String, List<String>) - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator
Creates a streaming translation context.
createTranslator() - Static method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator
Creates a batch translator.
createTypeConversion(boolean) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.DefaultTypeConversionsFactory
 
createTypeConversion(boolean) - Method in interface org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversionsFactory
 
createUdfOperator(String, Class<?>, String, SqlReturnTypeInference, List<RelDataType>) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
createUdfOperator(String, Method) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
createUdfOperator(String, Method, SqlSyntax) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
createUnboundedTableStatistics(Double) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
 
createUrl(String, int) - Method in interface org.apache.beam.runners.fnexecution.ServerFactory.UrlFactory
 
createValue(String, T) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
Create a OneOfType.Value specifying which field to set and the value to set.
createValue(int, T) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
Create a OneOfType.Value specifying which field to set and the value to set.
createValue(EnumerationType.Value, T) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
Create a OneOfType.Value specifying which field to set and the value to set.
createWatermarkPolicy() - Method in interface org.apache.beam.sdk.io.kinesis.WatermarkPolicyFactory
 
createWithPortSupplier(Supplier<Integer>) - Static method in class org.apache.beam.runners.fnexecution.ServerFactory
Create a ServerFactory.InetSocketAddressServerFactory that uses ports from a supplier.
createWithUrlFactory(ServerFactory.UrlFactory) - Static method in class org.apache.beam.runners.fnexecution.ServerFactory
Create a ServerFactory.InetSocketAddressServerFactory that uses the given url factory.
createWithUrlFactoryAndPortSupplier(ServerFactory.UrlFactory, Supplier<Integer>) - Static method in class org.apache.beam.runners.fnexecution.ServerFactory
Create a ServerFactory.InetSocketAddressServerFactory that uses the given url factory and ports from a supplier.
createWriteOperation() - Method in class org.apache.beam.sdk.io.FileBasedSink
Return a subclass of FileBasedSink.WriteOperation that will manage the write to the sink.
createWriter() - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
Clients must implement to return a subclass of FileBasedSink.Writer.
createZetaSqlStructTypeFromBeamSchema(Schema) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlUtils
 
CredentialFactory - Interface in org.apache.beam.sdk.extensions.gcp.auth
Construct an oauth credential to be used by the SDK and the SDK workers.
crossProductJoin() - Method in class org.apache.beam.sdk.schemas.transforms.CoGroup.Impl
Expand the join into individual rows, similar to SQL joins.
csvLines2BeamRows(CSVFormat, String, Schema) - Static method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamTableUtils
Decode zero or more CSV records from the given string, according to the specified CSVFormat, and converts them to Rows with the specified Schema.
CsvRecorderDecoder(Schema, CSVFormat) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaCSVTable.CsvRecorderDecoder
 
CsvRecorderEncoder(Schema, CSVFormat) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaCSVTable.CsvRecorderEncoder
 
CsvSink - Class in org.apache.beam.runners.spark.metrics.sink
A Spark Sink that is tailored to report AggregatorMetric metrics to a CSV file.
CsvSink(Properties, MetricRegistry, SecurityManager) - Constructor for class org.apache.beam.runners.spark.metrics.sink.CsvSink
 
CsvToRow(Schema, CSVFormat) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTableProvider.CsvToRow
 
ctxt - Variable in class org.apache.beam.runners.spark.SparkRunner.Evaluator
 
current() - Method in interface org.apache.beam.runners.spark.aggregators.NamedAggregators.State
 
current() - Method in interface org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregators.State
 
currentEventTime() - Method in interface org.apache.beam.sdk.state.Timers
Returns the current event time.
currentInputWatermarkTime() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
 
currentOutputWatermarkTime() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
 
currentProcessingTime() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
 
currentProcessingTime() - Method in interface org.apache.beam.sdk.state.Timers
Returns the current processing time.
currentRestriction() - Method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
 
currentRestriction() - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
 
currentRestriction() - Method in class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker
Returns a restriction accurately describing the full range of work the current DoFn.ProcessElement call will do, including already completed work.
currentSynchronizedProcessingTime() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
 
currentSynchronizedProcessingTime() - Method in interface org.apache.beam.sdk.state.Timers
Returns the current synchronized processing time or null if unknown.
currentWatermark - Variable in class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.LogAppendTimePolicy
 
currentWatermark() - Method in interface org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimator
Return estimated output watermark.
currentWatermark() - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.Manual
 
currentWatermark() - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.MonotonicallyIncreasing
 
currentWatermark() - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.WallTime
 
CUSTOM_SOURCE_FORMAT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
CustomCoder<T> - Class in org.apache.beam.sdk.coders
An abstract base class that implements all methods of Coder except Coder.encode(T, java.io.OutputStream) and Coder.decode(java.io.InputStream).
CustomCoder() - Constructor for class org.apache.beam.sdk.coders.CustomCoder
 
Customer - Class in org.apache.beam.sdk.extensions.sql.example.model
Describes a customer.
Customer(int, String, String) - Constructor for class org.apache.beam.sdk.extensions.sql.example.model.Customer
 
Customer() - Constructor for class org.apache.beam.sdk.extensions.sql.example.model.Customer
 
CustomHttpErrors - Class in org.apache.beam.sdk.extensions.gcp.util
An optional component to use with the RetryHttpRequestInitializer in order to provide custom errors for failing http calls.
CustomHttpErrors.Builder - Class in org.apache.beam.sdk.extensions.gcp.util
A Builder which allows building immutable CustomHttpErrors object.
CustomHttpErrors.MatcherAndError - Class in org.apache.beam.sdk.extensions.gcp.util
A simple Tuple class for creating a list of HttpResponseMatcher and HttpResponseCustomError to print for the responses.
CustomTableResolver - Interface in org.apache.beam.sdk.extensions.sql.meta
Interface that table providers can implement if they require custom table name resolution.
CustomTimestampPolicyWithLimitedDelay<K,V> - Class in org.apache.beam.sdk.io.kafka
A policy for custom record timestamps where timestamps within a partition are expected to be roughly monotonically increasing with a cap on out of order event delays (say 1 minute).
CustomTimestampPolicyWithLimitedDelay(SerializableFunction<KafkaRecord<K, V>, Instant>, Duration, Optional<Instant>) - Constructor for class org.apache.beam.sdk.io.kafka.CustomTimestampPolicyWithLimitedDelay
A policy for custom record timestamps where timestamps are expected to be roughly monotonically increasing with out of order event delays less than maxDelay.

D

DAGBuilder - Class in org.apache.beam.runners.jet
Utility class for wiring up Jet DAGs based on Beam pipelines.
DAGBuilder.WiringListener - Interface in org.apache.beam.runners.jet
Listener that can be registered with a DAGBuilder in order to be notified when edges are being registered.
data(StreamObserver<BeamFnApi.Elements>) - Method in class org.apache.beam.runners.fnexecution.data.GrpcDataService
 
data(String, String) - Static method in class org.apache.beam.sdk.fn.data.LogicalEndpoint
 
DATA_BUFFER_SIZE_LIMIT - Static variable in interface org.apache.beam.sdk.fn.data.BeamFnDataBufferingOutboundObserver
 
DATA_BUFFER_TIME_LIMIT_MS - Static variable in interface org.apache.beam.sdk.fn.data.BeamFnDataBufferingOutboundObserver
 
database() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
 
DataCatalogPipelineOptions - Interface in org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog
Pipeline options for Data Catalog table provider.
DataCatalogPipelineOptionsRegistrar - Class in org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog
 
DataCatalogPipelineOptionsRegistrar() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogPipelineOptionsRegistrar
 
DataCatalogTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog
Uses DataCatalog to get the source type and schema for a table.
DataflowClient - Class in org.apache.beam.runners.dataflow
Wrapper around the generated Dataflow client to provide common functionality.
DataflowClientFactory() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions.DataflowClientFactory
 
DataflowJobAlreadyExistsException - Exception in org.apache.beam.runners.dataflow
An exception that is thrown if the unique job name constraint of the Dataflow service is broken because an existing job with the same job name is currently active.
DataflowJobAlreadyExistsException(DataflowPipelineJob, String) - Constructor for exception org.apache.beam.runners.dataflow.DataflowJobAlreadyExistsException
Create a new DataflowJobAlreadyExistsException with the specified DataflowPipelineJob and message.
DataflowJobAlreadyUpdatedException - Exception in org.apache.beam.runners.dataflow
An exception that is thrown if the existing job has already been updated within the Dataflow service and is no longer able to be updated.
DataflowJobAlreadyUpdatedException(DataflowPipelineJob, String) - Constructor for exception org.apache.beam.runners.dataflow.DataflowJobAlreadyUpdatedException
Create a new DataflowJobAlreadyUpdatedException with the specified DataflowPipelineJob and message.
DataflowJobException - Exception in org.apache.beam.runners.dataflow
A RuntimeException that contains information about a DataflowPipelineJob.
DataflowPipelineDebugOptions - Interface in org.apache.beam.runners.dataflow.options
Internal.
DataflowPipelineDebugOptions.DataflowClientFactory - Class in org.apache.beam.runners.dataflow.options
Returns the default Dataflow client built from the passed in PipelineOptions.
DataflowPipelineDebugOptions.StagerFactory - Class in org.apache.beam.runners.dataflow.options
Creates a Stager object using the class specified in DataflowPipelineDebugOptions.getStagerClass().
DataflowPipelineJob - Class in org.apache.beam.runners.dataflow
A DataflowPipelineJob represents a job submitted to Dataflow using DataflowRunner.
DataflowPipelineJob(DataflowClient, String, DataflowPipelineOptions, Map<AppliedPTransform<?, ?, ?>, String>) - Constructor for class org.apache.beam.runners.dataflow.DataflowPipelineJob
Constructs the job.
DataflowPipelineOptions - Interface in org.apache.beam.runners.dataflow.options
Options that can be used to configure the DataflowRunner.
DataflowPipelineOptions.DefaultGcpRegionFactory - Class in org.apache.beam.runners.dataflow.options
Factory for a default value for Google Cloud region according to https://cloud.google.com/compute/docs/gcloud-compute/#default-properties.
DataflowPipelineOptions.FlexResourceSchedulingGoal - Enum in org.apache.beam.runners.dataflow.options
Set of available Flexible Resource Scheduling goals.
DataflowPipelineOptions.StagingLocationFactory - Class in org.apache.beam.runners.dataflow.options
Returns a default staging location under GcpOptions.getGcpTempLocation().
DataflowPipelineRegistrar - Class in org.apache.beam.runners.dataflow
Contains the PipelineOptionsRegistrar and PipelineRunnerRegistrar for the DataflowRunner.
DataflowPipelineRegistrar.Options - Class in org.apache.beam.runners.dataflow
DataflowPipelineRegistrar.Runner - Class in org.apache.beam.runners.dataflow
Register the DataflowRunner.
DataflowPipelineTranslator - Class in org.apache.beam.runners.dataflow
DataflowPipelineTranslator knows how to translate Pipeline objects into Cloud Dataflow Service API Jobs.
DataflowPipelineTranslator.JobSpecification - Class in org.apache.beam.runners.dataflow
The result of a job translation.
DataflowPipelineWorkerPoolOptions - Interface in org.apache.beam.runners.dataflow.options
Options that are used to configure the Dataflow pipeline worker pool.
DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType - Enum in org.apache.beam.runners.dataflow.options
Type of autoscaling algorithm to use.
DataflowPipelineWorkerPoolOptions.WorkerHarnessContainerImageFactory - Class in org.apache.beam.runners.dataflow.options
Returns the default Docker container image that executes Dataflow worker harness, residing in Google Container Registry.
DataflowPortabilityApiUnsupported - Interface in org.apache.beam.sdk.testing
Category tag for the dataflow fn-api worker unsupported tests.
DataflowProfilingAgentConfiguration() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowProfilingOptions.DataflowProfilingAgentConfiguration
 
DataflowProfilingOptions - Interface in org.apache.beam.runners.dataflow.options
Options for controlling profiling of pipeline execution.
DataflowProfilingOptions.DataflowProfilingAgentConfiguration - Class in org.apache.beam.runners.dataflow.options
Configuration the for profiling agent.
DataflowRunner - Class in org.apache.beam.runners.dataflow
A PipelineRunner that executes the operations in the pipeline by first translating them to the Dataflow representation using the DataflowPipelineTranslator and then submitting them to a Dataflow service for execution.
DataflowRunner(DataflowPipelineOptions) - Constructor for class org.apache.beam.runners.dataflow.DataflowRunner
 
DataflowRunner.StreamingPCollectionViewWriterFn<T> - Class in org.apache.beam.runners.dataflow
A marker DoFn for writing the contents of a PCollection to a streaming PCollectionView backend implementation.
DataflowRunnerHooks - Class in org.apache.beam.runners.dataflow
An instance of this class can be passed to the DataflowRunner to add user defined hooks to be invoked at various times during pipeline execution.
DataflowRunnerHooks() - Constructor for class org.apache.beam.runners.dataflow.DataflowRunnerHooks
 
DataflowRunnerInfo - Class in org.apache.beam.runners.dataflow
Populates versioning and other information for DataflowRunner.
DataflowServiceException - Exception in org.apache.beam.runners.dataflow
Signals there was an error retrieving information about a job from the Cloud Dataflow Service.
DataflowTemplateJob - Class in org.apache.beam.runners.dataflow.util
A DataflowPipelineJob that is returned when --templateRunner is set.
DataflowTemplateJob() - Constructor for class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
 
DataflowTransport - Class in org.apache.beam.runners.dataflow.util
Helpers for cloud communication.
DataflowTransport() - Constructor for class org.apache.beam.runners.dataflow.util.DataflowTransport
 
DataflowWorkerHarnessOptions - Interface in org.apache.beam.runners.dataflow.options
Options that are used exclusively within the Dataflow worker harness.
DataflowWorkerLoggingOptions - Interface in org.apache.beam.runners.dataflow.options
Deprecated.
This interface will no longer be the source of truth for worker logging configuration once jobs are executed using a dedicated SDK harness instead of user code being co-located alongside Dataflow worker code. Please set the option below and also the corresponding option within SdkHarnessOptions to ensure forward compatibility.
DataflowWorkerLoggingOptions.Level - Enum in org.apache.beam.runners.dataflow.options
Deprecated.
The set of log levels that can be used on the Dataflow worker.
DataflowWorkerLoggingOptions.WorkerLogLevelOverrides - Class in org.apache.beam.runners.dataflow.options
Deprecated.
Defines a log level override for a specific class, package, or name.
DatasetSourceBatch - Class in org.apache.beam.runners.spark.structuredstreaming.translation.batch
This is a spark structured streaming DataSourceV2 implementation that wraps an BoundedSource.As Continuous streaming is tagged experimental in spark (no aggregation support + no exactly once guaranty), this class does no implement ContinuousReadSupport.
DatasetSourceBatch() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.batch.DatasetSourceBatch
 
DataSourceConfiguration() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
 
DatastoreIO - Class in org.apache.beam.sdk.io.gcp.datastore
DatastoreIO provides an API for reading from and writing to Google Cloud Datastore over different versions of the Cloud Datastore Client libraries.
DatastoreV1 - Class in org.apache.beam.sdk.io.gcp.datastore
DatastoreV1 provides an API to Read, Write and Delete PCollections of Google Cloud Datastore version v1 Entity objects.
DatastoreV1.DeleteEntity - Class in org.apache.beam.sdk.io.gcp.datastore
A PTransform that deletes Entities from Cloud Datastore.
DatastoreV1.DeleteKey - Class in org.apache.beam.sdk.io.gcp.datastore
A PTransform that deletes Entities associated with the given Keys from Cloud Datastore.
DatastoreV1.Read - Class in org.apache.beam.sdk.io.gcp.datastore
A PTransform that reads the result rows of a Cloud Datastore query as Entity objects.
DatastoreV1.Write - Class in org.apache.beam.sdk.io.gcp.datastore
A PTransform that writes Entity objects to Cloud Datastore.
DataStoreV1TableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.datastore
TableProvider for DataStoreV1Table.
DataStoreV1TableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.datastore.DataStoreV1TableProvider
 
DataStreamDecoder(Coder<T>, InputStream) - Constructor for class org.apache.beam.sdk.fn.stream.DataStreams.DataStreamDecoder
 
DataStreams - Class in org.apache.beam.sdk.fn.stream
DataStreams.inbound(Iterator) treats multiple ByteStrings as a single input stream and DataStreams.outbound(OutputChunkConsumer) treats a single OutputStream as multiple ByteStrings.
DataStreams() - Constructor for class org.apache.beam.sdk.fn.stream.DataStreams
 
DataStreams.BlockingQueueIterator<T> - Class in org.apache.beam.sdk.fn.stream
Allows for one or more writing threads to append values to this iterator while one reading thread reads values.
DataStreams.DataStreamDecoder<T> - Class in org.apache.beam.sdk.fn.stream
An adapter which converts an InputStream to an Iterator of T values using the specified Coder.
DataStreams.ElementDelimitedOutputStream - Class in org.apache.beam.sdk.fn.stream
An adapter which wraps an DataStreams.OutputChunkConsumer as an OutputStream.
DataStreams.OutputChunkConsumer<T> - Interface in org.apache.beam.sdk.fn.stream
dataType() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.DecodeUsingBeamCoder
 
dataType() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.EncodeUsingBeamCoder
 
DATE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
date(Integer, Integer, Integer) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.DateFunctions
 
date(DateTime) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.DateFunctions
 
date(DateTime, String) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.DateFunctions
 
DATE - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
DATE_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamBuiltinMethods
 
DATE_OP - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
DATE_TYPES - Static variable in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
DateFunctions - Class in org.apache.beam.sdk.extensions.sql.zetasql
DateFunctions.
DateFunctions() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.DateFunctions
 
DATETIME - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
DATETIME - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
The type of datetime fields.
DateTimeUtils - Class in org.apache.beam.sdk.extensions.sql.zetasql
DateTimeUtils.
DateTimeUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
DateType() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.DateType
 
days(int) - Static method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows
Returns a WindowFn that windows elements into periods measured by days.
dec() - Method in class org.apache.beam.runners.jet.metrics.CounterImpl
 
dec(long) - Method in class org.apache.beam.runners.jet.metrics.CounterImpl
 
dec() - Method in interface org.apache.beam.sdk.metrics.Counter
 
dec(long) - Method in interface org.apache.beam.sdk.metrics.Counter
 
dec() - Method in class org.apache.beam.sdk.metrics.DelegatingCounter
 
dec(long) - Method in class org.apache.beam.sdk.metrics.DelegatingCounter
 
DECIMAL - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
DECIMAL - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
The type of decimal fields.
decode(InputStream) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
 
decode(String) - Static method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService.StagingSessionToken
 
decode(boolean, byte[], Coder<T>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.DecodeUsingBeamCoder
Convert value from byte array (invoked by generated code in EncoderHelpers.DecodeUsingBeamCoder.doGenCode(CodegenContext, ExprCode)).
decode(InputStream) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.MultiOuputCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.AvroCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.BigEndianShortCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.BitSetCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.BitSetCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.BooleanCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.ByteCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.Coder
Decodes a value of type T from the given input stream in the given context.
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.Coder
Deprecated.
only implement and call Coder.decode(InputStream)
decode(InputStream) - Method in class org.apache.beam.sdk.coders.DelegateCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.DelegateCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.DoubleCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.DurationCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.FloatCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.InstantCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.KvCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.KvCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.MapCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.MapCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.NullableCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.NullableCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.SerializableCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.ShardedKeyCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.SnappyCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.VarIntCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.VarLongCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.coders.VoidCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.HyperLogLogPlusCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow.Coder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.amqp.AmqpMessageCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.aws.dynamodb.AttributeValueCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.AttributeValueCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.aws2.sns.PublishResponseCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.aws2.sqs.MessageCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.aws2.sqs.SendMessageRequestCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.ParamsCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.fs.MetadataCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.fs.MetadataCoderV2
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.fs.ResourceIdCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertErrorCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV2
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV3
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteResultCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessagePayloadOnlyCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessagePayloadOnlyCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesAndMessageIdCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithMessageIdCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.hadoop.WritableCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.range.OffsetRange.Coder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.ReadableFileCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.io.xml.JAXBCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.io.xml.JAXBCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.schemas.SchemaCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.testing.TestStream.TestStreamCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
 
decode(InputStream) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo.PaneInfoCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
 
decode(InputStream) - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
 
decode(InputStream) - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
 
decode(InputStream, Coder.Context) - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
 
decodePane(byte) - Static method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
 
decodeQueryResult(String) - Static method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
 
decodeTimerDataTimerId(String) - Static method in class org.apache.beam.runners.fnexecution.control.TimerReceiverFactory
Decodes a string into the transform and timer family ids.
decodeToIterable(List<T>) - Method in class org.apache.beam.sdk.coders.CollectionCoder
Builds an instance of IterableT, this coder's associated Iterable-like subtype, from a list of decoded elements.
decodeToIterable(List<T>) - Method in class org.apache.beam.sdk.coders.IterableCoder
 
decodeToIterable(List<T>) - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
Builds an instance of IterableT, this coder's associated Iterable-like subtype, from a list of decoded elements.
decodeToIterable(List<T>) - Method in class org.apache.beam.sdk.coders.ListCoder
 
decodeToIterable(List<T>) - Method in class org.apache.beam.sdk.coders.SetCoder
Builds an instance of IterableT, this coder's associated Iterable-like subtype, from a list of decoded elements.
DecodeUsingBeamCoder(Expression, ClassTag<T>, Coder<T>) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.DecodeUsingBeamCoder
 
decodeWindowedValue(byte[], Coder) - Static method in class org.apache.beam.runners.jet.Utils
 
decrementingLongs() - Static method in class org.apache.beam.sdk.fn.IdGenerators
Returns an IdGenerators that will provide successive decrementing longs.
Deduplicate - Class in org.apache.beam.sdk.transforms
A set of PTransforms which deduplicate input records over a time domain and threshold.
Deduplicate.KeyedValues<K,V> - Class in org.apache.beam.sdk.transforms
Deduplicates keyed values using the key over a specified time domain and threshold.
Deduplicate.Values<T> - Class in org.apache.beam.sdk.transforms
Deduplicates values over a specified time domain and threshold.
Deduplicate.WithRepresentativeValues<T,IdT> - Class in org.apache.beam.sdk.transforms
A PTransform that uses a SerializableFunction to obtain a representative value for each input element used for deduplication.
deepEquals(Object, Object, Schema.FieldType) - Static method in class org.apache.beam.sdk.values.Row.Equals
 
deepHashCode(Object, Schema.FieldType) - Static method in class org.apache.beam.sdk.values.Row.Equals
 
DEF - Static variable in interface org.apache.beam.sdk.extensions.sql.impl.planner.NodeStatsMetadata
 
DEFAULT - Static variable in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamBigQuerySqlDialect
 
Default - Annotation Type in org.apache.beam.sdk.options
Default represents a set of annotations that can be used to annotate getter properties on PipelineOptions with information representing the default value to be returned if no value is specified.
Default.Boolean - Annotation Type in org.apache.beam.sdk.options
This represents that the default of the option is the specified boolean primitive value.
Default.Byte - Annotation Type in org.apache.beam.sdk.options
This represents that the default of the option is the specified byte primitive value.
Default.Character - Annotation Type in org.apache.beam.sdk.options
This represents that the default of the option is the specified char primitive value.
Default.Class - Annotation Type in org.apache.beam.sdk.options
This represents that the default of the option is the specified Class value.
Default.Double - Annotation Type in org.apache.beam.sdk.options
This represents that the default of the option is the specified double primitive value.
Default.Enum - Annotation Type in org.apache.beam.sdk.options
This represents that the default of the option is the specified enum.
Default.Float - Annotation Type in org.apache.beam.sdk.options
This represents that the default of the option is the specified float primitive value.
Default.InstanceFactory - Annotation Type in org.apache.beam.sdk.options
Value must be of type DefaultValueFactory and have a default constructor.
Default.Integer - Annotation Type in org.apache.beam.sdk.options
This represents that the default of the option is the specified int primitive value.
Default.Long - Annotation Type in org.apache.beam.sdk.options
This represents that the default of the option is the specified long primitive value.
Default.Short - Annotation Type in org.apache.beam.sdk.options
This represents that the default of the option is the specified short primitive value.
Default.String - Annotation Type in org.apache.beam.sdk.options
This represents that the default of the option is the specified String value.
DEFAULT_BUFFER_LIMIT_BYTES - Static variable in interface org.apache.beam.sdk.fn.data.BeamFnDataBufferingOutboundObserver
 
DEFAULT_BUFFER_LIMIT_TIME_MS - Static variable in interface org.apache.beam.sdk.fn.data.BeamFnDataBufferingOutboundObserver
 
DEFAULT_BYTE_ARRAY_CODER - Static variable in class org.apache.beam.sdk.io.TFRecordIO
The default coder, which returns each record of the input file as a byte array.
DEFAULT_CONTEXT - Static variable in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamBigQuerySqlDialect
 
DEFAULT_DURATION - Static variable in class org.apache.beam.sdk.transforms.Deduplicate
The default duration is 10 mins.
DEFAULT_INITIAL_BACKOFF - Static variable in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO
 
DEFAULT_MASTER_URL - Static variable in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
 
DEFAULT_MAX_CUMULATIVE_BACKOFF - Static variable in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO
 
DEFAULT_MAX_INSERT_BLOCK_SIZE - Static variable in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO
 
DEFAULT_MAX_INVOCATION_HISTORY - Static variable in class org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService
The default maximum number of completed invocations to keep.
DEFAULT_MAX_NUM_ELEMENTS - Static variable in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
The cost (in time and space) to compute quantiles to a given accuracy is a function of the total number of elements in the data set.
DEFAULT_MAX_RETRIES - Static variable in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO
 
DEFAULT_OUTBOUND_BUFFER_LIMIT_BYTES - Static variable in class org.apache.beam.sdk.fn.stream.DataStreams
 
DEFAULT_PRECISION - Static variable in class org.apache.beam.sdk.extensions.zetasketch.HllCount
The default precision value used in HllCount.Init.Builder.withPrecision(int) is 15.
DEFAULT_SCHEME - Static variable in class org.apache.beam.sdk.io.FileSystems
 
DEFAULT_TIME_DOMAIN - Static variable in class org.apache.beam.sdk.transforms.Deduplicate
The default is the processing time domain.
DEFAULT_UNWINDOWED_SHARD_TEMPLATE - Static variable in class org.apache.beam.sdk.io.DefaultFilenamePolicy
The default sharding name template.
DEFAULT_WINDOWED_SHARD_TEMPLATE - Static variable in class org.apache.beam.sdk.io.DefaultFilenamePolicy
The default windowed sharding name template used when writing windowed files.
DefaultCoder - Annotation Type in org.apache.beam.sdk.coders
The DefaultCoder annotation specifies a Coder class to handle encoding and decoding instances of the annotated class.
DefaultCoder.DefaultCoderProviderRegistrar - Class in org.apache.beam.sdk.coders
A CoderProviderRegistrar that registers a CoderProvider which can use the @DefaultCoder annotation to provide coder providers that creates Coders.
DefaultCoderCloudObjectTranslatorRegistrar - Class in org.apache.beam.runners.dataflow.util
The CoderCloudObjectTranslatorRegistrar containing the default collection of Coder Cloud Object Translators.
DefaultCoderCloudObjectTranslatorRegistrar() - Constructor for class org.apache.beam.runners.dataflow.util.DefaultCoderCloudObjectTranslatorRegistrar
 
DefaultCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.coders.DefaultCoder.DefaultCoderProviderRegistrar
 
DefaultConcludeTransform() - Constructor for class org.apache.beam.sdk.testing.PAssert.DefaultConcludeTransform
 
defaultConfig(JdbcConnection, RuleSet[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner
 
DefaultExecutableStageContext - Class in org.apache.beam.runners.fnexecution.control
Implementation of a ExecutableStageContext.
DefaultFilenamePolicy - Class in org.apache.beam.sdk.io
A default FileBasedSink.FilenamePolicy for windowed and unwindowed files.
DefaultFilenamePolicy.Params - Class in org.apache.beam.sdk.io
Encapsulates constructor parameters to DefaultFilenamePolicy.
DefaultFilenamePolicy.ParamsCoder - Class in org.apache.beam.sdk.io
DefaultGcpRegionFactory() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowPipelineOptions.DefaultGcpRegionFactory
 
DefaultJobBundleFactory - Class in org.apache.beam.runners.fnexecution.control
A JobBundleFactory for which the implementation can specify a custom EnvironmentFactory for environment management.
DefaultJobBundleFactory.ServerInfo - Class in org.apache.beam.runners.fnexecution.control
A container for EnvironmentFactory and its corresponding Grpc servers.
DefaultJobBundleFactory.WrappedSdkHarnessClient - Class in org.apache.beam.runners.fnexecution.control
Holder for an SdkHarnessClient along with its associated state and data servers.
DefaultJobServerConfigFactory() - Constructor for class org.apache.beam.runners.portability.testing.TestPortablePipelineOptions.DefaultJobServerConfigFactory
 
defaultNaming(String, String) - Static method in class org.apache.beam.sdk.io.FileIO.Write
 
defaultNaming(ValueProvider<String>, ValueProvider<String>) - Static method in class org.apache.beam.sdk.io.FileIO.Write
Defines a default FileIO.Write.FileNaming which will use the prefix and suffix supplied to create a name based on the window, pane, number of shards, shard index, and compression.
DefaultPipelineOptionsRegistrar - Class in org.apache.beam.sdk.options
A PipelineOptionsRegistrar containing the PipelineOptions subclasses available by default.
DefaultPipelineOptionsRegistrar() - Constructor for class org.apache.beam.sdk.options.DefaultPipelineOptionsRegistrar
 
DefaultProjectFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.DefaultProjectFactory
 
defaultPublishResult() - Static method in class org.apache.beam.sdk.io.aws.sns.PublishResultCoders
Returns a new PublishResult coder which by default serializes only the messageId.
DefaultRetryStrategy() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.DefaultRetryStrategy
 
Defaults() - Constructor for class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
 
DefaultS3ClientBuilderFactory - Class in org.apache.beam.sdk.io.aws.s3
Construct AmazonS3ClientBuilder with default values of S3 client properties like path style access, accelerated mode, etc.
DefaultS3ClientBuilderFactory() - Constructor for class org.apache.beam.sdk.io.aws.s3.DefaultS3ClientBuilderFactory
 
DefaultSchema - Annotation Type in org.apache.beam.sdk.schemas.annotations
The DefaultSchema annotation specifies a SchemaProvider class to handle obtaining a schema and row for the specified class.
DefaultSchema.DefaultSchemaProvider - Class in org.apache.beam.sdk.schemas.annotations
SchemaProvider for default schemas.
DefaultSchema.DefaultSchemaProviderRegistrar - Class in org.apache.beam.sdk.schemas.annotations
Registrar for default schemas.
DefaultSchemaProvider() - Constructor for class org.apache.beam.sdk.schemas.annotations.DefaultSchema.DefaultSchemaProvider
 
DefaultSchemaProviderRegistrar() - Constructor for class org.apache.beam.sdk.schemas.annotations.DefaultSchema.DefaultSchemaProviderRegistrar
 
DefaultStopPipelineWatermarkFactory() - Constructor for class org.apache.beam.runners.spark.TestSparkPipelineOptions.DefaultStopPipelineWatermarkFactory
 
DefaultTableFilter - Class in org.apache.beam.sdk.extensions.sql.meta
This default implementation of BeamSqlTableFilter interface.
DefaultTrigger - Class in org.apache.beam.sdk.transforms.windowing
A trigger that is equivalent to Repeatedly.forever(AfterWatermark.pastEndOfWindow()).
defaultType() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.Column
 
DefaultTypeConversionsFactory() - Constructor for class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.DefaultTypeConversionsFactory
 
defaultValue() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.Column
 
defaultValue() - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
Returns the default value when there are no values added to the accumulator.
defaultValue() - Method in interface org.apache.beam.sdk.transforms.CombineFnBase.GlobalCombineFn
Returns the default value when there are no values added to the accumulator.
defaultValue() - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
 
defaultValue() - Method in class org.apache.beam.sdk.transforms.View.AsSingleton
Returns the default value of this transform, or null if there isn't one.
DefaultValueFactory<T> - Interface in org.apache.beam.sdk.options
An interface used with the Default.InstanceFactory annotation to specify the class that will be an instance factory to produce default values for a given getter on PipelineOptions.
delay(Duration) - Static method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform
For internal use only; no backwards-compatibility guarantees.
Delay() - Constructor for class org.apache.beam.sdk.transforms.windowing.TimestampTransform.Delay
 
DelayIntervalRateLimiter() - Constructor for class org.apache.beam.sdk.io.kinesis.RateLimitPolicyFactory.DelayIntervalRateLimiter
 
DelayIntervalRateLimiter(Supplier<Duration>) - Constructor for class org.apache.beam.sdk.io.kinesis.RateLimitPolicyFactory.DelayIntervalRateLimiter
 
delegate() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.ParDoSingle
 
delegate(HasDisplayData) - Method in interface org.apache.beam.sdk.transforms.display.DisplayData.Builder
Register display data from the specified component on behalf of the current component.
delegateBasedUponType(EnumMap<BeamFnApi.StateKey.TypeCase, StateRequestHandler>) - Static method in class org.apache.beam.runners.fnexecution.state.StateRequestHandlers
Returns a StateRequestHandler which delegates to the supplied handler depending on the BeamFnApi.StateRequests type.
DelegateCoder<T,IntermediateT> - Class in org.apache.beam.sdk.coders
A DelegateCoder<T, IntermediateT> wraps a Coder for IntermediateT and encodes/decodes values of type T by converting to/from IntermediateT and then encoding/decoding using the underlying Coder<IntermediateT>.
DelegateCoder(Coder<IntermediateT>, DelegateCoder.CodingFunction<T, IntermediateT>, DelegateCoder.CodingFunction<IntermediateT, T>, TypeDescriptor<T>) - Constructor for class org.apache.beam.sdk.coders.DelegateCoder
 
DelegateCoder.CodingFunction<InputT,OutputT> - Interface in org.apache.beam.sdk.coders
A CodingFunction<InputT, OutputT> is a serializable function from InputT to OutputT that may throw any Exception.
DelegatingCounter - Class in org.apache.beam.sdk.metrics
Implementation of Counter that delegates to the instance for the current context.
DelegatingCounter(MetricName) - Constructor for class org.apache.beam.sdk.metrics.DelegatingCounter
 
delete() - Static method in class org.apache.beam.sdk.io.cassandra.CassandraIO
Provide a CassandraIO.Write PTransform to delete data to a Cassandra database.
delete(Collection<ResourceIdT>) - Method in class org.apache.beam.sdk.io.FileSystem
Deletes a collection of resources.
delete(Collection<ResourceId>, MoveOptions...) - Static method in class org.apache.beam.sdk.io.FileSystems
Deletes a collection of resources.
deleteAsync(T) - Method in interface org.apache.beam.sdk.io.cassandra.Mapper
This method is called for each delete event.
deleteDataset(String, String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
Deletes the dataset specified by the datasetId value.
deleteDataset(String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
 
deleteDataset(String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
deletedTimer(TimerInternals.TimerData) - Method in class org.apache.beam.runners.direct.WatermarkManager.TimerUpdate.TimerUpdateBuilder
Adds the provided timer to the collection of deleted timers, removing it from set timers if it has previously been set.
deleteEntity() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1
Returns an empty DatastoreV1.DeleteEntity builder.
deleteKey() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1
Returns an empty DatastoreV1.DeleteKey builder.
deleteSubscription(PubsubClient.SubscriptionPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
Delete subscription.
deleteSubscription(PubsubClient.SubscriptionPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
 
deleteSubscription(PubsubClient.SubscriptionPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
 
deleteSubscription(PubsubClient.SubscriptionPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
deleteTable(TableReference) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
Deletes the table specified by tableId from the dataset.
deleteTable(String, String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
 
deleteTable(TableReference) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
deleteTimer(StateNamespace, String, TimeDomain) - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
 
deleteTimer(TimerInternals.TimerData) - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
 
deleteTimer(StateNamespace, String, String) - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
 
deleteTimer(Instant, TimeDomain) - Method in interface org.apache.beam.sdk.state.Timers
Removes the timer set in this context for the timestmap and timeDomain.
deleteTopic(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
 
deleteTopic(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
 
deleteTopic(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
 
deleteTopic(PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
delimitElement() - Method in class org.apache.beam.sdk.fn.stream.DataStreams.ElementDelimitedOutputStream
 
dependsOnlyOnEarliestTimestamp() - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
Returns true if the result of combination of many output timestamps actually depends only on the earliest.
dependsOnlyOnWindow() - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
Returns true if the result does not depend on what outputs were combined but only the window they are in.
deregister() - Method in interface org.apache.beam.runners.fnexecution.state.StateDelegator.Registration
De-registers the handler for all future requests for state for the registered process bundle instruction id.
deriveIterableValueCoder(WindowedValue.FullWindowedValueCoder) - Static method in class org.apache.beam.runners.jet.Utils
 
deriveRowType() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnnestRel
 
describe(Set<Class<? extends PipelineOptions>>) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
Outputs the set of options available to be set for the passed in PipelineOptions interfaces.
describeMismatchSafely(BigqueryMatcher.TableAndQuery, Description) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher
 
describeMismatchSafely(ShardedFile, Description) - Method in class org.apache.beam.sdk.testing.FileChecksumMatcher
 
describeMismatchSafely(T, Description) - Method in class org.apache.beam.sdk.testing.JsonMatcher
 
describePipelineOptions(JobApi.DescribePipelineOptionsRequest, StreamObserver<JobApi.DescribePipelineOptionsResponse>) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService
 
describeTo(Description) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher
 
describeTo(Description) - Method in class org.apache.beam.sdk.testing.FileChecksumMatcher
 
describeTo(Description) - Method in class org.apache.beam.sdk.testing.JsonMatcher
 
describeTo(Description) - Method in class org.apache.beam.sdk.testing.RegexMatcher
 
describeTo(Description) - Method in class org.apache.beam.sdk.testing.TestPipelineOptions.AlwaysPassMatcher
 
Description - Annotation Type in org.apache.beam.sdk.options
Descriptions are used to generate human readable output when the --help command is specified.
deserialize(String, byte[]) - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantDeserializer
 
deserialize(JsonParser, DeserializationContext) - Method in class org.apache.beam.sdk.options.ValueProvider.Deserializer
 
deserializeTimers(Collection<byte[]>, TimerInternals.TimerDataCoderV2) - Static method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
 
detect(String) - Static method in enum org.apache.beam.sdk.io.Compression
 
detectTranslationMode(Pipeline, StreamingOptions) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
Visit the pipeline to determine the translation mode (batch/streaming) and update options accordingly.
DirectOptions - Interface in org.apache.beam.runners.direct
Options that can be used to configure the DirectRunner.
DirectOptions.AvailableParallelismFactory - Class in org.apache.beam.runners.direct
A DefaultValueFactory that returns the result of Runtime.availableProcessors() from the DirectOptions.AvailableParallelismFactory.create(PipelineOptions) method.
DIRECTORY_CONTAINER - Static variable in class org.apache.beam.sdk.io.ShardNameTemplate
Shard is a file within a directory.
DirectRegistrar - Class in org.apache.beam.runners.direct
Contains the PipelineRunnerRegistrar and PipelineOptionsRegistrar for the DirectRunner.
DirectRegistrar.Options - Class in org.apache.beam.runners.direct
Registers the DirectOptions.
DirectRegistrar.Runner - Class in org.apache.beam.runners.direct
Registers the DirectRunner.
DirectRunner - Class in org.apache.beam.runners.direct
A PipelineRunner that executes a Pipeline within the process that constructed the Pipeline.
DirectRunner() - Constructor for class org.apache.beam.sdk.options.PipelineOptions.DirectRunner
 
DirectRunner.DirectPipelineResult - Class in org.apache.beam.runners.direct
The result of running a Pipeline with the DirectRunner.
DirectStreamObserver<T> - Class in org.apache.beam.sdk.fn.stream
A StreamObserver which uses synchronization on the underlying CallStreamObserver to provide thread safety.
DirectStreamObserver(Phaser, CallStreamObserver<T>) - Constructor for class org.apache.beam.sdk.fn.stream.DirectStreamObserver
 
DirectTestOptions - Interface in org.apache.beam.runners.direct
Internal-only options for tweaking the behavior of the DirectRunner in ways that users should never do.
DISALLOW_COMBINER_LIFTING - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
discardingFiredPanes() - Method in class org.apache.beam.sdk.transforms.windowing.Window
Returns a new Window PTransform that uses the registered WindowFn and Triggering behavior, and that discards elements in a pane after they are triggered.
dispatchBag(Coder<?>) - Method in interface org.apache.beam.sdk.state.StateSpec.Cases
 
dispatchBag(Coder<?>) - Method in class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
 
dispatchCombining(Combine.CombineFn<?, ?, ?>, Coder<?>) - Method in interface org.apache.beam.sdk.state.StateSpec.Cases
 
dispatchCombining(Combine.CombineFn<?, ?, ?>, Coder<?>) - Method in class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
 
dispatchDefault() - Method in class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
 
dispatchMap(Coder<?>, Coder<?>) - Method in interface org.apache.beam.sdk.state.StateSpec.Cases
 
dispatchMap(Coder<?>, Coder<?>) - Method in class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
 
dispatchSet(Coder<?>) - Method in interface org.apache.beam.sdk.state.StateSpec.Cases
 
dispatchSet(Coder<?>) - Method in class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
 
dispatchValue(Coder<?>) - Method in interface org.apache.beam.sdk.state.StateSpec.Cases
 
dispatchValue(Coder<?>) - Method in class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
 
DISPLAY_DATA - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
DisplayData - Class in org.apache.beam.sdk.transforms.display
Static display data associated with a pipeline component.
DisplayData.Builder - Interface in org.apache.beam.sdk.transforms.display
Utility to build up display data from a component and its included subcomponents.
DisplayData.Identifier - Class in org.apache.beam.sdk.transforms.display
Unique identifier for a display data item within a component.
DisplayData.Item - Class in org.apache.beam.sdk.transforms.display
Items are the unit of display data.
DisplayData.ItemSpec<T> - Class in org.apache.beam.sdk.transforms.display
Specifies an DisplayData.Item to register as display data.
DisplayData.Path - Class in org.apache.beam.sdk.transforms.display
Structured path of registered display data within a component hierarchy.
DisplayData.Type - Enum in org.apache.beam.sdk.transforms.display
Display data type.
Distinct<T> - Class in org.apache.beam.sdk.transforms
Distinct<T> takes a PCollection<T> and returns a PCollection<T> that has all distinct elements of the input.
Distinct() - Constructor for class org.apache.beam.sdk.transforms.Distinct
 
Distinct.WithRepresentativeValues<T,IdT> - Class in org.apache.beam.sdk.transforms
A Distinct PTransform that uses a SerializableFunction to obtain a representative value for each input element.
Distribution - Interface in org.apache.beam.sdk.metrics
A metric that reports information about the distribution of reported values.
distribution(String, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
Create a metric that records various statistics about the distribution of reported values.
distribution(Class<?>, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
Create a metric that records various statistics about the distribution of reported values.
DistributionImpl - Class in org.apache.beam.runners.jet.metrics
Implementation of Distribution.
DistributionImpl(MetricName) - Constructor for class org.apache.beam.runners.jet.metrics.DistributionImpl
 
DistributionResult - Class in org.apache.beam.sdk.metrics
The result of a Distribution metric.
DistributionResult() - Constructor for class org.apache.beam.sdk.metrics.DistributionResult
 
divideBy(RelOptCost) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
doChecks(PAssert.PAssertionSite, ActualT, SerializableFunction<ActualT, Void>) - Static method in class org.apache.beam.sdk.testing.PAssert
 
DockerEnvironmentFactory - Class in org.apache.beam.runners.fnexecution.environment
An EnvironmentFactory that creates docker containers by shelling out to docker.
DockerEnvironmentFactory.Provider - Class in org.apache.beam.runners.fnexecution.environment
Provider for DockerEnvironmentFactory.
DoFn<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
The argument to ParDo providing the code to use to process elements of the input PCollection.
DoFn() - Constructor for class org.apache.beam.sdk.transforms.DoFn
 
DoFn.AlwaysFetched - Annotation Type in org.apache.beam.sdk.transforms
Annotation for declaring that a state parameter is always fetched.
DoFn.BoundedPerElement - Annotation Type in org.apache.beam.sdk.transforms
Annotation on a splittable DoFn specifying that the DoFn performs a bounded amount of work per input element, so applying it to a bounded PCollection will produce also a bounded PCollection.
DoFn.BundleFinalizer - Interface in org.apache.beam.sdk.transforms
A parameter that is accessible during @StartBundle, @ProcessElement and @FinishBundle that allows the caller to register a callback that will be invoked after the bundle has been successfully completed and the runner has commit the output.
DoFn.BundleFinalizer.Callback - Interface in org.apache.beam.sdk.transforms
An instance of a function that will be invoked after bundle finalization.
DoFn.Element - Annotation Type in org.apache.beam.sdk.transforms
DoFn.FieldAccess - Annotation Type in org.apache.beam.sdk.transforms
Annotation for specifying specific fields that are accessed in a Schema PCollection.
DoFn.FinishBundle - Annotation Type in org.apache.beam.sdk.transforms
Annotation for the method to use to finish processing a batch of elements.
DoFn.FinishBundleContext - Class in org.apache.beam.sdk.transforms
Information accessible while within the DoFn.FinishBundle method.
DoFn.GetInitialRestriction - Annotation Type in org.apache.beam.sdk.transforms
Annotation for the method that maps an element to an initial restriction for a splittable DoFn.
DoFn.GetInitialWatermarkEstimatorState - Annotation Type in org.apache.beam.sdk.transforms
Annotation for the method that maps an element and restriction to initial watermark estimator state for a splittable DoFn.
DoFn.GetRestrictionCoder - Annotation Type in org.apache.beam.sdk.transforms
Annotation for the method that returns the coder to use for the restriction of a splittable DoFn.
DoFn.GetSize - Annotation Type in org.apache.beam.sdk.transforms
Annotation for the method that returns the corresponding size for an element and restriction pair.
DoFn.GetWatermarkEstimatorStateCoder - Annotation Type in org.apache.beam.sdk.transforms
Annotation for the method that returns the coder to use for the watermark estimator state of a splittable DoFn.
DoFn.MultiOutputReceiver - Interface in org.apache.beam.sdk.transforms
Receives tagged output for a multi-output function.
DoFn.NewTracker - Annotation Type in org.apache.beam.sdk.transforms
Annotation for the method that creates a new RestrictionTracker for the restriction of a splittable DoFn.
DoFn.NewWatermarkEstimator - Annotation Type in org.apache.beam.sdk.transforms
Annotation for the method that creates a new WatermarkEstimator for the watermark state of a splittable DoFn.
DoFn.OnTimer - Annotation Type in org.apache.beam.sdk.transforms
Annotation for registering a callback for a timer.
DoFn.OnTimerContext - Class in org.apache.beam.sdk.transforms
Information accessible when running a DoFn.OnTimer method.
DoFn.OnTimerFamily - Annotation Type in org.apache.beam.sdk.transforms
Annotation for registering a callback for a timerFamily.
DoFn.OnWindowExpiration - Annotation Type in org.apache.beam.sdk.transforms
Annotation for the method to use for performing actions on window expiration.
DoFn.OutputReceiver<T> - Interface in org.apache.beam.sdk.transforms
Receives values of the given type.
DoFn.ProcessContext - Class in org.apache.beam.sdk.transforms
Information accessible when running a DoFn.ProcessElement method.
DoFn.ProcessContinuation - Class in org.apache.beam.sdk.transforms
When used as a return value of DoFn.ProcessElement, indicates whether there is more work to be done for the current element.
DoFn.ProcessElement - Annotation Type in org.apache.beam.sdk.transforms
Annotation for the method to use for processing elements.
DoFn.RequiresStableInput - Annotation Type in org.apache.beam.sdk.transforms
Experimental - no backwards compatibility guarantees.
DoFn.RequiresTimeSortedInput - Annotation Type in org.apache.beam.sdk.transforms
Experimental - no backwards compatibility guarantees.
DoFn.Restriction - Annotation Type in org.apache.beam.sdk.transforms
DoFn.Setup - Annotation Type in org.apache.beam.sdk.transforms
Annotation for the method to use to prepare an instance for processing bundles of elements.
DoFn.SideInput - Annotation Type in org.apache.beam.sdk.transforms
Parameter annotation for the SideInput for a DoFn.ProcessElement method.
DoFn.SplitRestriction - Annotation Type in org.apache.beam.sdk.transforms
Annotation for the method that splits restriction of a splittable DoFn into multiple parts to be processed in parallel.
DoFn.StartBundle - Annotation Type in org.apache.beam.sdk.transforms
Annotation for the method to use to prepare an instance for processing a batch of elements.
DoFn.StartBundleContext - Class in org.apache.beam.sdk.transforms
Information accessible while within the DoFn.StartBundle method.
DoFn.StateId - Annotation Type in org.apache.beam.sdk.transforms
Annotation for declaring and dereferencing state cells.
DoFn.Teardown - Annotation Type in org.apache.beam.sdk.transforms
Annotation for the method to use to clean up this instance before it is discarded.
DoFn.TimerFamily - Annotation Type in org.apache.beam.sdk.transforms
Parameter annotation for the TimerMap for a DoFn.ProcessElement method.
DoFn.TimerId - Annotation Type in org.apache.beam.sdk.transforms
Annotation for declaring and dereferencing timers.
DoFn.Timestamp - Annotation Type in org.apache.beam.sdk.transforms
DoFn.UnboundedPerElement - Annotation Type in org.apache.beam.sdk.transforms
Annotation on a splittable DoFn specifying that the DoFn performs an unbounded amount of work per input element, so applying it to a bounded PCollection will produce an unbounded PCollection.
DoFn.WatermarkEstimatorState - Annotation Type in org.apache.beam.sdk.transforms
Parameter annotation for the watermark estimator state for the DoFn.NewWatermarkEstimator method.
DoFn.WindowedContext - Class in org.apache.beam.sdk.transforms
Information accessible to all methods in this DoFn where the context is in some window.
DoFnFunction<InputT,OutputT> - Class in org.apache.beam.runners.gearpump.translators.functions
Gearpump FlatMapFunction wrapper over Beam DoFn.
DoFnFunction(GearpumpPipelineOptions, DoFn<InputT, OutputT>, WindowingStrategy<?, ?>, Collection<PCollectionView<?>>, Map<String, PCollectionView<?>>, TupleTag<OutputT>, Map<TupleTag<?>, Coder<?>>, List<TupleTag<?>>, DoFnSchemaInformation, Map<String, PCollectionView<?>>) - Constructor for class org.apache.beam.runners.gearpump.translators.functions.DoFnFunction
 
DoFnFunction<InputT,OutputT> - Class in org.apache.beam.runners.spark.structuredstreaming.translation.batch
Encapsulates a DoFn inside a Spark MapPartitionsFunction.
DoFnFunction(MetricsContainerStepMapAccumulator, String, DoFn<InputT, OutputT>, WindowingStrategy<?, ?>, Map<PCollectionView<?>, WindowingStrategy<?, ?>>, SerializablePipelineOptions, List<TupleTag<?>>, TupleTag<OutputT>, Coder<InputT>, Map<TupleTag<?>, Coder<?>>, SideInputBroadcast, DoFnSchemaInformation, Map<String, PCollectionView<?>>) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.batch.DoFnFunction
 
DoFnOutputReceivers - Class in org.apache.beam.sdk.transforms
DoFnOutputReceivers() - Constructor for class org.apache.beam.sdk.transforms.DoFnOutputReceivers
 
DoFnRunnerFactory<InputT,OutputT> - Class in org.apache.beam.runners.gearpump.translators.utils
a serializable SimpleDoFnRunner.
DoFnRunnerFactory(GearpumpPipelineOptions, DoFn<InputT, OutputT>, Collection<PCollectionView<?>>, DoFnRunners.OutputManager, TupleTag<OutputT>, List<TupleTag<?>>, StepContext, Map<TupleTag<?>, Coder<?>>, WindowingStrategy<?, ?>, DoFnSchemaInformation, Map<String, PCollectionView<?>>) - Constructor for class org.apache.beam.runners.gearpump.translators.utils.DoFnRunnerFactory
 
DoFnRunnerWithMetricsUpdate<InputT,OutputT> - Class in org.apache.beam.runners.flink.metrics
DoFnRunner decorator which registers MetricsContainerImpl.
DoFnRunnerWithMetricsUpdate(String, DoFnRunner<InputT, OutputT>, FlinkMetricContainer) - Constructor for class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
 
DoFnSchemaInformation - Class in org.apache.beam.sdk.transforms
Represents information about how a DoFn extracts schemas.
DoFnSchemaInformation() - Constructor for class org.apache.beam.sdk.transforms.DoFnSchemaInformation
 
DoFnSchemaInformation.Builder - Class in org.apache.beam.sdk.transforms
The builder object.
DoFnTester<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
Deprecated.
Use TestPipeline with the DirectRunner.
DoFnTester.CloningBehavior - Enum in org.apache.beam.sdk.transforms
Deprecated.
Use TestPipeline with the DirectRunner.
doGenCode(CodegenContext, ExprCode) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.DecodeUsingBeamCoder
 
doGenCode(CodegenContext, ExprCode) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.EncodeUsingBeamCoder
 
dotExpression() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
dotExpression() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.FieldSpecifierContext
 
dotExpressionComponent() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
dotExpressionComponent() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionContext
 
dotExpressionComponent(int) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionContext
 
DotExpressionComponentContext(ParserRuleContext, int) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionComponentContext
 
DotExpressionComponentContext() - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionComponentContext
 
DotExpressionContext(ParserRuleContext, int) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionContext
 
DOUBLE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
DOUBLE - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
The type of double fields.
DoubleCoder - Class in org.apache.beam.sdk.coders
A DoubleCoder encodes Double values in 8 bytes using Java serialization.
doubles() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for Double.
doublesGlobally() - Static method in class org.apache.beam.sdk.transforms.Max
Returns a PTransform that takes an input PCollection<Double> and returns a PCollection<Double> whose contents is the maximum of the input PCollection's elements, or Double.NEGATIVE_INFINITY if there are no elements.
doublesGlobally() - Static method in class org.apache.beam.sdk.transforms.Min
Returns a PTransform that takes an input PCollection<Double> and returns a PCollection<Double> whose contents is the minimum of the input PCollection's elements, or Double.POSITIVE_INFINITY if there are no elements.
doublesGlobally() - Static method in class org.apache.beam.sdk.transforms.Sum
Returns a PTransform that takes an input PCollection<Double> and returns a PCollection<Double> whose contents is the sum of the input PCollection's elements, or 0 if there are no elements.
doublesPerKey() - Static method in class org.apache.beam.sdk.transforms.Max
Returns a PTransform that takes an input PCollection<KV<K, Double>> and returns a PCollection<KV<K, Double>> that contains an output element mapping each distinct key in the input PCollection to the maximum of the values associated with that key in the input PCollection.
doublesPerKey() - Static method in class org.apache.beam.sdk.transforms.Min
Returns a PTransform that takes an input PCollection<KV<K, Double>> and returns a PCollection<KV<K, Double>> that contains an output element mapping each distinct key in the input PCollection to the minimum of the values associated with that key in the input PCollection.
doublesPerKey() - Static method in class org.apache.beam.sdk.transforms.Sum
Returns a PTransform that takes an input PCollection<KV<K, Double>> and returns a PCollection<KV<K, Double>> that contains an output element mapping each distinct key in the input PCollection to the sum of the values associated with that key in the input PCollection.
drive() - Method in interface org.apache.beam.runners.local.ExecutionDriver
 
dropExpiredTimers(SparkTimerInternals, WindowingStrategy<?, W>) - Static method in class org.apache.beam.runners.spark.util.TimerUtils
 
DropFields - Class in org.apache.beam.sdk.schemas.transforms
A transform to drop fields from a schema.
DropFields() - Constructor for class org.apache.beam.sdk.schemas.transforms.DropFields
 
DropFields.Inner<T> - Class in org.apache.beam.sdk.schemas.transforms
Implementation class for DropFields.
dropTable(SqlParserPos, boolean, SqlIdentifier) - Static method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlDdlNodes
Creates a DROP TABLE.
dropTable(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
 
dropTable(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.InMemoryMetaTableProvider
 
dropTable(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.ReadOnlyTableProvider
 
dropTable(String) - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.TableProvider
Drops a table.
dropTable(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
 
dropTable(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.store.InMemoryMetaStore
 
dryRunQuery(String, JobConfigurationQuery, String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.JobService
Dry runs the query in the given project.
dryRunQuery(String, JobConfigurationQuery, String) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
 
DurationCoder - Class in org.apache.beam.sdk.coders
A Coder that encodes a joda Duration as a Long using the format of VarLongCoder.
DurationConvert() - Constructor for class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.DurationConvert
 
DynamicAvroDestinations<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io
A specialization of FileBasedSink.DynamicDestinations for AvroIO.
DynamicAvroDestinations() - Constructor for class org.apache.beam.sdk.io.DynamicAvroDestinations
 
DynamicDestinations() - Constructor for class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
 
DynamicDestinations<T,DestinationT> - Class in org.apache.beam.sdk.io.gcp.bigquery
This class provides the most general way of specifying dynamic BigQuery table destinations.
DynamicDestinations() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
 
DynamicFileDestinations - Class in org.apache.beam.sdk.io
Some helper classes that derive from FileBasedSink.DynamicDestinations.
DynamicFileDestinations() - Constructor for class org.apache.beam.sdk.io.DynamicFileDestinations
 
DynamicProtoCoder - Class in org.apache.beam.sdk.extensions.protobuf
A Coder using Google Protocol Buffers binary format.
DynamoDbClientProvider - Interface in org.apache.beam.sdk.io.aws2.dynamodb
Provides instances of DynamoDB clients.
DynamoDBIO - Class in org.apache.beam.sdk.io.aws.dynamodb
PTransforms to read/write from/to Amazon DynamoDB.
DynamoDBIO() - Constructor for class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO
 
DynamoDBIO - Class in org.apache.beam.sdk.io.aws2.dynamodb
PTransforms to read/write from/to Amazon DynamoDB.
DynamoDBIO() - Constructor for class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO
 
DynamoDBIO.Read<T> - Class in org.apache.beam.sdk.io.aws.dynamodb
Read data from DynamoDB and return ScanResult.
DynamoDBIO.Read<T> - Class in org.apache.beam.sdk.io.aws2.dynamodb
Read data from DynamoDB and return ScanResult.
DynamoDBIO.RetryConfiguration - Class in org.apache.beam.sdk.io.aws.dynamodb
A POJO encapsulating a configuration for retry behavior when issuing requests to DynamoDB.
DynamoDBIO.RetryConfiguration - Class in org.apache.beam.sdk.io.aws2.dynamodb
A POJO encapsulating a configuration for retry behavior when issuing requests to DynamoDB.
DynamoDBIO.Write<T> - Class in org.apache.beam.sdk.io.aws.dynamodb
Write a PCollection data into DynamoDB.
DynamoDBIO.Write<T> - Class in org.apache.beam.sdk.io.aws2.dynamodb
Write a PCollection data into DynamoDB.

E

eitherOf(Watch.Growth.TerminationCondition<InputT, FirstStateT>, Watch.Growth.TerminationCondition<InputT, SecondStateT>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
Returns a Watch.Growth.TerminationCondition that holds when at least one of the given two conditions holds.
ElasticsearchIO - Class in org.apache.beam.sdk.io.elasticsearch
Transforms for reading and writing data from/to Elasticsearch.
ElasticsearchIO.BoundedElasticsearchSource - Class in org.apache.beam.sdk.io.elasticsearch
A BoundedSource reading from Elasticsearch.
ElasticsearchIO.ConnectionConfiguration - Class in org.apache.beam.sdk.io.elasticsearch
A POJO describing a connection configuration to Elasticsearch.
ElasticsearchIO.Read - Class in org.apache.beam.sdk.io.elasticsearch
A PTransform reading data from Elasticsearch.
ElasticsearchIO.RetryConfiguration - Class in org.apache.beam.sdk.io.elasticsearch
A POJO encapsulating a configuration for retry behavior when issuing requests to ES.
ElasticsearchIO.Write - Class in org.apache.beam.sdk.io.elasticsearch
A PTransform writing data to Elasticsearch.
ElasticsearchIO.Write.FieldValueExtractFn - Interface in org.apache.beam.sdk.io.elasticsearch
Interface allowing a specific field value to be returned from a parsed JSON document.
element() - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContext
Returns the input element to be processed.
element() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn.AssignContext
Returns the current element.
element() - Method in class org.apache.beam.sdk.transforms.WithFailures.ExceptionElement
 
elementCoder() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.SideInputSpec
 
elementCoder() - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.IterableSideInputHandler
Returns the Coder to use for the elements of the resulting values iterable.
elementCountAtLeast(int) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterPane
Creates a trigger that fires when the pane contains at least countElems elements.
ElementDelimitedOutputStream(DataStreams.OutputChunkConsumer<ByteString>, int) - Constructor for class org.apache.beam.sdk.fn.stream.DataStreams.ElementDelimitedOutputStream
 
ElementEvent() - Constructor for class org.apache.beam.sdk.testing.TestStream.ElementEvent
 
elements() - Static method in class org.apache.beam.sdk.transforms.ToString
Transforms each element of the input PCollection to a String using the Object.toString() method.
elementsRead() - Static method in class org.apache.beam.sdk.metrics.SourceMetrics
Counter of elements read by a source.
elementsReadBySplit(String) - Static method in class org.apache.beam.sdk.metrics.SourceMetrics
Counter of elements read by a source split.
elementsWritten() - Static method in class org.apache.beam.sdk.metrics.SinkMetrics
Counter of elements written to a sink.
EmbeddedEnvironmentFactory - Class in org.apache.beam.runners.fnexecution.environment
An EnvironmentFactory that communicates to a FnHarness which is executing in the same process.
EmbeddedEnvironmentFactory.Provider - Class in org.apache.beam.runners.fnexecution.environment
Provider of EmbeddedEnvironmentFactory.
EmbeddedMetastoreService - Class in org.apache.beam.sdk.io.hcatalog.test
Implementation of a light-weight embedded metastore.
EmbeddedMetastoreService(String) - Constructor for class org.apache.beam.sdk.io.hcatalog.test.EmbeddedMetastoreService
 
empty() - Static method in class org.apache.beam.runners.direct.WatermarkManager.TimerUpdate
Returns a TimerUpdate for a null key with no timers.
empty() - Static method in class org.apache.beam.runners.local.StructuralKey
Get the empty StructuralKey.
EMPTY - Static variable in class org.apache.beam.sdk.io.range.ByteKey
An empty key.
empty() - Static method in class org.apache.beam.sdk.metrics.GaugeResult
 
EMPTY - Static variable in class org.apache.beam.sdk.schemas.utils.SchemaZipFold.Context
 
empty() - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
Asserts that the iterable in question is empty.
empty() - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
 
empty() - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.empty().
empty(Schema) - Static method in class org.apache.beam.sdk.transforms.Create
Returns a new Create.Values transform that produces an empty PCollection of rows.
empty(Coder<T>) - Static method in class org.apache.beam.sdk.transforms.Create
Returns a new Create.Values transform that produces an empty PCollection.
empty(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.transforms.Create
Returns a new Create.Values transform that produces an empty PCollection.
empty() - Static method in class org.apache.beam.sdk.transforms.join.CoGbkResult
Returns an empty CoGbkResult.
empty(Pipeline) - Static method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
Returns an empty KeyedPCollectionTuple<K> on the given pipeline.
empty() - Static method in class org.apache.beam.sdk.transforms.Requirements
Describes an empty set of requirements.
empty(Pipeline) - Static method in class org.apache.beam.sdk.values.PCollectionList
Returns an empty PCollectionList that is part of the given Pipeline.
empty(Pipeline) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
Returns an empty PCollectionTuple that is part of the given Pipeline.
empty() - Static method in class org.apache.beam.sdk.values.TupleTagList
Returns an empty TupleTagList.
EMPTY_ROW - Static variable in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.AggregationCombineFnAdapter
 
EMPTY_SCHEMA - Static variable in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.AggregationCombineFnAdapter
 
emptyArray() - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.emptyArray().
emptyBatch() - Method in class org.apache.beam.runners.spark.io.CreateStream
Adds an empty batch.
EmptyCheckpointMark - Class in org.apache.beam.runners.spark.io
Passing null values to Spark's Java API may cause problems because of Guava preconditions.
emptyDataset() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
emptyIterable() - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.emptyIterable().
EmptyListenersList() - Constructor for class org.apache.beam.runners.spark.SparkContextOptions.EmptyListenersList
 
EmptyMatchTreatment - Enum in org.apache.beam.sdk.io.fs
Options for allowing or disallowing filepatterns that match no resources in FileSystems.match(java.util.List<java.lang.String>).
emulateNullDirection(SqlNode, boolean, boolean) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamBigQuerySqlDialect
 
enableAbandonedNodeEnforcement(boolean) - Method in class org.apache.beam.sdk.testing.TestPipeline
Enables the abandoned node detection.
enableAutoRunIfMissing(boolean) - Method in class org.apache.beam.sdk.testing.TestPipeline
If enabled, a pipeline.run() statement will be added automatically in case it is missing in the test.
enableSSL() - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
Enable SSL connection to Redis server.
EnableStreamingEngineFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.EnableStreamingEngineFactory
 
encode(RandomAccessData, OutputStream) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
 
encode(RandomAccessData, OutputStream, Coder.Context) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
 
encode() - Method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService.StagingSessionToken
 
encode(T, Coder<T>) - Static method in class org.apache.beam.runners.jet.Utils
 
encode(boolean, T, Coder<T>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.EncodeUsingBeamCoder
Convert value to byte array (invoked by generated code in EncoderHelpers.EncodeUsingBeamCoder.doGenCode(CodegenContext, ExprCode)).
encode(Tuple2<TupleTag<T>, WindowedValue<T>>, OutputStream) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.MultiOuputCoder
 
encode(T, OutputStream) - Method in class org.apache.beam.sdk.coders.AvroCoder
 
encode(BigDecimal, OutputStream) - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
 
encode(BigDecimal, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
 
encode(Integer, OutputStream) - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
 
encode(Long, OutputStream) - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
 
encode(Short, OutputStream) - Method in class org.apache.beam.sdk.coders.BigEndianShortCoder
 
encode(BigInteger, OutputStream) - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
 
encode(BigInteger, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
 
encode(BitSet, OutputStream) - Method in class org.apache.beam.sdk.coders.BitSetCoder
 
encode(BitSet, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.BitSetCoder
 
encode(Boolean, OutputStream) - Method in class org.apache.beam.sdk.coders.BooleanCoder
 
encode(byte[], OutputStream) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
 
encode(byte[], OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
 
encode(Byte, OutputStream) - Method in class org.apache.beam.sdk.coders.ByteCoder
 
encode(T, OutputStream) - Method in class org.apache.beam.sdk.coders.Coder
Encodes the given value of type T onto the given output stream.
encode(T, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.Coder
Deprecated.
encode(T, OutputStream) - Method in class org.apache.beam.sdk.coders.DelegateCoder
 
encode(T, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.DelegateCoder
 
encode(Double, OutputStream) - Method in class org.apache.beam.sdk.coders.DoubleCoder
 
encode(ReadableDuration, OutputStream) - Method in class org.apache.beam.sdk.coders.DurationCoder
 
encode(Float, OutputStream) - Method in class org.apache.beam.sdk.coders.FloatCoder
 
encode(Instant, OutputStream) - Method in class org.apache.beam.sdk.coders.InstantCoder
 
encode(IterableT, OutputStream) - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
 
encode(KV<K, V>, OutputStream) - Method in class org.apache.beam.sdk.coders.KvCoder
 
encode(KV<K, V>, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.KvCoder
 
encode(T, OutputStream) - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
 
encode(Map<K, V>, OutputStream) - Method in class org.apache.beam.sdk.coders.MapCoder
 
encode(Map<K, V>, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.MapCoder
 
encode(T, OutputStream) - Method in class org.apache.beam.sdk.coders.NullableCoder
 
encode(T, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.NullableCoder
 
encode(T, OutputStream) - Method in class org.apache.beam.sdk.coders.SerializableCoder
 
encode(ShardedKey<KeyT>, OutputStream) - Method in class org.apache.beam.sdk.coders.ShardedKeyCoder
 
encode(T, OutputStream) - Method in class org.apache.beam.sdk.coders.SnappyCoder
 
encode(T, OutputStream) - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
 
encode(T, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
 
encode(String, OutputStream) - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
 
encode(String, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
 
encode(Integer, OutputStream) - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
 
encode(Integer, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
 
encode(Integer, OutputStream) - Method in class org.apache.beam.sdk.coders.VarIntCoder
 
encode(Long, OutputStream) - Method in class org.apache.beam.sdk.coders.VarLongCoder
 
encode(Void, OutputStream) - Method in class org.apache.beam.sdk.coders.VoidCoder
 
encode(ByteString, OutputStream) - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
 
encode(ByteString, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
 
encode(T, OutputStream) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
 
encode(T, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
 
encode(HyperLogLogPlus, OutputStream) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.HyperLogLogPlusCoder
 
encode(EncodedBoundedWindow, OutputStream) - Method in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow.Coder
 
encode(Message, OutputStream) - Method in class org.apache.beam.sdk.io.amqp.AmqpMessageCoder
 
encode(AttributeValue, OutputStream) - Method in class org.apache.beam.sdk.io.aws.dynamodb.AttributeValueCoder
 
encode(AttributeValue, OutputStream) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.AttributeValueCoder
 
encode(PublishResponse, OutputStream) - Method in class org.apache.beam.sdk.io.aws2.sns.PublishResponseCoder
 
encode(Message, OutputStream) - Method in class org.apache.beam.sdk.io.aws2.sqs.MessageCoder
 
encode(SendMessageRequest, OutputStream) - Method in class org.apache.beam.sdk.io.aws2.sqs.SendMessageRequestCoder
 
encode(DefaultFilenamePolicy.Params, OutputStream) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.ParamsCoder
 
encode(FileBasedSink.FileResult<DestinationT>, OutputStream) - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
 
encode(MatchResult.Metadata, OutputStream) - Method in class org.apache.beam.sdk.io.fs.MetadataCoder
 
encode(MatchResult.Metadata, OutputStream) - Method in class org.apache.beam.sdk.io.fs.MetadataCoderV2
 
encode(ResourceId, OutputStream) - Method in class org.apache.beam.sdk.io.fs.ResourceIdCoder
 
encode(BigQueryInsertError, OutputStream) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertErrorCoder
 
encode(TableDestination, OutputStream) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoder
 
encode(TableDestination, OutputStream) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV2
 
encode(TableDestination, OutputStream) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV3
 
encode(TableRow, OutputStream) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
 
encode(TableRow, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
 
encode(BigtableWriteResult, OutputStream) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteResultCoder
 
encode(PubsubMessage, OutputStream) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessagePayloadOnlyCoder
 
encode(PubsubMessage, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessagePayloadOnlyCoder
 
encode(PubsubMessage, OutputStream) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesAndMessageIdCoder
 
encode(PubsubMessage, OutputStream) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
 
encode(PubsubMessage, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
 
encode(PubsubMessage, OutputStream) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithMessageIdCoder
 
encode(T, OutputStream) - Method in class org.apache.beam.sdk.io.hadoop.WritableCoder
 
encode(KafkaRecord<K, V>, OutputStream) - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
 
encode(ProducerRecord<K, V>, OutputStream) - Method in class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
 
encode(OffsetRange, OutputStream) - Method in class org.apache.beam.sdk.io.range.OffsetRange.Coder
 
encode(FileIO.ReadableFile, OutputStream) - Method in class org.apache.beam.sdk.io.ReadableFileCoder
 
encode(T, OutputStream) - Method in class org.apache.beam.sdk.io.xml.JAXBCoder
 
encode(T, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.io.xml.JAXBCoder
 
encode(T, OutputStream) - Method in class org.apache.beam.sdk.schemas.SchemaCoder
 
encode(TestStream<T>, OutputStream) - Method in class org.apache.beam.sdk.testing.TestStream.TestStreamCoder
 
encode(CoGbkResult, OutputStream) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
 
encode(RawUnionValue, OutputStream) - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
 
encode(RawUnionValue, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
 
encode(GlobalWindow, OutputStream) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
 
encode(IntervalWindow, OutputStream) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
 
encode(PaneInfo, OutputStream) - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo.PaneInfoCoder
 
encode(TimestampedValue<T>, OutputStream) - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
 
encode(ValueInSingleWindow<T>, OutputStream) - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
 
encode(ValueInSingleWindow<T>, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
 
encode(ValueWithRecordId<ValueT>, OutputStream) - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
 
encode(ValueWithRecordId<ValueT>, OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
 
encodeAndOwn(byte[], OutputStream, Coder.Context) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
Encodes the provided value with the identical encoding to ByteArrayCoder.encode(byte[], java.io.OutputStream), but with optimizations that take ownership of the value.
EncodedBoundedWindow - Class in org.apache.beam.sdk.fn.windowing
An encoded BoundedWindow used within Runners to track window information without needing to decode the window.
EncodedBoundedWindow() - Constructor for class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow
 
EncodedBoundedWindow.Coder - Class in org.apache.beam.sdk.fn.windowing
encodeQueryResult(Table) - Static method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
 
encodeQueryResult(Table, List<TableRow>) - Static method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
 
EncoderHelpers - Class in org.apache.beam.runners.spark.structuredstreaming.translation.helpers
Encoders utility class.
EncoderHelpers() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers
 
EncoderHelpers.DecodeUsingBeamCoder<T> - Class in org.apache.beam.runners.spark.structuredstreaming.translation.helpers
Catalyst Expression that deserializes elements using Beam Coder.
EncoderHelpers.EncodeUsingBeamCoder<T> - Class in org.apache.beam.runners.spark.structuredstreaming.translation.helpers
Catalyst Expression that serializes elements using Beam Coder.
encodeToTimerDataTimerId(String, String) - Static method in class org.apache.beam.runners.fnexecution.control.TimerReceiverFactory
Encodes transform and timer family ids into a single string which retains the human readable format len(transformId):transformId:timerId.
EncodeUsingBeamCoder(Expression, Coder<T>) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.EncodeUsingBeamCoder
 
ENCODING - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
end() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
Returns the end of this window, exclusive.
ENDS_WITH - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
ENDS_WITH_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamBuiltinMethods
 
endsWith(Path) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
endsWith(String) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
endsWith(String, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
endsWith(String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
endsWith(String) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
enterArrayQualifier(FieldSpecifierNotationParser.ArrayQualifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Enter a parse tree produced by FieldSpecifierNotationParser.arrayQualifier().
enterArrayQualifier(FieldSpecifierNotationParser.ArrayQualifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Enter a parse tree produced by FieldSpecifierNotationParser.arrayQualifier().
enterArrayQualifierList(FieldSpecifierNotationParser.ArrayQualifierListContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Enter a parse tree produced by the arrayQualifierList labeled alternative in FieldSpecifierNotationParser.qualifierList().
enterArrayQualifierList(FieldSpecifierNotationParser.ArrayQualifierListContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Enter a parse tree produced by the arrayQualifierList labeled alternative in FieldSpecifierNotationParser.qualifierList().
enterCompositeTransform(TransformHierarchy.Node) - Method in class org.apache.beam.runners.gearpump.translators.GearpumpPipelineTranslator
 
enterCompositeTransform(TransformHierarchy.Node) - Method in class org.apache.beam.runners.spark.SparkNativePipelineVisitor
 
enterCompositeTransform(TransformHierarchy.Node) - Method in class org.apache.beam.runners.spark.SparkRunner.Evaluator
 
enterCompositeTransform(TransformHierarchy.Node) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
 
enterCompositeTransform(TransformHierarchy.Node) - Method in class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
 
enterCompositeTransform(TransformHierarchy.Node) - Method in interface org.apache.beam.sdk.Pipeline.PipelineVisitor
Called for each composite transform after all topological predecessors have been visited but before any of its component transforms.
enterDotExpression(FieldSpecifierNotationParser.DotExpressionContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Enter a parse tree produced by FieldSpecifierNotationParser.dotExpression().
enterDotExpression(FieldSpecifierNotationParser.DotExpressionContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Enter a parse tree produced by FieldSpecifierNotationParser.dotExpression().
enterEveryRule(ParserRuleContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
enterFieldSpecifier(FieldSpecifierNotationParser.FieldSpecifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Enter a parse tree produced by FieldSpecifierNotationParser.fieldSpecifier().
enterFieldSpecifier(FieldSpecifierNotationParser.FieldSpecifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Enter a parse tree produced by FieldSpecifierNotationParser.fieldSpecifier().
enterMapQualifier(FieldSpecifierNotationParser.MapQualifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Enter a parse tree produced by FieldSpecifierNotationParser.mapQualifier().
enterMapQualifier(FieldSpecifierNotationParser.MapQualifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Enter a parse tree produced by FieldSpecifierNotationParser.mapQualifier().
enterMapQualifierList(FieldSpecifierNotationParser.MapQualifierListContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Enter a parse tree produced by the mapQualifierList labeled alternative in FieldSpecifierNotationParser.qualifierList().
enterMapQualifierList(FieldSpecifierNotationParser.MapQualifierListContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Enter a parse tree produced by the mapQualifierList labeled alternative in FieldSpecifierNotationParser.qualifierList().
enterPipeline(Pipeline) - Method in class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
 
enterPipeline(Pipeline) - Method in interface org.apache.beam.sdk.Pipeline.PipelineVisitor
Called before visiting anything values or transforms, as many uses of a visitor require access to the Pipeline object itself.
enterQualifiedComponent(FieldSpecifierNotationParser.QualifiedComponentContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
enterQualifiedComponent(FieldSpecifierNotationParser.QualifiedComponentContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
enterQualifyComponent(FieldSpecifierNotationParser.QualifyComponentContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Enter a parse tree produced by the qualifyComponent labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
enterQualifyComponent(FieldSpecifierNotationParser.QualifyComponentContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Enter a parse tree produced by the qualifyComponent labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
enterRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierContext
 
enterRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierListContext
 
enterRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionContext
 
enterRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.FieldSpecifierContext
 
enterRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierContext
 
enterRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierListContext
 
enterRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifiedComponentContext
 
enterRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifyComponentContext
 
enterRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.SimpleIdentifierContext
 
enterRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.WildcardContext
 
enterSimpleIdentifier(FieldSpecifierNotationParser.SimpleIdentifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Enter a parse tree produced by the simpleIdentifier labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
enterSimpleIdentifier(FieldSpecifierNotationParser.SimpleIdentifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Enter a parse tree produced by the simpleIdentifier labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
enterWildcard(FieldSpecifierNotationParser.WildcardContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Enter a parse tree produced by the wildcard labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
enterWildcard(FieldSpecifierNotationParser.WildcardContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Enter a parse tree produced by the wildcard labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
entries() - Method in interface org.apache.beam.sdk.state.MapState
Returns an Iterable over the key-value pairs contained in this map.
entrySet() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedMap
 
entrySet() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
enum16(Map<String, Integer>) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
enum8(Map<String, Integer>) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
EnumerationType - Class in org.apache.beam.sdk.schemas.logicaltypes
This Schema.LogicalType represent an enumeration over a fixed set of values.
EnumerationType.Value - Class in org.apache.beam.sdk.schemas.logicaltypes
This class represents a single enum value.
enumValues() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
ENVIRONMENT_VERSION_JOB_TYPE_KEY - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
ENVIRONMENT_VERSION_MAJOR_KEY - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
EnvironmentFactory - Interface in org.apache.beam.runners.fnexecution.environment
Creates environments which communicate to an SdkHarnessClient.
EnvironmentFactory.Provider - Interface in org.apache.beam.runners.fnexecution.environment
Provider for a EnvironmentFactory and ServerFactory for the environment.
equal(T) - Static method in class org.apache.beam.sdk.transforms.Filter
Returns a PTransform that takes an input PCollection<T> and returns a PCollection<T> with elements that equals to a given value.
equals(Object) - Method in class org.apache.beam.runners.dataflow.util.CloudObject
 
equals(Object) - Method in class org.apache.beam.runners.dataflow.util.OutputReference
 
equals(Object) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
 
equals(Object) - Method in class org.apache.beam.runners.direct.WatermarkManager.TimerUpdate
 
equals(Object) - Method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils.RawUnionValue
 
equals(Object) - Method in class org.apache.beam.runners.jet.Utils.ByteArrayKey
 
equals(Object) - Method in class org.apache.beam.runners.spark.io.EmptyCheckpointMark
 
equals(Object) - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
 
equals(Object) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.DecodeUsingBeamCoder
 
equals(Object) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.EncodeUsingBeamCoder
 
equals(Object) - Method in class org.apache.beam.runners.spark.util.ByteArray
 
equals(Object) - Method in class org.apache.beam.sdk.coders.AtomicCoder
.
equals(Object) - Method in class org.apache.beam.sdk.coders.AvroCoder
 
equals(Object) - Method in class org.apache.beam.sdk.coders.Coder.Context
Deprecated.
 
equals(Object) - Method in class org.apache.beam.sdk.coders.DelegateCoder
 
equals(Object) - Method in class org.apache.beam.sdk.coders.RowCoder
 
equals(Object) - Method in class org.apache.beam.sdk.coders.SerializableCoder
 
equals(Object) - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
 
equals(Object) - Method in class org.apache.beam.sdk.coders.StructuralByteArray
 
equals(Object) - Method in class org.apache.beam.sdk.coders.StructuredCoder
equals(Object) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
 
equals(Object) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
equals(Object) - Method in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
 
equals(Object) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
 
equals(Object) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
 
equals(Object) - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
 
equals(Object) - Method in class org.apache.beam.sdk.extensions.sql.example.model.Order
 
equals(RelOptCost) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
equals(Object) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
equals(Object) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
 
equals(Object) - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
 
equals(Object) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertError
 
equals(Object) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
equals(Object) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.ProjectPath
 
equals(Object) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.SubscriptionPath
 
equals(Object) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.TopicPath
 
equals(Object) - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
 
equals(Object) - Method in class org.apache.beam.sdk.io.hadoop.WritableCoder
 
equals(Object) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
 
equals(Object) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
 
equals(Object) - Method in class org.apache.beam.sdk.io.jms.JmsRecord
 
equals(Object) - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
 
equals(Object) - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
equals(Object) - Method in class org.apache.beam.sdk.io.range.ByteKey
 
equals(Object) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
 
equals(Object) - Method in class org.apache.beam.sdk.io.range.OffsetRange
 
equals(Object) - Method in class org.apache.beam.sdk.io.tika.ParseResult
 
equals(Object) - Method in class org.apache.beam.sdk.io.xml.JAXBCoder
 
equals(Object) - Method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
 
equals(Object) - Method in class org.apache.beam.sdk.options.ValueProvider.RuntimeValueProvider
 
equals(Object) - Method in class org.apache.beam.sdk.options.ValueProvider.StaticValueProvider
 
equals(Object) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
 
equals(Object) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema
 
equals(Object) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema.GetterTypeSupplier
 
equals(Object) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema.SetterTypeSupplier
 
equals(Object) - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType.Value
 
equals(Object) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType.Value
 
equals(Object) - Method in class org.apache.beam.sdk.schemas.Schema
Returns true if two Schemas have the same fields in the same order.
equals(Object) - Method in class org.apache.beam.sdk.schemas.Schema.Field
 
equals(Object) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
 
equals(Object) - Method in class org.apache.beam.sdk.schemas.Schema.Options
 
equals(Object) - Method in class org.apache.beam.sdk.schemas.SchemaCoder
 
equals(Object) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
equals(Object) - Method in class org.apache.beam.sdk.testing.PAssert.PAssertionSite
 
equals(Object) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
Deprecated.
Object.equals(Object) is not supported on PAssert objects. If you meant to test object equality, use a variant of PAssert.PCollectionContentsAssert.containsInAnyOrder(T...) instead.
equals(Object) - Method in class org.apache.beam.sdk.testing.SuccessOrFailure
 
equals(Object) - Method in class org.apache.beam.sdk.testing.TestStream
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn.LargestUnique
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.CombineFns.CoCombineResult
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.display.DisplayData
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Path
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.join.RawUnionValue
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.windowing.AfterSynchronizedProcessingTime
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
 
equals(Object) - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
 
equals(Object) - Method in class org.apache.beam.sdk.values.KV
 
equals(Object) - Method in class org.apache.beam.sdk.values.PCollectionList
 
equals(Object) - Method in class org.apache.beam.sdk.values.PCollectionTuple
 
equals(Object) - Method in class org.apache.beam.sdk.values.PCollectionViews.ListViewFn
 
equals(Object) - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
 
equals(Object) - Method in class org.apache.beam.sdk.values.Row
 
Equals() - Constructor for class org.apache.beam.sdk.values.Row.Equals
 
equals(Object) - Method in class org.apache.beam.sdk.values.RowWithGetters
 
equals(Object) - Method in class org.apache.beam.sdk.values.ShardedKey
 
equals(Object) - Method in class org.apache.beam.sdk.values.TimestampedValue
 
equals(Object) - Method in class org.apache.beam.sdk.values.TupleTag
 
equals(Object) - Method in class org.apache.beam.sdk.values.TypeDescriptor
Two type descriptor are equal if and only if they represent the same type.
equals(Object) - Method in class org.apache.beam.sdk.values.TypeParameter
 
equals(Object) - Method in class org.apache.beam.sdk.values.ValueWithRecordId
 
equals(Object) - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
equalTo(T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.equalTo(Object).
equalTo(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.equalTo(Object).
equivalent(Schema) - Method in class org.apache.beam.sdk.schemas.Schema
Returns true if two Schemas have the same fields, but possibly in different orders.
equivalent(Schema.FieldType, Schema.EquivalenceNullablePolicy) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
Check whether two types are equivalent.
ErrorContainer<T> - Interface in org.apache.beam.sdk.io.gcp.bigquery
ErrorContainer interface.
estimateCount(T, Coder<T>) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.Sketch
Utility class to retrieve the estimate frequency of an element from a CountMinSketch.
estimateFractionForKey(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
Returns the fraction of this range [startKey, endKey) that is in the interval [startKey, key).
estimateNodeStats(RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.AbstractBeamCalcRel
 
estimateNodeStats(RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamAggregationRel
 
estimateNodeStats(RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIntersectRel
 
estimateNodeStats(RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
 
estimateNodeStats(RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
 
estimateNodeStats(RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
 
estimateNodeStats(RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamMinusRel
 
estimateNodeStats(RelMetadataQuery) - Method in interface org.apache.beam.sdk.extensions.sql.impl.rel.BeamRelNode
This method is called by org.apache.beam.sdk.extensions.sql.impl.planner.RelMdNodeStats.
estimateNodeStats(RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSortRel
 
estimateNodeStats(RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUncollectRel
 
estimateNodeStats(RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnionRel
 
estimateNodeStats(RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnnestRel
 
estimateNodeStats(RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamValuesRel
 
estimateRowCount(RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
 
estimateRowCount(PipelineOptions) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
Estimates the number of non empty rows.
Evaluator(SparkPipelineTranslator, EvaluationContext) - Constructor for class org.apache.beam.runners.spark.SparkRunner.Evaluator
 
eventually(Matcher<Iterable<? extends Row>>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery.RowsAssertion
 
ever() - Static method in class org.apache.beam.sdk.transforms.windowing.Never
Returns a trigger which never fires.
every(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
Returns a new SlidingWindows with the original size, that assigns timestamps into half-open intervals of the form [N * period, N * period + size), where 0 is the epoch.
EXACTLY_ONCE - Static variable in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
exception() - Method in class org.apache.beam.sdk.transforms.WithFailures.ExceptionElement
 
ExceptionAsMapHandler() - Constructor for class org.apache.beam.sdk.transforms.WithFailures.ExceptionAsMapHandler
 
ExceptionElement() - Constructor for class org.apache.beam.sdk.transforms.WithFailures.ExceptionElement
 
exceptionsInto(TypeDescriptor<NewFailureT>) - Method in class org.apache.beam.sdk.extensions.jackson.AsJsons
Returns a new AsJsons.AsJsonsWithFailures transform that catches exceptions raised while writing JSON elements, with the given type descriptor used for the failure collection but the exception handler yet to be specified using AsJsons.AsJsonsWithFailures.exceptionsVia(ProcessFunction).
exceptionsInto(TypeDescriptor<NewFailureT>) - Method in class org.apache.beam.sdk.extensions.jackson.ParseJsons
Returns a new ParseJsons.ParseJsonsWithFailures transform that catches exceptions raised while parsing elements, with the given type descriptor used for the failure collection but the exception handler yet to be specified using ParseJsons.ParseJsonsWithFailures.exceptionsVia(ProcessFunction).
exceptionsInto(TypeDescriptor<NewFailureT>) - Method in class org.apache.beam.sdk.transforms.FlatMapElements
Returns a new FlatMapElements.FlatMapWithFailures transform that catches exceptions raised while mapping elements, with the given type descriptor used for the failure collection but the exception handler yet to be specified using FlatMapElements.FlatMapWithFailures.exceptionsVia(ProcessFunction).
exceptionsInto(TypeDescriptor<NewFailureT>) - Method in class org.apache.beam.sdk.transforms.MapElements
Returns a new MapElements.MapWithFailures transform that catches exceptions raised while mapping elements, with the given type descriptor used for the failure collection but the exception handler yet to be specified using MapElements.MapWithFailures.exceptionsVia(ProcessFunction).
exceptionsVia(ProcessFunction<WithFailures.ExceptionElement<InputT>, FailureT>) - Method in class org.apache.beam.sdk.extensions.jackson.AsJsons.AsJsonsWithFailures
Returns a new AsJsons.AsJsonsWithFailures transform that catches exceptions raised while writing JSON elements, passing the raised exception instance and the input element being processed through the given exceptionHandler and emitting the result to a failure collection.
exceptionsVia(InferableFunction<WithFailures.ExceptionElement<InputT>, FailureT>) - Method in class org.apache.beam.sdk.extensions.jackson.AsJsons
Returns a new AsJsons.AsJsonsWithFailures transform that catches exceptions raised while writing JSON elements, passing the raised exception instance and the input element being processed through the given exceptionHandler and emitting the result to a failure collection.
exceptionsVia() - Method in class org.apache.beam.sdk.extensions.jackson.AsJsons
Returns a new AsJsons.AsJsonsWithFailures transform that catches exceptions raised while writing JSON elements, passing the raised exception instance and the input element being processed through the default exception handler DefaultExceptionAsMapHandler and emitting the result to a failure collection.
exceptionsVia(InferableFunction<WithFailures.ExceptionElement<String>, FailureT>) - Method in class org.apache.beam.sdk.extensions.jackson.ParseJsons
Returns a new ParseJsons.ParseJsonsWithFailures transform that catches exceptions raised while parsing elements, passing the raised exception instance and the input element being processed through the given exceptionHandler and emitting the result to a failure collection.
exceptionsVia() - Method in class org.apache.beam.sdk.extensions.jackson.ParseJsons
Returns a new ParseJsons.ParseJsonsWithFailures transform that catches exceptions raised while parsing elements, passing the raised exception instance and the input element being processed through the default exception handler DefaultExceptionAsMapHandler and emitting the result to a failure collection.
exceptionsVia(ProcessFunction<WithFailures.ExceptionElement<String>, FailureT>) - Method in class org.apache.beam.sdk.extensions.jackson.ParseJsons.ParseJsonsWithFailures
Returns a new ParseJsons.ParseJsonsWithFailures transform that catches exceptions raised while parsing elements, passing the raised exception instance and the input element being processed through the given exceptionHandler and emitting the result to a failure collection.
exceptionsVia(InferableFunction<WithFailures.ExceptionElement<InputT>, FailureT>) - Method in class org.apache.beam.sdk.transforms.FlatMapElements
Returns a new FlatMapElements.FlatMapWithFailures transform that catches exceptions raised while mapping elements, passing the raised exception instance and the input element being processed through the given exceptionHandler and emitting the result to a failure collection.
exceptionsVia(ProcessFunction<WithFailures.ExceptionElement<InputT>, FailureT>) - Method in class org.apache.beam.sdk.transforms.FlatMapElements.FlatMapWithFailures
Returns a new FlatMapElements.FlatMapWithFailures transform that catches exceptions raised while mapping elements, passing the raised exception instance and the input element being processed through the given exceptionHandler and emitting the result to a failure collection.
exceptionsVia(InferableFunction<WithFailures.ExceptionElement<InputT>, FailureT>) - Method in class org.apache.beam.sdk.transforms.MapElements
Returns a new MapElements.MapWithFailures transform that catches exceptions raised while mapping elements, passing the raised exception instance and the input element being processed through the given exceptionHandler and emitting the result to a failure collection.
exceptionsVia(ProcessFunction<WithFailures.ExceptionElement<InputT>, FailureT>) - Method in class org.apache.beam.sdk.transforms.MapElements.MapWithFailures
Returns a PTransform that catches exceptions raised while mapping elements, passing the raised exception instance and the input element being processed through the given exceptionHandler and emitting the result to a failure collection.
ExecutableGraph<ExecutableT,CollectionT> - Interface in org.apache.beam.runners.direct
The interface that enables querying of a graph of independently executable stages and the inputs and outputs of those stages.
ExecutableProcessBundleDescriptor() - Constructor for class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
 
ExecutableStageContext - Interface in org.apache.beam.runners.fnexecution.control
The context required in order to execute stages.
ExecutableStageContext.Factory - Interface in org.apache.beam.runners.fnexecution.control
Creates ExecutableStageContext instances.
execute(String) - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.BatchTranslationContext
 
execute(String) - Method in interface org.apache.beam.runners.flink.FlinkPortablePipelineTranslator.Executor
 
execute(String) - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext
 
execute(String) - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlCli
Executes the given sql.
execute(Expression, Class<T>) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
execute(Expression, Type) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
execute(CalcitePrepare.Context) - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlCreateExternalTable
 
execute(CalcitePrepare.Context) - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlSetOptionBeam
 
executeDdl(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
 
executeQuery(Queryable<T>) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
executeQuery(String) - Method in class org.apache.beam.sdk.io.hcatalog.test.EmbeddedMetastoreService
Executes the passed query on the embedded metastore service.
ExecutionDriver - Interface in org.apache.beam.runners.local
Drives the execution of a Pipeline by scheduling work.
ExecutionDriver.DriverState - Enum in org.apache.beam.runners.local
The state of the driver.
ExecutorServiceFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcsOptions.ExecutorServiceFactory
 
exitArrayQualifier(FieldSpecifierNotationParser.ArrayQualifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Exit a parse tree produced by FieldSpecifierNotationParser.arrayQualifier().
exitArrayQualifier(FieldSpecifierNotationParser.ArrayQualifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Exit a parse tree produced by FieldSpecifierNotationParser.arrayQualifier().
exitArrayQualifierList(FieldSpecifierNotationParser.ArrayQualifierListContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Exit a parse tree produced by the arrayQualifierList labeled alternative in FieldSpecifierNotationParser.qualifierList().
exitArrayQualifierList(FieldSpecifierNotationParser.ArrayQualifierListContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Exit a parse tree produced by the arrayQualifierList labeled alternative in FieldSpecifierNotationParser.qualifierList().
exitDotExpression(FieldSpecifierNotationParser.DotExpressionContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Exit a parse tree produced by FieldSpecifierNotationParser.dotExpression().
exitDotExpression(FieldSpecifierNotationParser.DotExpressionContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Exit a parse tree produced by FieldSpecifierNotationParser.dotExpression().
exitEveryRule(ParserRuleContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
exitFieldSpecifier(FieldSpecifierNotationParser.FieldSpecifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Exit a parse tree produced by FieldSpecifierNotationParser.fieldSpecifier().
exitFieldSpecifier(FieldSpecifierNotationParser.FieldSpecifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Exit a parse tree produced by FieldSpecifierNotationParser.fieldSpecifier().
exitMapQualifier(FieldSpecifierNotationParser.MapQualifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Exit a parse tree produced by FieldSpecifierNotationParser.mapQualifier().
exitMapQualifier(FieldSpecifierNotationParser.MapQualifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Exit a parse tree produced by FieldSpecifierNotationParser.mapQualifier().
exitMapQualifierList(FieldSpecifierNotationParser.MapQualifierListContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Exit a parse tree produced by the mapQualifierList labeled alternative in FieldSpecifierNotationParser.qualifierList().
exitMapQualifierList(FieldSpecifierNotationParser.MapQualifierListContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Exit a parse tree produced by the mapQualifierList labeled alternative in FieldSpecifierNotationParser.qualifierList().
exitQualifiedComponent(FieldSpecifierNotationParser.QualifiedComponentContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
exitQualifiedComponent(FieldSpecifierNotationParser.QualifiedComponentContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
exitQualifyComponent(FieldSpecifierNotationParser.QualifyComponentContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Exit a parse tree produced by the qualifyComponent labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
exitQualifyComponent(FieldSpecifierNotationParser.QualifyComponentContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Exit a parse tree produced by the qualifyComponent labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
exitRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierContext
 
exitRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierListContext
 
exitRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionContext
 
exitRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.FieldSpecifierContext
 
exitRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierContext
 
exitRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierListContext
 
exitRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifiedComponentContext
 
exitRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifyComponentContext
 
exitRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.SimpleIdentifierContext
 
exitRule(ParseTreeListener) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.WildcardContext
 
exitSimpleIdentifier(FieldSpecifierNotationParser.SimpleIdentifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Exit a parse tree produced by the simpleIdentifier labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
exitSimpleIdentifier(FieldSpecifierNotationParser.SimpleIdentifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Exit a parse tree produced by the simpleIdentifier labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
exitWildcard(FieldSpecifierNotationParser.WildcardContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
Exit a parse tree produced by the wildcard labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
exitWildcard(FieldSpecifierNotationParser.WildcardContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationListener
Exit a parse tree produced by the wildcard labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
expand(PCollection<ElemT>) - Method in class org.apache.beam.runners.apex.ApexRunner.CreateApexPCollectionView
 
expand(PCollection<ElemT>) - Method in class org.apache.beam.runners.dataflow.CreateDataflowView
 
expand(PCollection<? extends InputT>) - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.ParDoSingle
 
expand(PCollection<T>) - Method in class org.apache.beam.runners.spark.io.ConsoleIO.Write.Unbound
 
expand(PBegin) - Method in class org.apache.beam.runners.spark.io.CreateStream
 
expand(PCollection<String>) - Method in class org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.CountWords
 
expand(ExpansionApi.ExpansionRequest, StreamObserver<ExpansionApi.ExpansionResponse>) - Method in class org.apache.beam.sdk.expansion.service.ExpansionService
 
expand(GcsPath) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
Expands a pattern into matched paths.
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.extensions.jackson.AsJsons.AsJsonsWithFailures
 
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.extensions.jackson.AsJsons
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.extensions.jackson.ParseJsons
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.extensions.jackson.ParseJsons.ParseJsonsWithFailures
 
expand(PCollection<KV<K, V1>>) - Method in class org.apache.beam.sdk.extensions.joinlibrary.Join.FullOuterJoin
 
expand(PCollection<KV<K, V1>>) - Method in class org.apache.beam.sdk.extensions.joinlibrary.Join.InnerJoin
 
expand(PCollection<KV<K, V1>>) - Method in class org.apache.beam.sdk.extensions.joinlibrary.Join.LeftOuterJoin
 
expand(PCollection<KV<K, V1>>) - Method in class org.apache.beam.sdk.extensions.joinlibrary.Join.RightOuterJoin
 
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.GloballyDistinct
 
expand(PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.PerKeyDistinct
 
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.GlobalSketch
 
expand(PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.PerKeySketch
 
expand(PCollection<Double>) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.GlobalDigest
 
expand(PCollection<KV<K, Double>>) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.PerKeyDigest
 
expand(PCollection<KV<PrimaryKeyT, Iterable<KV<SecondaryKeyT, ValueT>>>>) - Method in class org.apache.beam.sdk.extensions.sorter.SortValues
 
expand(PCollectionList<Row>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSetOperatorRelBase
 
expand(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamJoinTransforms.JoinAsLookup
 
expand(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.avro.GenericRecordWriteConverter
 
expand(PCollection<KV<byte[], byte[]>>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaCSVTable.CsvRecorderDecoder
 
expand(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaCSVTable.CsvRecorderEncoder
 
expand(PCollection<Document>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable.DocumentToRow
 
expand(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable.RowToDocument
 
expand(PCollection<GenericRecord>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.parquet.GenericRecordReadConverter
 
expand(PCollection<PubsubMessage>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubMessageToRow
 
expand(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.RowToPubsubMessage
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTableProvider.CsvToRow
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTableProvider.LinesReadConverter
 
expand(PCollection<Row>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTableProvider.LinesWriteConverter
 
expand(PInput) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.amqp.AmqpIO.Read
 
expand(PCollection<Message>) - Method in class org.apache.beam.sdk.io.amqp.AmqpIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.io.AvroIO.ParseAll
Deprecated.
 
expand(PCollection<FileIO.ReadableFile>) - Method in class org.apache.beam.sdk.io.AvroIO.ParseFiles
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.AvroIO.Read
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.io.AvroIO.ReadAll
Deprecated.
 
expand(PCollection<FileIO.ReadableFile>) - Method in class org.apache.beam.sdk.io.AvroIO.ReadFiles
 
expand(PCollection<UserT>) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.io.AvroIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Read
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Write
 
expand(PCollection<PublishRequest>) - Method in class org.apache.beam.sdk.io.aws.sns.SnsIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.aws.sqs.SqsIO.Read
 
expand(PCollection<SendMessageRequest>) - Method in class org.apache.beam.sdk.io.aws.sqs.SqsIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Write
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.io.aws2.sns.SnsIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Read
 
expand(PCollection<SendMessageRequest>) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.BoundedReadFromUnboundedSource
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.FileIO.Match
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.io.FileIO.MatchAll
 
expand(PCollection<MatchResult.Metadata>) - Method in class org.apache.beam.sdk.io.FileIO.ReadMatches
 
expand(PCollection<UserT>) - Method in class org.apache.beam.sdk.io.FileIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
 
expand(PCollection<KV<DestinationT, ElementT>>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.CreateTables
 
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.PrepareWrite
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.ReifyAsIterable
 
expand(PCollection<KV<DestinationT, ElementT>>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StreamingInserts
 
expand(PCollection<KV<TableDestination, ElementT>>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StreamingWriteTables
 
expand() - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
 
expand(PCollection<KV<ByteString, Iterable<Mutation>>>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
 
expand(PCollection<KV<ByteString, Iterable<Mutation>>>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.WriteWithResults
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
 
expand(PCollection<PubsubMessage>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
expand(PCollection<ReadOperation>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
 
expand(PCollection<Mutation>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
 
expand(PCollection<MutationGroup>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.WriteGrouped
 
expand() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteResult
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.GenerateSequence
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
 
expand(PCollection<KV<KeyT, ValueT>>) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
 
expand(PCollection<HBaseIO.Read>) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.ReadAll
 
expand(PCollection<Mutation>) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
 
expand(PCollection<HCatRecord>) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
 
expand(PCollection<ParameterT>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.TypedWithoutMetadata
 
expand(PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
 
expand(PCollection<ProducerRecord<K, V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
 
expand(PCollection<byte[]>) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
 
expand(PCollection<Document>) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Read
 
expand(PCollection<byte[]>) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Read
 
expand(PCollection<FileIO.ReadableFile>) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.ReadFiles
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.Read.Bounded
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.Read.Unbounded
 
expand(PCollection<FileIO.ReadableFile>) - Method in class org.apache.beam.sdk.io.ReadAllViaFileBasedSource
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadAll
 
expand(PCollection<KV<String, String>>) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Read
 
expand(PCollection<SolrInputDocument>) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.TextIO.Read
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
Deprecated.
 
expand(PCollection<FileIO.ReadableFile>) - Method in class org.apache.beam.sdk.io.TextIO.ReadFiles
 
expand(PCollection<UserT>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.io.TextIO.Write
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
 
expand(PCollection<FileIO.ReadableFile>) - Method in class org.apache.beam.sdk.io.TFRecordIO.ReadFiles
 
expand(PCollection<byte[]>) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
 
expand(PCollection<FileIO.ReadableFile>) - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.ReadFiles
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Parse
 
expand(PCollection<FileIO.ReadableFile>) - Method in class org.apache.beam.sdk.io.tika.TikaIO.ParseFiles
 
expand(PCollection<UserT>) - Method in class org.apache.beam.sdk.io.WriteFiles
 
expand() - Method in class org.apache.beam.sdk.io.WriteFilesResult
 
expand(PBegin) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
 
expand(PCollection<FileIO.ReadableFile>) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.schemas.transforms.AddFields.Inner
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.schemas.transforms.Cast
 
expand(PCollectionTuple) - Method in class org.apache.beam.sdk.schemas.transforms.CoGroup.ExpandCrossProduct
 
expand(PCollectionTuple) - Method in class org.apache.beam.sdk.schemas.transforms.CoGroup.Impl
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.schemas.transforms.DropFields.Inner
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.schemas.transforms.Filter.Inner
 
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
 
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
 
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsGlobally
 
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineGlobally
 
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.schemas.transforms.Group.Global
 
expand(PCollection) - Method in class org.apache.beam.sdk.schemas.transforms.Join.Impl
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.schemas.transforms.RenameFields.Inner
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.schemas.transforms.Select.Fields
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.schemas.transforms.Select.Flattened
 
expand(PCollection<SuccessOrFailure>) - Method in class org.apache.beam.sdk.testing.PAssert.DefaultConcludeTransform
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.testing.PAssert.GroupThenAssert
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.testing.PAssert.GroupThenAssertForSingleton
 
expand(PBegin) - Method in class org.apache.beam.sdk.testing.PAssert.OneSideInputAssert
 
expand(PBegin) - Method in class org.apache.beam.sdk.testing.TestStream
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.Globally
 
expand(PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.PerKey
 
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.transforms.Combine.Globally
 
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.transforms.Combine.GloballyAsSingletonView
 
expand(PCollection<? extends KV<K, ? extends Iterable<InputT>>>) - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
 
expand(PCollection<KV<K, InputT>>) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
 
expand(PCollection<KV<K, InputT>>) - Method in class org.apache.beam.sdk.transforms.Combine.PerKeyWithHotKeyFanout
 
expand(PBegin) - Method in class org.apache.beam.sdk.transforms.Create.OfValueProvider
 
expand(PBegin) - Method in class org.apache.beam.sdk.transforms.Create.TimestampedValues
 
expand(PBegin) - Method in class org.apache.beam.sdk.transforms.Create.Values
 
expand(PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.transforms.Deduplicate.KeyedValues
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.Deduplicate.Values
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.Deduplicate.WithRepresentativeValues
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.Distinct
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.Distinct.WithRepresentativeValues
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.Filter
 
expand(PCollection<? extends InputT>) - Method in class org.apache.beam.sdk.transforms.FlatMapElements
 
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.transforms.FlatMapElements.FlatMapWithFailures
 
expand(PCollection<? extends Iterable<T>>) - Method in class org.apache.beam.sdk.transforms.Flatten.Iterables
 
expand(PCollectionList<T>) - Method in class org.apache.beam.sdk.transforms.Flatten.PCollections
 
expand(PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.transforms.GroupByKey
 
expand(PCollection<KV<K, InputT>>) - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches
 
expand(PBegin) - Method in class org.apache.beam.sdk.transforms.Impulse
 
expand(KeyedPCollectionTuple<K>) - Method in class org.apache.beam.sdk.transforms.join.CoGroupByKey
 
expand() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
Expands the component PCollections, stripping off any tag-specific information.
expand(PCollection<? extends KV<K, ?>>) - Method in class org.apache.beam.sdk.transforms.Keys
 
expand(PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.transforms.KvSwap
 
expand(PCollection<? extends InputT>) - Method in class org.apache.beam.sdk.transforms.MapElements
 
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.transforms.MapElements.MapWithFailures
 
expand(PCollection<? extends InputT>) - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
 
expand(PCollection<? extends InputT>) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.Partition
 
expand(InputT) - Method in class org.apache.beam.sdk.transforms.PTransform
Override this method to specify how this PTransform should be expanded on the given InputT.
expand(PCollection<String>) - Method in class org.apache.beam.sdk.transforms.Regex.AllMatches
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.transforms.Regex.Find
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.transforms.Regex.FindAll
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.transforms.Regex.FindKV
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.transforms.Regex.FindName
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.transforms.Regex.FindNameKV
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.transforms.Regex.Matches
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.transforms.Regex.MatchesKV
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.transforms.Regex.MatchesName
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.transforms.Regex.MatchesNameKV
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.transforms.Regex.ReplaceAll
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.transforms.Regex.ReplaceFirst
 
expand(PCollection<String>) - Method in class org.apache.beam.sdk.transforms.Regex.Split
 
expand(PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.transforms.Reshuffle
Deprecated.
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.Reshuffle.ViaRandomKey
Deprecated.
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.ToJson
 
expand(PCollection<? extends KV<?, V>>) - Method in class org.apache.beam.sdk.transforms.Values
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.View.AsIterable
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.View.AsList
 
expand(PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.transforms.View.AsMap
 
expand(PCollection<KV<K, V>>) - Method in class org.apache.beam.sdk.transforms.View.AsMultimap
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.View.AsSingleton
 
expand(PCollection<ElemT>) - Method in class org.apache.beam.sdk.transforms.View.CreatePCollectionView
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.Wait.OnSignal
 
expand(PCollection<InputT>) - Method in class org.apache.beam.sdk.transforms.Watch.Growth
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.windowing.Window.Assign
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.windowing.Window
 
expand() - Method in class org.apache.beam.sdk.transforms.WithFailures.Result
 
expand(PCollection<V>) - Method in class org.apache.beam.sdk.transforms.WithKeys
 
expand(PCollection<T>) - Method in class org.apache.beam.sdk.transforms.WithTimestamps
 
expand() - Method in class org.apache.beam.sdk.values.PBegin
 
expand() - Method in class org.apache.beam.sdk.values.PCollection
 
expand() - Method in class org.apache.beam.sdk.values.PCollectionList
 
expand() - Method in class org.apache.beam.sdk.values.PCollectionTuple
 
expand() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
 
expand() - Method in class org.apache.beam.sdk.values.PDone
A PDone contains no PValues.
expand() - Method in interface org.apache.beam.sdk.values.PInput
Expands this PInput into a list of its component output PValues.
expand() - Method in interface org.apache.beam.sdk.values.POutput
Expands this POutput into a list of its component output PValues.
expand() - Method in interface org.apache.beam.sdk.values.PValue
Deprecated.
A PValue always expands into itself. Calling PValue.expand() on a PValue is almost never appropriate.
ExpansionServer - Class in org.apache.beam.sdk.expansion.service
A gRPC Server for an ExpansionService.
ExpansionService - Class in org.apache.beam.sdk.expansion.service
A service that allows pipeline expand transforms from a remote SDK.
ExpansionService() - Constructor for class org.apache.beam.sdk.expansion.service.ExpansionService
 
ExpansionService.ExpansionServiceRegistrar - Interface in org.apache.beam.sdk.expansion.service
A registrar that creates ExpansionService.TransformProvider instances from RunnerApi.FunctionSpecs.
ExpansionService.ExternalTransformRegistrarLoader - Class in org.apache.beam.sdk.expansion.service
Exposes Java transforms via ExternalTransformRegistrar.
ExpansionService.TransformProvider<InputT extends PInput,OutputT extends POutput> - Interface in org.apache.beam.sdk.expansion.service
Provides a mapping of RunnerApi.FunctionSpec to a PTransform, together with mappings of its inputs and outputs to maps of PCollections.
expectDryRunQuery(String, String, JobStatistics) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
 
Experimental - Annotation Type in org.apache.beam.sdk.annotations
Signifies that a public API (public class, method or field) is subject to incompatible changes, or even removal, in a future release.
Experimental.Kind - Enum in org.apache.beam.sdk.annotations
An enumeration of various kinds of experimental APIs.
ExperimentalOptions - Interface in org.apache.beam.sdk.options
Apache Beam provides a number of experimental features that can be enabled with this flag.
explain(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
 
explainQuery(String) - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlCli
Returns a human readable representation of the query execution plan.
explainTerms(RelWriter) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamAggregationRel
 
explainTerms(RelWriter) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamPushDownIOSourceRel
 
explainTerms(RelWriter) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamUnnestRel
 
ExpressionConverter - Class in org.apache.beam.sdk.extensions.sql.zetasql.translation
Extracts expressions (function calls, field accesses) from the resolve query nodes, converts them to RexNodes.
ExpressionConverter(RelOptCluster, QueryPlanner.QueryParameters) - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
 
expressionsInFilter(List<RexNode>) - Static method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTableFilter
Count a number of RexNodes involved in all supported filters.
extend(String) - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Path
Extend the path by appending a sub-component path.
External() - Constructor for class org.apache.beam.sdk.io.GenerateSequence.External
 
External() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External
 
External() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.Write.External
 
ExternalConfiguration() - Constructor for class org.apache.beam.sdk.io.GenerateSequence.External.ExternalConfiguration
 
ExternalEnvironmentFactory - Class in org.apache.beam.runners.fnexecution.environment
An EnvironmentFactory which requests workers via the given URL in the Environment.
ExternalEnvironmentFactory.Provider - Class in org.apache.beam.runners.fnexecution.environment
Provider of ExternalEnvironmentFactory.
ExternalRead - Class in org.apache.beam.sdk.io.gcp.pubsub
Exposes PubsubIO.Read as an external transform for cross-language usage.
ExternalRead() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead
 
ExternalRead.Configuration - Class in org.apache.beam.sdk.io.gcp.pubsub
Parameters class to expose the transform to an external SDK.
ExternalRead.ReadBuilder - Class in org.apache.beam.sdk.io.gcp.pubsub
 
ExternalSorter - Class in org.apache.beam.sdk.extensions.sorter
Does an external sort of the provided values.
ExternalSorter.Options - Class in org.apache.beam.sdk.extensions.sorter
ExternalSorter.Options contains configuration of the sorter.
ExternalSorter.Options.SorterType - Enum in org.apache.beam.sdk.extensions.sorter
Sorter type.
ExternalSynchronization - Interface in org.apache.beam.sdk.io.hadoop.format
Provides mechanism for acquiring locks related to the job.
ExternalTransformBuilder<ConfigT,InputT extends PInput,OutputT extends POutput> - Interface in org.apache.beam.sdk.transforms
An interface for building a transform from an externally provided configuration.
ExternalTransformRegistrar - Interface in org.apache.beam.sdk.expansion
A registrar which contains a mapping from URNs to available ExternalTransformBuilders.
ExternalTransformRegistrarLoader() - Constructor for class org.apache.beam.sdk.expansion.service.ExpansionService.ExternalTransformRegistrarLoader
 
ExternalWorkerService - Class in org.apache.beam.runners.portability
Implements the BeamFnExternalWorkerPool service by starting a fresh SDK harness for each request.
ExternalWorkerService(PipelineOptions) - Constructor for class org.apache.beam.runners.portability.ExternalWorkerService
 
ExternalWrite - Class in org.apache.beam.sdk.io.gcp.pubsub
Exposes PubsubIO.Write as an external transform for cross-language usage.
ExternalWrite() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite
 
ExternalWrite.Configuration - Class in org.apache.beam.sdk.io.gcp.pubsub
Parameters class to expose the transform to an external SDK.
ExternalWrite.WriteBuilder - Class in org.apache.beam.sdk.io.gcp.pubsub
 
extractFiredTimers() - Method in class org.apache.beam.runners.direct.WatermarkManager
Returns a map of each PTransform that has pending timers to those timers.
extractFromTypeParameters(T, Class<? super T>, TypeDescriptors.TypeVariableExtractor<T, V>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
Extracts a type from the actual type parameters of a parameterized class, subject to Java type erasure.
extractFromTypeParameters(TypeDescriptor<T>, Class<? super T>, TypeDescriptors.TypeVariableExtractor<T, V>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
Like TypeDescriptors.extractFromTypeParameters(Object, Class, TypeVariableExtractor), but takes a TypeDescriptor of the instance being analyzed rather than the instance itself.
extractKey() - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.KVHelpers
A Spark MapFunction for extracting the key out of a KV for GBK for example.
extractOutput(JavaPairRDD<K, <any>>, <any>) - Static method in class org.apache.beam.runners.spark.util.SparkCompat
Extracts the output for a given collection of WindowedAccumulators.
extractOutput(HyperLogLogPlus) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.ApproximateDistinctFn
Output the whole structure so it can be queried, reused or stored easily.
extractOutput(SketchFrequencies.Sketch<InputT>) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.CountMinSketchFn
Output the whole structure so it can be queried, reused or stored easily.
extractOutput(MergingDigest) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.TDigestQuantilesFn
Output the whole structure so it can be queried, reused or stored easily.
extractOutput(CovarianceAccumulator) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CovarianceFn
 
extractOutput(VarianceAccumulator) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.VarianceFn
 
extractOutput(ApproximateUnique.ApproximateUniqueCombineFn.LargestUnique) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn
 
extractOutput() - Method in interface org.apache.beam.sdk.transforms.Combine.AccumulatingCombineFn.Accumulator
Returns the output value that is the result of combining all the input values represented by this accumulator.
extractOutput(AccumT) - Method in class org.apache.beam.sdk.transforms.Combine.AccumulatingCombineFn
 
extractOutput(double[]) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
 
extractOutput(Combine.Holder<V>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
 
extractOutput(int[]) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
 
extractOutput(long[]) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
 
extractOutput(AccumT) - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
Returns the output value that is the result of combining all the input values represented by the given accumulator.
extractOutput(List<V>) - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
 
extractOutput(Object[]) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
 
extractOutput(Object[], CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
 
extractOutput(AccumT, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
Returns the output value that is the result of combining all the input values represented by the given accumulator.
extractOutput(Top.BoundedHeap<KV<Integer, T>, SerializableComparator<KV<Integer, T>>>) - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
 
extractOutputs(OutputT) - Method in interface org.apache.beam.sdk.expansion.service.ExpansionService.TransformProvider
 
extractTableNamesFromNode(SqlNode) - Static method in class org.apache.beam.sdk.extensions.sql.TableNameExtractionUtils
 
extractTimestamp(String, String, Map<String, String>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
Return the timestamp (in ms since unix epoch) to use for a Pubsub message with attributes and pubsubTimestamp.
extractTimestampsFromValues() - Static method in class org.apache.beam.sdk.transforms.Reify
Extracts the timestamps from each value in a KV.
extractWindowedValueFromRowMapFunction(WindowedValue.WindowedValueCoder<T>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.RowHelpers
A Spark MapFunction for extracting a WindowedValue from a Row in which the WindowedValue was serialized to bytes using its WindowedValue.WindowedValueCoder.

F

factory - Variable in class org.apache.beam.sdk.extensions.sql.impl.CalciteFactoryWrapper
 
FACTORY - Static variable in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
Factory() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel.Factory
 
FACTORY - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
Factory for creating Pubsub clients using gRCP transport.
FACTORY - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
Factory for creating Pubsub clients using Json transport.
Factory<T> - Interface in org.apache.beam.sdk.schemas
A Factory interface for schema-related objects for a specific Java type.
fail(Throwable) - Method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
 
fail(Throwable) - Method in class org.apache.beam.sdk.fn.data.CompletableFutureInboundDataClient
 
fail(Throwable) - Method in interface org.apache.beam.sdk.fn.data.InboundDataClient
Mark the client as completed with an exception.
failed(Exception) - Method in interface org.apache.beam.runners.local.PipelineMessageReceiver
Report that a failure has occurred.
failed(Error) - Method in interface org.apache.beam.runners.local.PipelineMessageReceiver
Report that a failure has occurred.
FailedRunningPipelineResults - Class in org.apache.beam.runners.jet
Alternative implementation of PipelineResult used to avoid throwing Exceptions in certain situations.
FailedRunningPipelineResults(RuntimeException) - Constructor for class org.apache.beam.runners.jet.FailedRunningPipelineResults
 
failOnInsert(Map<TableRow, List<TableDataInsertAllResponse.InsertErrors>>) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
Cause a given TableRow object to fail when it's inserted.
failure(String, String, Metadata, Throwable) - Static method in class org.apache.beam.sdk.io.tika.ParseResult
 
failure(PAssert.PAssertionSite, Throwable) - Static method in class org.apache.beam.sdk.testing.SuccessOrFailure
 
FAILURE_COUNTER - Static variable in class org.apache.beam.sdk.testing.PAssert
 
failures() - Method in class org.apache.beam.sdk.transforms.WithFailures.Result
 
failuresTo(List<PCollection<FailureElementT>>) - Method in class org.apache.beam.sdk.transforms.WithFailures.Result
Adds the failure collection to the passed list and returns just the output collection.
FakeBigQueryServerStream(List<T>) - Constructor for class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices.FakeBigQueryServerStream
 
FakeBigQueryServices - Class in org.apache.beam.sdk.io.gcp.testing
A fake implementation of BigQuery's query service..
FakeBigQueryServices() - Constructor for class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
 
FakeBigQueryServices.FakeBigQueryServerStream<T> - Class in org.apache.beam.sdk.io.gcp.testing
An implementation of BigQueryServerStream which takes a List as the Iterable to simulate a server stream.
FakeDatasetService - Class in org.apache.beam.sdk.io.gcp.testing
A fake dataset service that can be serialized, for use in testReadFromTable.
FakeDatasetService() - Constructor for class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
FakeJobService - Class in org.apache.beam.sdk.io.gcp.testing
A fake implementation of BigQuery's job service.
FakeJobService() - Constructor for class org.apache.beam.sdk.io.gcp.testing.FakeJobService
 
FakeJobService(int) - Constructor for class org.apache.beam.sdk.io.gcp.testing.FakeJobService
 
featureList - Variable in class org.apache.beam.sdk.extensions.ml.AnnotateVideo
 
fewKeys() - Method in class org.apache.beam.sdk.transforms.GroupByKey
Returns whether it groups just few keys.
field(Row, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedRow
 
Field() - Constructor for class org.apache.beam.sdk.schemas.Schema.Field
 
field(String, Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.transforms.AddFields.Inner
Add a new field of the specified type.
field(String, Schema.FieldType, Object) - Method in class org.apache.beam.sdk.schemas.transforms.AddFields.Inner
Add a new field of the specified type.
fieldAccess(FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.transforms.Select
Select a set of fields described in a FieldAccessDescriptor.
FieldAccessDescriptor - Class in org.apache.beam.sdk.schemas
Used inside of a DoFn to describe which fields in a schema type need to be accessed for processing.
FieldAccessDescriptor() - Constructor for class org.apache.beam.sdk.schemas.FieldAccessDescriptor
 
fieldAccessDescriptor(FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.transforms.CoGroup.By
Join by the following field access descriptor.
FieldAccessDescriptor.FieldDescriptor - Class in org.apache.beam.sdk.schemas
Description of a single field.
FieldAccessDescriptor.FieldDescriptor.Builder - Class in org.apache.beam.sdk.schemas
Builder class.
FieldAccessDescriptor.FieldDescriptor.ListQualifier - Enum in org.apache.beam.sdk.schemas
Qualifier for a list selector.
FieldAccessDescriptor.FieldDescriptor.MapQualifier - Enum in org.apache.beam.sdk.schemas
Qualifier for a map selector.
FieldAccessDescriptor.FieldDescriptor.Qualifier - Class in org.apache.beam.sdk.schemas
OneOf union for a collection selector.
FieldAccessDescriptor.FieldDescriptor.Qualifier.Kind - Enum in org.apache.beam.sdk.schemas
The kind of qualifier.
FieldAccessDescriptorParser - Class in org.apache.beam.sdk.schemas.parser
Parser for textual field-access selector.
FieldAccessDescriptorParser() - Constructor for class org.apache.beam.sdk.schemas.parser.FieldAccessDescriptorParser
 
FieldDescriptor() - Constructor for class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor
 
fieldFromType(TypeDescriptor, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.StaticSchemaInference
Map a Java field type to a Beam Schema FieldType.
fieldIds(Integer...) - Static method in class org.apache.beam.sdk.schemas.transforms.CoGroup.By
Join by the following field ids.
fieldIds(Integer...) - Static method in class org.apache.beam.sdk.schemas.transforms.Select
Select a set of top-level field ids from the row.
fieldIdsAccessed() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return the field ids accessed.
fieldNames(String...) - Static method in class org.apache.beam.sdk.schemas.transforms.CoGroup.By
Join by the following field names.
fieldNames(String...) - Static method in class org.apache.beam.sdk.schemas.transforms.Select
Select a set of top-level field names from the row.
fieldNamesAccessed() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return the field names accessed.
fields(String...) - Static method in class org.apache.beam.sdk.schemas.transforms.DropFields
 
fields(Integer...) - Static method in class org.apache.beam.sdk.schemas.transforms.DropFields
 
fields(FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.transforms.DropFields
 
Fields() - Constructor for class org.apache.beam.sdk.schemas.transforms.Select.Fields
 
FieldsEqual() - Constructor for class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual
 
fieldSpecifier() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
FieldSpecifierContext(ParserRuleContext, int) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.FieldSpecifierContext
 
FieldSpecifierNotationBaseListener - Class in org.apache.beam.sdk.schemas.parser.generated
This class provides an empty implementation of FieldSpecifierNotationListener, which can be extended to create a listener which only needs to handle a subset of the available methods.
FieldSpecifierNotationBaseListener() - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
 
FieldSpecifierNotationBaseVisitor<T> - Class in org.apache.beam.sdk.schemas.parser.generated
This class provides an empty implementation of FieldSpecifierNotationVisitor, which can be extended to create a visitor which only needs to handle a subset of the available methods.
FieldSpecifierNotationBaseVisitor() - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
 
FieldSpecifierNotationLexer - Class in org.apache.beam.sdk.schemas.parser.generated
 
FieldSpecifierNotationLexer(CharStream) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
FieldSpecifierNotationListener - Interface in org.apache.beam.sdk.schemas.parser.generated
This interface defines a complete listener for a parse tree produced by FieldSpecifierNotationParser.
FieldSpecifierNotationParser - Class in org.apache.beam.sdk.schemas.parser.generated
 
FieldSpecifierNotationParser(TokenStream) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
FieldSpecifierNotationParser.ArrayQualifierContext - Class in org.apache.beam.sdk.schemas.parser.generated
 
FieldSpecifierNotationParser.ArrayQualifierListContext - Class in org.apache.beam.sdk.schemas.parser.generated
 
FieldSpecifierNotationParser.DotExpressionComponentContext - Class in org.apache.beam.sdk.schemas.parser.generated
 
FieldSpecifierNotationParser.DotExpressionContext - Class in org.apache.beam.sdk.schemas.parser.generated
 
FieldSpecifierNotationParser.FieldSpecifierContext - Class in org.apache.beam.sdk.schemas.parser.generated
 
FieldSpecifierNotationParser.MapQualifierContext - Class in org.apache.beam.sdk.schemas.parser.generated
 
FieldSpecifierNotationParser.MapQualifierListContext - Class in org.apache.beam.sdk.schemas.parser.generated
 
FieldSpecifierNotationParser.QualifiedComponentContext - Class in org.apache.beam.sdk.schemas.parser.generated
 
FieldSpecifierNotationParser.QualifierListContext - Class in org.apache.beam.sdk.schemas.parser.generated
 
FieldSpecifierNotationParser.QualifyComponentContext - Class in org.apache.beam.sdk.schemas.parser.generated
 
FieldSpecifierNotationParser.SimpleIdentifierContext - Class in org.apache.beam.sdk.schemas.parser.generated
 
FieldSpecifierNotationParser.WildcardContext - Class in org.apache.beam.sdk.schemas.parser.generated
 
FieldSpecifierNotationVisitor<T> - Interface in org.apache.beam.sdk.schemas.parser.generated
This interface defines a complete generic visitor for a parse tree produced by FieldSpecifierNotationParser.
FieldType() - Constructor for class org.apache.beam.sdk.schemas.Schema.FieldType
 
FieldTypeDescriptors - Class in org.apache.beam.sdk.schemas
Utilities for converting between Schema field types and TypeDescriptors that define Java objects which can represent these field types.
FieldTypeDescriptors() - Constructor for class org.apache.beam.sdk.schemas.FieldTypeDescriptors
 
fieldTypeForJavaType(TypeDescriptor) - Static method in class org.apache.beam.sdk.schemas.FieldTypeDescriptors
FieldValueGetter<ObjectT,ValueT> - Interface in org.apache.beam.sdk.schemas
For internal use only; no backwards-compatibility guarantees.
fieldValueGetters(Class<?>, Schema) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoMessageSchema
 
fieldValueGetters(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.AutoValueSchema
 
fieldValueGetters(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.AvroRecordSchema
 
fieldValueGetters(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
Implementing class should override to return FieldValueGetters.
fieldValueGetters(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema
 
fieldValueGetters(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.JavaFieldSchema
 
FieldValueSetter<ObjectT,ValueT> - Interface in org.apache.beam.sdk.schemas
For internal use only; no backwards-compatibility guarantees.
FieldValueTypeInformation - Class in org.apache.beam.sdk.schemas
Represents type information for a Java type that will be used to infer a Schema type.
FieldValueTypeInformation() - Constructor for class org.apache.beam.sdk.schemas.FieldValueTypeInformation
 
FieldValueTypeInformation.Builder - Class in org.apache.beam.sdk.schemas
 
fieldValueTypeInformations(Class<?>, Schema) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoMessageSchema
 
fieldValueTypeInformations(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.AutoValueSchema
 
fieldValueTypeInformations(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.AvroRecordSchema
 
fieldValueTypeInformations(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
Implementing class should override to return a list of type-informations.
fieldValueTypeInformations(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema
 
fieldValueTypeInformations(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.JavaFieldSchema
 
FieldValueTypeSupplier - Interface in org.apache.beam.sdk.schemas.utils
A naming policy for schema fields.
FileBasedReader(FileBasedSource<T>) - Constructor for class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
Subclasses should not perform IO operations at the constructor.
FileBasedSink<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io
Abstract class for file-based output.
FileBasedSink(ValueProvider<ResourceId>, FileBasedSink.DynamicDestinations<?, DestinationT, OutputT>) - Constructor for class org.apache.beam.sdk.io.FileBasedSink
Construct a FileBasedSink with the given temp directory, producing uncompressed files.
FileBasedSink(ValueProvider<ResourceId>, FileBasedSink.DynamicDestinations<?, DestinationT, OutputT>, FileBasedSink.WritableByteChannelFactory) - Constructor for class org.apache.beam.sdk.io.FileBasedSink
Construct a FileBasedSink with the given temp directory and output channel type.
FileBasedSink(ValueProvider<ResourceId>, FileBasedSink.DynamicDestinations<?, DestinationT, OutputT>, Compression) - Constructor for class org.apache.beam.sdk.io.FileBasedSink
Construct a FileBasedSink with the given temp directory and output channel type.
FileBasedSink.CompressionType - Enum in org.apache.beam.sdk.io
Deprecated.
FileBasedSink.DynamicDestinations<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io
A class that allows value-dependent writes in FileBasedSink.
FileBasedSink.FilenamePolicy - Class in org.apache.beam.sdk.io
A naming policy for output files.
FileBasedSink.FileResult<DestinationT> - Class in org.apache.beam.sdk.io
Result of a single bundle write.
FileBasedSink.FileResultCoder<DestinationT> - Class in org.apache.beam.sdk.io
A coder for FileBasedSink.FileResult objects.
FileBasedSink.OutputFileHints - Interface in org.apache.beam.sdk.io
Provides hints about how to generate output files, such as a suggested filename suffix (e.g.
FileBasedSink.WritableByteChannelFactory - Interface in org.apache.beam.sdk.io
Implementations create instances of WritableByteChannel used by FileBasedSink and related classes to allow decorating, or otherwise transforming, the raw data that would normally be written directly to the WritableByteChannel passed into FileBasedSink.WritableByteChannelFactory.create(WritableByteChannel).
FileBasedSink.WriteOperation<DestinationT,OutputT> - Class in org.apache.beam.sdk.io
Abstract operation that manages the process of writing to FileBasedSink.
FileBasedSink.Writer<DestinationT,OutputT> - Class in org.apache.beam.sdk.io
Abstract writer that writes a bundle to a FileBasedSink.
FileBasedSource<T> - Class in org.apache.beam.sdk.io
A common base class for all file-based Sources.
FileBasedSource(ValueProvider<String>, EmptyMatchTreatment, long) - Constructor for class org.apache.beam.sdk.io.FileBasedSource
Create a FileBaseSource based on a file or a file pattern specification, with the given strategy for treating filepatterns that do not match any files.
FileBasedSource(ValueProvider<String>, long) - Constructor for class org.apache.beam.sdk.io.FileBasedSource
FileBasedSource(MatchResult.Metadata, long, long, long) - Constructor for class org.apache.beam.sdk.io.FileBasedSource
Create a FileBasedSource based on a single file.
FileBasedSource.FileBasedReader<T> - Class in org.apache.beam.sdk.io
A reader that implements code common to readers of FileBasedSources.
FileBasedSource.Mode - Enum in org.apache.beam.sdk.io
A given FileBasedSource represents a file resource of one of these types.
FileChecksumMatcher - Class in org.apache.beam.sdk.testing
Matcher to verify checksum of the contents of an ShardedFile in E2E test.
fileContentsHaveChecksum(String) - Static method in class org.apache.beam.sdk.testing.FileChecksumMatcher
 
FileIO - Class in org.apache.beam.sdk.io
General-purpose transforms for working with files: listing files (matching), reading and writing.
FileIO() - Constructor for class org.apache.beam.sdk.io.FileIO
 
FileIO.Match - Class in org.apache.beam.sdk.io
Implementation of FileIO.match().
FileIO.MatchAll - Class in org.apache.beam.sdk.io
Implementation of FileIO.matchAll().
FileIO.MatchConfiguration - Class in org.apache.beam.sdk.io
Describes configuration for matching filepatterns, such as EmptyMatchTreatment and continuous watching for matching files.
FileIO.ReadableFile - Class in org.apache.beam.sdk.io
A utility class for accessing a potentially compressed file.
FileIO.ReadMatches - Class in org.apache.beam.sdk.io
Implementation of FileIO.readMatches().
FileIO.ReadMatches.DirectoryTreatment - Enum in org.apache.beam.sdk.io
Enum to control how directories are handled.
FileIO.Sink<ElementT> - Interface in org.apache.beam.sdk.io
Specifies how to write elements to individual files in FileIO.write() and FileIO.writeDynamic().
FileIO.Write<DestinationT,UserT> - Class in org.apache.beam.sdk.io
Implementation of FileIO.write() and FileIO.writeDynamic().
FileIO.Write.FileNaming - Interface in org.apache.beam.sdk.io
A policy for generating names for shard files.
FilenamePolicy() - Constructor for class org.apache.beam.sdk.io.FileBasedSink.FilenamePolicy
 
filepattern(String) - Method in class org.apache.beam.sdk.io.FileIO.Match
Matches the given filepattern.
filepattern(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.FileIO.Match
filepattern(String) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Parse
Matches the given filepattern.
filepattern(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Parse
FileReporter - Class in org.apache.beam.runners.flink.metrics
Flink metrics reporter for writing metrics to a file specified via the "metrics.reporter.file.path" config key (assuming an alias of "file" for this reporter in the "metrics.reporters" setting).
FileReporter() - Constructor for class org.apache.beam.runners.flink.metrics.FileReporter
 
FileResult(ResourceId, int, BoundedWindow, PaneInfo, DestinationT) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.FileResult
 
FileResultCoder(Coder<BoundedWindow>, Coder<DestinationT>) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
 
fileSize(GcsPath) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
Returns the file size from GCS or throws FileNotFoundException if the resource does not exist.
FileSystem<ResourceIdT extends ResourceId> - Class in org.apache.beam.sdk.io
File system interface in Beam.
FileSystem() - Constructor for class org.apache.beam.sdk.io.FileSystem
 
FileSystemRegistrar - Interface in org.apache.beam.sdk.io
A registrar that creates FileSystem instances from PipelineOptions.
FileSystems - Class in org.apache.beam.sdk.io
Clients facing FileSystem utility.
FileSystems() - Constructor for class org.apache.beam.sdk.io.FileSystems
 
Filter - Class in org.apache.beam.sdk.schemas.transforms
A PTransform for filtering a collection of schema types.
Filter() - Constructor for class org.apache.beam.sdk.schemas.transforms.Filter
 
Filter<T> - Class in org.apache.beam.sdk.transforms
PTransforms for filtering from a PCollection the elements satisfying a predicate, or satisfying an inequality with a given value based on the elements' natural ordering.
Filter.Inner<T> - Class in org.apache.beam.sdk.schemas.transforms
Implementation of the filter.
filterCharacters(String) - Method in class org.apache.beam.runners.flink.metrics.FileReporter
 
finalizeAllOutstandingBundles() - Method in class org.apache.beam.runners.fnexecution.control.BundleFinalizationHandlers.InMemoryFinalizer
All finalization requests will be sent without waiting for the responses.
finalizeCheckpoint() - Method in class org.apache.beam.runners.spark.io.EmptyCheckpointMark
 
finalizeCheckpoint() - Method in class org.apache.beam.sdk.io.CountingSource.CounterMark
 
finalizeCheckpoint() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark
 
finalizeCheckpoint() - Method in interface org.apache.beam.sdk.io.UnboundedSource.CheckpointMark
Called by the system to signal that this checkpoint mark has been committed along with all the records which have been read from the UnboundedSource.UnboundedReader since the previous checkpoint was taken.
finalizeCheckpoint() - Method in class org.apache.beam.sdk.io.UnboundedSource.CheckpointMark.NoopCheckpointMark
 
finalizeDestination(DestinationT, BoundedWindow, Integer, Collection<FileBasedSink.FileResult<DestinationT>>) - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
 
find(String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.Find PTransform that checks if a portion of the line matches the Regex.
find(Pattern) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.Find PTransform that checks if a portion of the line matches the Regex.
find(String, int) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.Find PTransform that checks if a portion of the line matches the Regex.
find(Pattern, int) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.Find PTransform that checks if a portion of the line matches the Regex.
find(String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.FindName PTransform that checks if a portion of the line matches the Regex.
find(Pattern, String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.FindName PTransform that checks if a portion of the line matches the Regex.
Find(Pattern, int) - Constructor for class org.apache.beam.sdk.transforms.Regex.Find
 
findAll(String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.FindAll PTransform that checks if a portion of the line matches the Regex.
findAll(Pattern) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.FindAll PTransform that checks if a portion of the line matches the Regex.
FindAll(Pattern) - Constructor for class org.apache.beam.sdk.transforms.Regex.FindAll
 
findDateTimePattern(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
findDateTimePattern(String, ImmutableMap<Enum, DateTimeFormatter>) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
findKV(String, int, int) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.FindKV PTransform that checks if a portion of the line matches the Regex.
findKV(Pattern, int, int) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.FindKV PTransform that checks if a portion of the line matches the Regex.
findKV(String, String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.FindNameKV PTransform that checks if a portion of the line matches the Regex.
findKV(Pattern, String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.FindNameKV PTransform that checks if a portion of the line matches the Regex.
FindKV(Pattern, int, int) - Constructor for class org.apache.beam.sdk.transforms.Regex.FindKV
 
FindName(Pattern, String) - Constructor for class org.apache.beam.sdk.transforms.Regex.FindName
 
FindNameKV(Pattern, String, String) - Constructor for class org.apache.beam.sdk.transforms.Regex.FindNameKV
 
FindQuery - Class in org.apache.beam.sdk.io.mongodb
Builds a MongoDB FindQuery object.
FindQuery() - Constructor for class org.apache.beam.sdk.io.mongodb.FindQuery
 
finishBundle() - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
 
finishBundle() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write.PubsubBoundedWriter
 
finishBundle() - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
FinishBundleContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.FinishBundleContext
 
finishSpecifying() - Method in interface org.apache.beam.sdk.state.StateSpec
For internal use only; no backwards-compatibility guarantees.
finishSpecifying(PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PCollection
After building, finalizes this PValue to make it ready for running.
finishSpecifying(PInput, PTransform<?, ?>) - Method in interface org.apache.beam.sdk.values.PValue
After building, finalizes this PValue to make it ready for being used as an input to a PTransform.
finishSpecifying(PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PValueBase
 
finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
 
finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteResult
 
finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.io.WriteFilesResult
 
finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.transforms.WithFailures.Result
 
finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PCollection
 
finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PCollectionList
 
finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
 
finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PDone
Does nothing; there is nothing to finish specifying.
finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in interface org.apache.beam.sdk.values.POutput
As part of applying the producing PTransform, finalizes this output to make it ready for being used as an input and for running.
finishSpecifyingOutput(String, PInput, PTransform<?, ?>) - Method in class org.apache.beam.sdk.values.PValueBase
 
finishWrite() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
fireEligibleTimers(InMemoryTimerInternals, Map<KV<String, String>, FnDataReceiver<Timer>>, Object) - Static method in class org.apache.beam.runners.fnexecution.translation.PipelineTranslatorUtils
Fires all timers which are ready to be fired.
fireTimestamp() - Method in class org.apache.beam.sdk.transforms.DoFn.OnTimerContext
Returns the output timestamp of the current timer.
fixDefaults() - Method in class org.apache.beam.sdk.values.WindowingStrategy
Fixes all the defaults so that equals can be used to check that two strategies are the same, regardless of the state of "defaulted-ness".
FixedBytes - Class in org.apache.beam.sdk.schemas.logicaltypes
A LogicalType representing a fixed-size byte array.
fixedSizeGlobally(int) - Static method in class org.apache.beam.sdk.transforms.Sample
Returns a PTransform that takes a PCollection<T>, selects sampleSize elements, uniformly at random, and returns a PCollection<Iterable<T>> containing the selected elements.
fixedSizePerKey(int) - Static method in class org.apache.beam.sdk.transforms.Sample
Returns a PTransform that takes an input PCollection<KV<K, V>> and returns a PCollection<KV<K, Iterable<V>>> that contains an output element mapping each distinct key in the input PCollection to a sample of sampleSize values associated with that key in the input PCollection, taken uniformly at random.
fixedString(int) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
fixedStringSize() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
FixedWindows - Class in org.apache.beam.sdk.transforms.windowing
A WindowFn that windows values into fixed-size timestamp-based windows.
flatMap(List<TranslatorUtils.RawUnionValue>) - Method in class org.apache.beam.runners.gearpump.translators.functions.DoFnFunction
 
flatMap(WindowedValue<T>) - Method in class org.apache.beam.runners.gearpump.translators.WindowAssignTranslator.AssignWindows
 
FlatMapElements<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
PTransforms for mapping a simple function that returns iterables over the elements of a PCollection and merging the results.
FlatMapElements.FlatMapWithFailures<InputT,OutputT,FailureT> - Class in org.apache.beam.sdk.transforms
A PTransform that adds exception handling to FlatMapElements.
Flatten - Class in org.apache.beam.sdk.transforms
Flatten<T> takes multiple PCollection<T>s bundled into a PCollectionList<T> and returns a single PCollection<T> containing all the elements in all the input PCollections.
Flatten() - Constructor for class org.apache.beam.sdk.transforms.Flatten
 
Flatten.Iterables<T> - Class in org.apache.beam.sdk.transforms
FlattenIterables<T> takes a PCollection<Iterable<T>> and returns a PCollection<T> that contains all the elements from each iterable.
Flatten.PCollections<T> - Class in org.apache.beam.sdk.transforms
A PTransform that flattens a PCollectionList into a PCollection containing all the elements of all the PCollections in its input.
Flattened() - Constructor for class org.apache.beam.sdk.schemas.transforms.Select.Flattened
 
flattenedSchema() - Static method in class org.apache.beam.sdk.schemas.transforms.Select
Selects every leaf-level field.
FlattenP - Class in org.apache.beam.runners.jet.processors
Jet Processor implementation for Beam's Flatten primitive.
FlattenP.Supplier - Class in org.apache.beam.runners.jet.processors
Jet Processor supplier that will provide instances of FlattenP.
FlattenPCollectionsTranslator<T> - Class in org.apache.beam.runners.gearpump.translators
Flatten.FlattenPCollectionList is translated to Gearpump merge function.
FlattenPCollectionsTranslator() - Constructor for class org.apache.beam.runners.gearpump.translators.FlattenPCollectionsTranslator
 
flattenRel(RelStructuredTypeFlattener) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
 
FlattenWithHeterogeneousCoders - Interface in org.apache.beam.sdk.testing
Category tag for tests that use a Flatten where the input PCollectionList contains PCollections heterogeneous coders.
FlinkBatchPortablePipelineTranslator - Class in org.apache.beam.runners.flink
A translator that translates bounded portable pipelines into executable Flink pipelines.
FlinkBatchPortablePipelineTranslator.BatchTranslationContext - Class in org.apache.beam.runners.flink
Batch translation context.
FlinkBatchPortablePipelineTranslator.IsFlinkNativeTransform - Class in org.apache.beam.runners.flink
Predicate to determine whether a URN is a Flink native transform.
FlinkDetachedRunnerResult - Class in org.apache.beam.runners.flink
Result of a detached execution of a Pipeline with Flink.
FlinkExecutionEnvironments - Class in org.apache.beam.runners.flink
Utilities for Flink execution environments.
FlinkExecutionEnvironments() - Constructor for class org.apache.beam.runners.flink.FlinkExecutionEnvironments
 
FlinkJobInvoker - Class in org.apache.beam.runners.flink
Job Invoker for the FlinkRunner.
FlinkJobInvoker(FlinkJobServerDriver.FlinkServerConfiguration) - Constructor for class org.apache.beam.runners.flink.FlinkJobInvoker
 
FlinkJobServerDriver - Class in org.apache.beam.runners.flink
Driver program that starts a job server for the Flink runner.
FlinkJobServerDriver.FlinkServerConfiguration - Class in org.apache.beam.runners.flink
Flink runner-specific Configuration for the jobServer.
FlinkMetricContainer - Class in org.apache.beam.runners.flink.metrics
Helper class for holding a MetricsContainerImpl and forwarding Beam metrics to Flink accumulators and metrics.
FlinkMetricContainer(RuntimeContext) - Constructor for class org.apache.beam.runners.flink.metrics.FlinkMetricContainer
 
FlinkMetricContainer.FlinkDistributionGauge - Class in org.apache.beam.runners.flink.metrics
Flink Gauge for DistributionResult.
FlinkMetricContainer.FlinkGauge - Class in org.apache.beam.runners.flink.metrics
Flink Gauge for GaugeResult.
FlinkMiniClusterEntryPoint - Class in org.apache.beam.runners.flink
Entry point for starting an embedded Flink cluster.
FlinkMiniClusterEntryPoint() - Constructor for class org.apache.beam.runners.flink.FlinkMiniClusterEntryPoint
 
FlinkPipelineOptions - Interface in org.apache.beam.runners.flink
Options which can be used to configure a Flink PortablePipelineRunner.
FlinkPipelineRunner - Class in org.apache.beam.runners.flink
Runs a Pipeline on Flink via FlinkRunner.
FlinkPipelineRunner(FlinkPipelineOptions, String, List<String>) - Constructor for class org.apache.beam.runners.flink.FlinkPipelineRunner
 
FlinkPortableClientEntryPoint - Class in org.apache.beam.runners.flink
Flink job entry point to launch a Beam pipeline by executing an external SDK driver program.
FlinkPortableClientEntryPoint(String) - Constructor for class org.apache.beam.runners.flink.FlinkPortableClientEntryPoint
 
FlinkPortablePipelineTranslator<T extends FlinkPortablePipelineTranslator.TranslationContext> - Interface in org.apache.beam.runners.flink
Interface for portable Flink translators.
FlinkPortablePipelineTranslator.Executor - Interface in org.apache.beam.runners.flink
A handle used to execute a translated pipeline.
FlinkPortablePipelineTranslator.TranslationContext - Interface in org.apache.beam.runners.flink
The context used for pipeline translation.
FlinkPortableRunnerResult - Class in org.apache.beam.runners.flink
Result of executing a portable Pipeline with Flink.
FlinkRunner - Class in org.apache.beam.runners.flink
A PipelineRunner that executes the operations in the pipeline by first translating them to a Flink Plan and then executing them either locally or on a Flink cluster, depending on the configuration.
FlinkRunner(FlinkPipelineOptions) - Constructor for class org.apache.beam.runners.flink.FlinkRunner
 
FlinkRunnerRegistrar - Class in org.apache.beam.runners.flink
AutoService registrar - will register FlinkRunner and FlinkOptions as possible pipeline runner services.
FlinkRunnerRegistrar.Options - Class in org.apache.beam.runners.flink
Pipeline options registrar.
FlinkRunnerRegistrar.Runner - Class in org.apache.beam.runners.flink
Pipeline runner registrar.
FlinkRunnerResult - Class in org.apache.beam.runners.flink
Result of executing a Pipeline with Flink.
FlinkServerConfiguration() - Constructor for class org.apache.beam.runners.flink.FlinkJobServerDriver.FlinkServerConfiguration
 
FlinkStateBackendFactory - Interface in org.apache.beam.runners.flink
Constructs a StateBackend to use from flink pipeline options.
FlinkStreamingPortablePipelineTranslator - Class in org.apache.beam.runners.flink
Translate an unbounded portable pipeline representation into a Flink pipeline representation.
FlinkStreamingPortablePipelineTranslator.IsFlinkNativeTransform - Class in org.apache.beam.runners.flink
Predicate to determine whether a URN is a Flink native transform.
FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext - Class in org.apache.beam.runners.flink
Streaming translation context.
FLOAT - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
FLOAT - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
The type of float fields.
FLOAT32 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
FLOAT64 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
FloatCoder - Class in org.apache.beam.sdk.coders
A FloatCoder encodes Float values in 4 bytes using Java serialization.
floats() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for Float.
flush(boolean) - Method in class org.apache.beam.runners.jet.metrics.JetMetricsContainer
 
flush() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataSizeBasedBufferingOutboundObserver
 
flush() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataTimeBasedBufferingOutboundObserver
 
flush() - Method in interface org.apache.beam.sdk.fn.data.CloseableFnDataReceiver
Deprecated.
to be removed once splitting/checkpointing are available in SDKs and rewinding in readers.
flush() - Method in class org.apache.beam.sdk.io.AvroIO.Sink
 
flush() - Method in interface org.apache.beam.sdk.io.FileIO.Sink
Flushes the buffered state (if any) before the channel is closed.
flush() - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Sink
 
flush() - Method in class org.apache.beam.sdk.io.TextIO.Sink
 
flush() - Method in class org.apache.beam.sdk.io.TFRecordIO.Sink
 
flush() - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.Sink
 
flush() - Method in class org.apache.beam.sdk.io.xml.XmlIO.Sink
 
fn(ProcessFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Contextful
fn(SerializableFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Contextful
Binary compatibility adapter for Contextful.fn(ProcessFunction).
fn(Contextful.Fn<InputT, OutputT>, Requirements) - Static method in class org.apache.beam.sdk.transforms.Contextful
Same with Contextful.of(ClosureT, org.apache.beam.sdk.transforms.Requirements) but with better type inference behavior for the case of Contextful.Fn.
FnApiControlClient - Class in org.apache.beam.runners.fnexecution.control
A client for the control plane of an SDK harness, which can issue requests to it over the Fn API.
FnApiControlClientPoolService - Class in org.apache.beam.runners.fnexecution.control
A Fn API control service which adds incoming SDK harness connections to a sink.
FnDataReceiver<T> - Interface in org.apache.beam.sdk.fn.data
A receiver of streamed data.
FnDataService - Interface in org.apache.beam.runners.fnexecution.data
The FnDataService is able to forward inbound elements to a consumer and is also a consumer of outbound elements.
FnService - Interface in org.apache.beam.runners.fnexecution
An interface sharing common behavior with services used during execution of user Fns.
fold(KV<Instant, WindowedValue<KV<K, List<V>>>>, KV<Instant, WindowedValue<KV<K, V>>>) - Method in class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator.Merge
 
forBagUserStateHandlerFactory(ProcessBundleDescriptors.ExecutableProcessBundleDescriptor, StateRequestHandlers.BagUserStateHandlerFactory) - Static method in class org.apache.beam.runners.fnexecution.state.StateRequestHandlers
forBatch(PCollectionView<ViewT>) - Static method in class org.apache.beam.runners.dataflow.CreateDataflowView
 
forBoolean(Boolean) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
Constructs a CloudObject representing the given value.
forBytes() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Init
Returns a HllCount.Init.Builder for a HllCount.Init combining PTransform that computes bytes-type HLL++ sketches.
forClass(Class<?>) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
Constructs a CloudObject to be used for serializing an instance of the supplied class for transport via the Dataflow API.
forClassName(String) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
Constructs a CloudObject to be used for serializing data to be deserialized using the supplied class name the supplied class name for transport via the Dataflow API.
forCoder(TypeDescriptor<?>, Coder<?>) - Static method in class org.apache.beam.sdk.coders.CoderProviders
Creates a CoderProvider that always returns the given coder for the specified type.
forConsumer(LogicalEndpoint, Coder<T>, FnDataReceiver<T>) - Static method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
 
forDescriptor(ProtoDomain, String) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoDynamicMessageSchema
Create a new ProtoDynamicMessageSchema from a ProtoDomain and for a message.
forDescriptor(ProtoDomain, Descriptors.Descriptor) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoDynamicMessageSchema
Create a new ProtoDynamicMessageSchema from a ProtoDomain and for a descriptor.
forDescriptor(Endpoints.ApiServiceDescriptor) - Method in class org.apache.beam.sdk.fn.channel.ManagedChannelFactory
 
forEncoding(ByteString) - Static method in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow
 
forever(Trigger) - Static method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
Create a composite trigger that repeatedly executes the trigger repeated, firing each time it fires and ignoring any indications to finish.
forField(Field) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
 
forFloat(Float) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
Constructs a CloudObject representing the given value.
forFloat(Double) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
Constructs a CloudObject representing the given value.
forGetter(Method) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
 
forHandler(RunnerApi.Environment, InstructionRequestHandler) - Static method in interface org.apache.beam.runners.fnexecution.environment.RemoteEnvironment
Create a new RemoteEnvironment for the provided RunnerApi.Environment and AutoCloseable InstructionRequestHandler.
forInteger(Long) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
Constructs a CloudObject representing the given value.
forInteger(Integer) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
Constructs a CloudObject representing the given value.
forIntegers() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Init
Returns a HllCount.Init.Builder for a HllCount.Init combining PTransform that computes integer-type HLL++ sketches.
forIterableSideInput(String, String, Coder<V>, Coder<W>) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.SideInputHandlerFactory
Returns an StateRequestHandlers.IterableSideInputHandler for the given pTransformId, sideInputId.
forIterableSideInput(String, String, Coder<V>, Coder<W>) - Method in class org.apache.beam.runners.fnexecution.translation.BatchSideInputHandlerFactory
 
forKnownType(Object) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
Constructs a CloudObject representing the given value of a well-known cloud object type.
forLocation(PipelineOptions, LogicalEndpoint, Coder<T>, StreamObserver<BeamFnApi.Elements>) - Static method in interface org.apache.beam.sdk.fn.data.BeamFnDataBufferingOutboundObserver
 
forLongs() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Init
Returns a HllCount.Init.Builder for a HllCount.Init combining PTransform that computes long-type HLL++ sketches.
FORMAT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
FormatAsTextFn() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.FormatAsTextFn
 
formatRecord(ElementT, Schema) - Method in interface org.apache.beam.sdk.io.AvroIO.RecordFormatter
Deprecated.
 
formatRecord(UserT) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
Convert an input record type into the output type.
formatTimestamp(Instant) - Static method in class org.apache.beam.sdk.transforms.windowing.BoundedWindow
Formats a Instant timestamp with additional Beam-specific metadata, such as indicating whether the timestamp is the end of the global window or one of the distinguished values BoundedWindow.TIMESTAMP_MIN_VALUE or BoundedWindow.TIMESTAMP_MIN_VALUE.
formatTimestampWithTimeZone(DateTime) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
forMultimapSideInput(String, String, KvCoder<K, V>, Coder<W>) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.SideInputHandlerFactory
Returns a StateRequestHandlers.MultimapSideInputHandler for the given pTransformId, sideInputId.
forMultimapSideInput(String, String, KvCoder<K, V>, Coder<W>) - Method in class org.apache.beam.runners.fnexecution.translation.BatchSideInputHandlerFactory
 
forNewInput(Instant, InputT) - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
Called by the Watch transform to create a new independent termination state for a newly arrived InputT.
forOneOf(String, boolean, Map<String, FieldValueTypeInformation>) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
 
forOrdinal(int) - Static method in enum org.apache.beam.sdk.io.kafka.KafkaTimestampType
 
forRegistry(MetricRegistry) - Static method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
 
forRegistry(MetricRegistry) - Static method in class org.apache.beam.runners.spark.structuredstreaming.metrics.WithMetricsSupport
 
forRequestObserver(String, StreamObserver<BeamFnApi.InstructionRequest>, ConcurrentMap<String, BeamFnApi.ProcessBundleDescriptor>) - Static method in class org.apache.beam.runners.fnexecution.control.FnApiControlClient
Returns a FnApiControlClient which will submit its requests to the provided observer.
forService(InstructionRequestHandler) - Static method in class org.apache.beam.runners.fnexecution.environment.StaticRemoteEnvironmentFactory
 
forSetter(Method) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
 
forSetter(Method, String) - Static method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
 
forSideInputHandlerFactory(Map<String, Map<String, ProcessBundleDescriptors.SideInputSpec>>, StateRequestHandlers.SideInputHandlerFactory) - Static method in class org.apache.beam.runners.fnexecution.state.StateRequestHandlers
Returns an adapter which converts a StateRequestHandlers.SideInputHandlerFactory to a StateRequestHandler.
forSqlType(Schema.TypeName) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.BigDecimalConverter
 
forStage(ExecutableStage) - Method in class org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory
 
forStage(ExecutableStage) - Method in interface org.apache.beam.runners.fnexecution.control.JobBundleFactory
 
forStage(ExecutableStage) - Method in class org.apache.beam.runners.fnexecution.control.SingleEnvironmentInstanceJobBundleFactory
Deprecated.
 
forStage(ExecutableStage, BatchSideInputHandlerFactory.SideInputGetter) - Static method in class org.apache.beam.runners.fnexecution.translation.BatchSideInputHandlerFactory
Creates a new state handler for the given stage.
forStreamFromSources(List<Integer>, Map<Integer, GlobalWatermarkHolder.SparkWatermarks>) - Static method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
Build the TimerInternals according to the feeding streams.
forStreaming(PCollectionView<ViewT>) - Static method in class org.apache.beam.runners.dataflow.CreateDataflowView
 
forString(String) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
Constructs a CloudObject representing the given value.
forStrings() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Init
Returns a HllCount.Init.Builder for a HllCount.Init combining PTransform that computes string-type HLL++ sketches.
forTransformHierarchy(TransformHierarchy, PipelineOptions) - Static method in class org.apache.beam.sdk.Pipeline
 
forTypeName(Schema.TypeName) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
 
forUserState(String, String, Coder<K>, Coder<V>, Coder<W>) - Method in class org.apache.beam.runners.fnexecution.state.InMemoryBagUserStateFactory
 
forUserState(String, String, Coder<K>, Coder<V>, Coder<W>) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.BagUserStateHandlerFactory
 
ForwardingClientResponseObserver<ReqT,RespT> - Class in org.apache.beam.sdk.fn.stream
A ClientResponseObserver which delegates all StreamObserver calls.
forWriter(LogWriter) - Static method in class org.apache.beam.runners.fnexecution.logging.GrpcLoggingService
 
freeze() - Method in class org.apache.beam.runners.jet.metrics.JetMetricResults
 
from(Map<String, String>) - Static method in class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
Deprecated.
Expects a map keyed by logger Names with values representing Levels.
from(ExecutorService) - Static method in class org.apache.beam.sdk.fn.test.TestExecutors
 
from(Supplier<ExecutorService>) - Static method in class org.apache.beam.sdk.fn.test.TestExecutors
 
from(String) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
Reads from the given filename or filepattern.
from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.AvroIO.Read
Reads from the given filename or filepattern.
from(String) - Method in class org.apache.beam.sdk.io.AvroIO.Read
from(ValueProvider<String>) - Static method in class org.apache.beam.sdk.io.AvroSource
Reads from the given file name or pattern ("glob").
from(MatchResult.Metadata) - Static method in class org.apache.beam.sdk.io.AvroSource
 
from(String) - Static method in class org.apache.beam.sdk.io.AvroSource
from(FileBasedSource<T>) - Static method in class org.apache.beam.sdk.io.CompressedSource
Creates a CompressedSource from an underlying FileBasedSource.
from(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
Reads a BigQuery table specified as "[project_id]:[dataset_id].[table_id]" or "[dataset_id].[table_id]" for tables within the current project.
from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
Same as from(String), but with a ValueProvider.
from(TableReference) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
Read from table specified by a TableReference.
from(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
from(TableReference) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
from(long) - Static method in class org.apache.beam.sdk.io.GenerateSequence
Specifies the minimum number to generate (inclusive).
from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Read
Reads from the given filename or filepattern.
from(String) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Read
from(BoundedSource<T>) - Method in class org.apache.beam.sdk.io.Read.Builder
Returns a new Read.Bounded PTransform reading from the given BoundedSource.
from(UnboundedSource<T, ?>) - Method in class org.apache.beam.sdk.io.Read.Builder
Returns a new Read.Unbounded PTransform reading from the given UnboundedSource.
from(BoundedSource<T>) - Static method in class org.apache.beam.sdk.io.Read
Returns a new Read.Bounded PTransform reading from the given BoundedSource.
from(UnboundedSource<T, ?>) - Static method in class org.apache.beam.sdk.io.Read
Returns a new Read.Unbounded PTransform reading from the given UnboundedSource.
from(String) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Read
Provide name of collection while reading from Solr.
from(String) - Method in class org.apache.beam.sdk.io.TextIO.Read
Reads text files that reads from the file(s) with the given filename or filename pattern.
from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.TextIO.Read
Same as from(filepattern), but accepting a ValueProvider.
from(String) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
Returns a transform for reading TFRecord files that reads from the file(s) with the given filename or filename pattern.
from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
Same as from(filepattern), but accepting a ValueProvider.
from(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
Reads a single XML file or a set of XML files defined by a Java "glob" file pattern.
from(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
Reads a single XML file or a set of XML files defined by a Java "glob" file pattern.
from(Map<String, String>) - Static method in class org.apache.beam.sdk.options.SdkHarnessOptions.SdkHarnessLogLevelOverrides
Expects a map keyed by logger Names with values representing LogLevels.
from(HasDisplayData) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
Collect the DisplayData from a component.
fromArgs(String...) - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
Sets the command line arguments to parse when constructing the PipelineOptions.
fromArgs(String...) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
Sets the command line arguments to parse when constructing the PipelineOptions.
fromAvroType(Schema) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils.FixedBytesField
Create a AvroUtils.FixedBytesField from an AVRO type.
fromBeamCoder(Coder<T>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers
Wrap a Beam coder into a Spark Encoder using Catalyst Expression Encoders (which uses java code generation).
fromBeamFieldType(Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils.FixedBytesField
fromByteArray(byte[], Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
Utility method for deserializing a byte array using the specified coder.
fromByteArray(byte[], Coder<T>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.CoderHelpers
Utility method for deserializing a byte array using the specified coder.
fromByteArrays(Collection<byte[]>, Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
Utility method for deserializing a Iterable of byte arrays using the specified coder.
fromByteFunction(Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
A function wrapper for converting a byte array to an object.
FromByteFunction(Coder<K>, Coder<V>) - Constructor for class org.apache.beam.runners.spark.coders.CoderHelpers.FromByteFunction
 
fromByteFunctionIterable(Coder<K>, Coder<V>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
A function wrapper for converting a byte array pair to a key-value pair, where values are Iterable.
fromCanonical(Compression) - Static method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
Deprecated.
 
fromCloudDuration(String) - Static method in class org.apache.beam.runners.dataflow.util.TimeUtil
Converts a Dataflow API duration string into a Duration.
fromCloudObject(CloudObject) - Method in interface org.apache.beam.runners.dataflow.util.CloudObjectTranslator
Converts back into the original object from a provided CloudObject.
fromCloudObject(CloudObject) - Method in class org.apache.beam.runners.dataflow.util.SchemaCoderCloudObjectTranslator
Convert from a cloud object.
fromCloudTime(String) - Static method in class org.apache.beam.runners.dataflow.util.TimeUtil
Converts a time value received via the Dataflow API into the corresponding Instant.
fromComponents(String, String) - Static method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
Creates a GcsPath from bucket and object components.
fromConfig(FlinkJobServerDriver.FlinkServerConfiguration) - Static method in class org.apache.beam.runners.flink.FlinkJobServerDriver
 
fromConfig(FlinkJobServerDriver.FlinkServerConfiguration, JobServerDriver.JobInvokerFactory) - Static method in class org.apache.beam.runners.flink.FlinkJobServerDriver
 
fromExecutableStage(String, ExecutableStage, Endpoints.ApiServiceDescriptor, Endpoints.ApiServiceDescriptor) - Static method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors
Note that the BeamFnApi.ProcessBundleDescriptor is constructed by: Adding gRPC read and write nodes wiring them to the specified data endpoint.
fromExecutableStage(String, ExecutableStage, Endpoints.ApiServiceDescriptor) - Static method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors
 
fromExisting(Storage.Stream) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource
Creates a new source with the same properties as this one, except with a different Storage.Stream.
fromFile(File, boolean) - Static method in class org.apache.beam.sdk.io.LocalResources
 
fromHex(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
fromJsonString(String, Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
 
fromLong(Long, Schema, LogicalType) - Method in class org.apache.beam.sdk.coders.AvroCoder.JodaTimestampConversion
 
fromObject(StorageObject) - Static method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
Creates a GcsPath from a StorageObject.
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.apex.ApexRunner
 
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.apex.TestApexRunner
 
fromOptions(DataflowPipelineOptions) - Static method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator
Constructs a translator from the provided options.
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.dataflow.DataflowRunner
Construct a runner from the provided options.
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.dataflow.TestDataflowRunner
Constructs a runner from the provided options.
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.dataflow.util.GcsStager
 
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.direct.DirectRunner
Construct a DirectRunner from the provided options.
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.flink.FlinkRunner
Construct a runner from the provided options.
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.flink.TestFlinkRunner
 
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.gearpump.GearpumpRunner
 
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.gearpump.TestGearpumpRunner
 
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.jet.JetRunner
 
fromOptions(PipelineOptions, Function<ClientConfig, JetInstance>) - Static method in class org.apache.beam.runners.jet.JetRunner
 
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.portability.PortableRunner
Constructs a runner from the provided options.
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.portability.testing.TestPortableRunner
 
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.spark.SparkRunner
Creates and returns a new SparkRunner with specified options.
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.spark.SparkRunnerDebugger
 
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingRunner
Creates and returns a new SparkStructuredStreamingRunner with specified options.
fromOptions(PipelineOptions) - Static method in class org.apache.beam.runners.spark.TestSparkRunner
 
fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.extensions.gcp.auth.GcpCredentialFactory
 
fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.extensions.gcp.auth.NoopCredentialFactory
 
fromOptions(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsFileSystemRegistrar
 
fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
 
fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
 
fromOptions(PipelineOptions) - Method in class org.apache.beam.sdk.io.aws.s3.S3FileSystemRegistrar
 
fromOptions(PipelineOptions) - Method in interface org.apache.beam.sdk.io.FileSystemRegistrar
Create zero or more filesystems from the given PipelineOptions.
fromOptions(PipelineOptions) - Method in class org.apache.beam.sdk.io.hdfs.HadoopFileSystemRegistrar
 
fromOptions(PipelineOptions) - Method in class org.apache.beam.sdk.io.LocalFileSystemRegistrar
 
fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.PipelineRunner
Constructs a runner from the provided PipelineOptions.
fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.testing.CrashingRunner
 
fromOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.testing.TestPipeline
 
fromParams(String[]) - Static method in class org.apache.beam.runners.flink.FlinkJobServerDriver
 
fromParams(DefaultFilenamePolicy.Params) - Static method in class org.apache.beam.sdk.io.DefaultFilenamePolicy
fromPath(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription
Creates a class representing a Pub/Sub subscription from the specified subscription path.
fromPath(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic
Creates a class representing a Cloud Pub/Sub topic from the specified topic path.
fromPath(Path, boolean) - Static method in class org.apache.beam.sdk.io.LocalResources
 
fromProcessFunctionWithOutputType(ProcessFunction<InputT, OutputT>, TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.transforms.InferableFunction
 
fromPTransform(RunnerApi.PTransform) - Static method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortRead
 
fromPTransform(RunnerApi.PTransform) - Static method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortWrite
 
fromQuery(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
Reads results received after executing the given query.
fromQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
Same as fromQuery(String), but with a ValueProvider.
fromQuery(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
fromQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
fromRawEvents(Coder<T>, List<TestStream.Event<T>>) - Static method in class org.apache.beam.sdk.testing.TestStream
For internal use only.
FromRawUnionValue() - Constructor for class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils.FromRawUnionValue
 
fromResourceName(String) - Static method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
Creates a GcsPath from a OnePlatform resource name in string form.
fromRow(Row) - Static method in class org.apache.beam.sdk.values.Row
Creates a row builder based on the specified row.
fromRowFunction(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.annotations.DefaultSchema.DefaultSchemaProvider
Given a type, returns a function that converts from a Row object to that type.
fromRowFunction(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
 
fromRowFunction(TypeDescriptor<T>) - Method in interface org.apache.beam.sdk.schemas.SchemaProvider
Given a type, returns a function that converts from a Row object to that type.
fromRows(Class<OutputT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Convert
Convert a PCollection<Row> into a PCollection<OutputT>.
fromRows(TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Convert
Convert a PCollection<Row> into a PCollection<Row>.
fromSerializableFunctionWithOutputType(SerializableFunction<InputT, OutputT>, TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.transforms.SimpleFunction
 
fromSpec(Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.CloudObject
Constructs a CloudObject by copying the supplied serialized object spec, which must represent an SDK object serialized for transport via the Dataflow API.
fromSpec(HCatalogIO.Read) - Static method in class org.apache.beam.sdk.io.hcatalog.HCatToRow
Wraps the HCatalogIO.read() to convert HCatRecords to Rows.
fromStandardParameters(ValueProvider<ResourceId>, String, String, boolean) - Static method in class org.apache.beam.sdk.io.DefaultFilenamePolicy
fromStaticMethods(Class<?>, Class<?>) - Static method in class org.apache.beam.sdk.coders.CoderProviders
Creates a CoderProvider from a class's static <T> Coder<T> of(TypeDescriptor<T>, List<Coder<?>>) method.
fromString(String, boolean) - Static method in class org.apache.beam.sdk.io.LocalResources
 
fromString(ValueProvider<String>, boolean) - Static method in class org.apache.beam.sdk.io.LocalResources
 
fromSubscription(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
Reads from the given subscription.
fromSubscription(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
Like subscription() but with a ValueProvider.
fromSupplier(SerializableMatchers.SerializableSupplier<Matcher<T>>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
Constructs a SerializableMatcher from a non-serializable Matcher via indirection through SerializableMatchers.SerializableSupplier.
fromTableConfig(PubsubJsonTableProvider.PubsubIOTableConfiguration) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.RowToPubsubMessage
 
fromTableSchema(TableSchema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
Convert a BigQuery TableSchema to a Beam Schema.
fromTopic(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
Creates and returns a transform for reading from a Cloud Pub/Sub topic.
fromTopic(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
Like topic() but with a ValueProvider.
fromUri(URI) - Static method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
Creates a GcsPath from a URI.
fromUri(String) - Static method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
Creates a GcsPath from a URI in string form.
FullNameTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider
Base class for table providers that look up table metadata using full table names, instead of querying it by parts of the name separately.
FullNameTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.FullNameTableProvider
 
fullOuterJoin(String, PCollection<KV<K, V1>>, PCollection<KV<K, V2>>, V1, V2) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
Full Outer Join of two collections of KV elements.
fullOuterJoin(PCollection<KV<K, V1>>, PCollection<KV<K, V2>>, V1, V2) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
Full Outer Join of two collections of KV elements.
fullOuterJoin(PCollection<RhsT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Join
Perform a full outer join.
fullPublishResult() - Static method in class org.apache.beam.sdk.io.aws.sns.PublishResultCoders
Returns a new PublishResult coder which serializes the sdkResponseMetadata and sdkHttpMetadata, including the HTTP response headers.
fullPublishResultWithoutHeaders() - Static method in class org.apache.beam.sdk.io.aws.sns.PublishResultCoders
Returns a new PublishResult coder which serializes the sdkResponseMetadata and sdkHttpMetadata, but does not include the HTTP response headers.

G

Gauge - Interface in org.apache.beam.sdk.metrics
A metric that reports the latest value out of reported values.
gauge(String, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
Create a metric that can have its new value set, and is aggregated by taking the last reported value.
gauge(Class<?>, String) - Static method in class org.apache.beam.sdk.metrics.Metrics
Create a metric that can have its new value set, and is aggregated by taking the last reported value.
GaugeImpl - Class in org.apache.beam.runners.jet.metrics
Implementation of Gauge.
GaugeImpl(MetricName) - Constructor for class org.apache.beam.runners.jet.metrics.GaugeImpl
 
GaugeResult - Class in org.apache.beam.sdk.metrics
The result of a Gauge metric.
GaugeResult() - Constructor for class org.apache.beam.sdk.metrics.GaugeResult
 
GaugeResult.EmptyGaugeResult - Class in org.apache.beam.sdk.metrics
Empty GaugeResult, representing no values reported.
GcpCredentialFactory - Class in org.apache.beam.sdk.extensions.gcp.auth
Construct an oauth credential to be used by the SDK and the SDK workers.
GcpCredentialFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.auth.GcpCredentialFactory
 
GcpIoPipelineOptionsRegistrar - Class in org.apache.beam.sdk.io.gcp.common
A registrar containing the default GCP options.
GcpIoPipelineOptionsRegistrar() - Constructor for class org.apache.beam.sdk.io.gcp.common.GcpIoPipelineOptionsRegistrar
 
GcpOptions - Interface in org.apache.beam.sdk.extensions.gcp.options
Options used to configure Google Cloud Platform specific options such as the project and credentials.
GcpOptions.DefaultProjectFactory - Class in org.apache.beam.sdk.extensions.gcp.options
Attempts to infer the default project based upon the environment this application is executing within.
GcpOptions.EnableStreamingEngineFactory - Class in org.apache.beam.sdk.extensions.gcp.options
EneableStreamingEngine defaults to false unless one of the two experiments is set.
GcpOptions.GcpTempLocationFactory - Class in org.apache.beam.sdk.extensions.gcp.options
Returns PipelineOptions.getTempLocation() as the default GCP temp location.
GcpOptions.GcpUserCredentialsFactory - Class in org.apache.beam.sdk.extensions.gcp.options
Attempts to load the GCP credentials.
GcpPipelineOptionsRegistrar - Class in org.apache.beam.sdk.extensions.gcp.options
A registrar containing the default GCP options.
GcpPipelineOptionsRegistrar() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcpPipelineOptionsRegistrar
 
GcpTempLocationFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpTempLocationFactory
 
GcpUserCredentialsFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpUserCredentialsFactory
 
GCS_URI - Static variable in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
Pattern that is used to parse a GCS URL.
GcsCreateOptions - Class in org.apache.beam.sdk.extensions.gcp.storage
An abstract class that contains common configuration options for creating resources.
GcsCreateOptions() - Constructor for class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions
 
GcsCreateOptions.Builder - Class in org.apache.beam.sdk.extensions.gcp.storage
A builder for GcsCreateOptions.
GcsFileSystemRegistrar - Class in org.apache.beam.sdk.extensions.gcp.storage
AutoService registrar for the GcsFileSystem.
GcsFileSystemRegistrar() - Constructor for class org.apache.beam.sdk.extensions.gcp.storage.GcsFileSystemRegistrar
 
GcsOptions - Interface in org.apache.beam.sdk.extensions.gcp.options
Options used to configure Google Cloud Storage.
GcsOptions.ExecutorServiceFactory - Class in org.apache.beam.sdk.extensions.gcp.options
Returns the default ExecutorService to use within the Apache Beam SDK.
GcsOptions.PathValidatorFactory - Class in org.apache.beam.sdk.extensions.gcp.options
Creates a PathValidator object using the class specified in GcsOptions.getPathValidatorClass().
GcsPath - Class in org.apache.beam.sdk.extensions.gcp.util.gcsfs
Implements the Java NIO Path API for Google Cloud Storage paths.
GcsPath(FileSystem, String, String) - Constructor for class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
Constructs a GcsPath.
GcsPathValidator - Class in org.apache.beam.sdk.extensions.gcp.storage
GCP implementation of PathValidator.
GcsResourceId - Class in org.apache.beam.sdk.extensions.gcp.storage
ResourceId implementation for Google Cloud Storage.
GcsStager - Class in org.apache.beam.runners.dataflow.util
Utility class for staging files to GCS.
gcsUploadBufferSizeBytes() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions
The buffer size (in bytes) to use when uploading files to GCS.
GcsUtil - Class in org.apache.beam.sdk.extensions.gcp.util
Provides operations on GCS.
GcsUtil.GcsUtilFactory - Class in org.apache.beam.sdk.extensions.gcp.util
This is a DefaultValueFactory able to create a GcsUtil using any transport flags specified on the PipelineOptions.
GcsUtil.StorageObjectOrIOException - Class in org.apache.beam.sdk.extensions.gcp.util
A class that holds either a StorageObject or an IOException.
GcsUtilFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.GcsUtilFactory
 
GearpumpPipelineOptions - Interface in org.apache.beam.runners.gearpump
Options that configure the Gearpump pipeline.
GearpumpPipelineResult - Class in org.apache.beam.runners.gearpump
Result of executing a Pipeline with Gearpump.
GearpumpPipelineResult(ClientContext, RunningApplication) - Constructor for class org.apache.beam.runners.gearpump.GearpumpPipelineResult
 
GearpumpPipelineTranslator - Class in org.apache.beam.runners.gearpump.translators
GearpumpPipelineTranslator knows how to translate Pipeline objects into Gearpump Graph.
GearpumpPipelineTranslator(TranslationContext) - Constructor for class org.apache.beam.runners.gearpump.translators.GearpumpPipelineTranslator
 
GearpumpRunner - Class in org.apache.beam.runners.gearpump
A PipelineRunner that executes the operations in the pipeline by first translating them to Gearpump Stream DSL and then executing them on a Gearpump cluster.
GearpumpRunner(GearpumpPipelineOptions) - Constructor for class org.apache.beam.runners.gearpump.GearpumpRunner
 
GearpumpRunnerRegistrar - Class in org.apache.beam.runners.gearpump
Contains the PipelineRunnerRegistrar and PipelineOptionsRegistrar for the GearpumpRunner.
GearpumpRunnerRegistrar.Options - Class in org.apache.beam.runners.gearpump
GearpumpRunnerRegistrar.Runner - Class in org.apache.beam.runners.gearpump
Registers the GearpumpRunner.
GearpumpSource<T> - Class in org.apache.beam.runners.gearpump.translators.io
GearpumpWindowFn(boolean) - Constructor for class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator.GearpumpWindowFn
 
generate(Schema) - Static method in class org.apache.beam.sdk.coders.RowCoderGenerator
 
GenerateSequence - Class in org.apache.beam.sdk.io
A PTransform that produces longs starting from the given value, and either up to the given limit or until Long.MAX_VALUE / until the given time elapses.
GenerateSequence() - Constructor for class org.apache.beam.sdk.io.GenerateSequence
 
GenerateSequence.External - Class in org.apache.beam.sdk.io
Exposes GenerateSequence as an external transform for cross-language usage.
GenerateSequence.External.ExternalConfiguration - Class in org.apache.beam.sdk.io
Parameters class to expose the transform to an external SDK.
GenerateSequenceTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.seqgen
Sequence generator table provider.
GenerateSequenceTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.seqgen.GenerateSequenceTableProvider
 
generateStagingSessionToken(String, String) - Static method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService
Generate a stagingSessionToken compatible with BeamFileSystemLegacyArtifactStagingService.
GenericRecordReadConverter - Class in org.apache.beam.sdk.extensions.sql.meta.provider.parquet
A PTransform to convert GenericRecord to Row.
GenericRecordReadConverter() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.parquet.GenericRecordReadConverter
 
GenericRecordWriteConverter - Class in org.apache.beam.sdk.extensions.sql.meta.provider.avro
A PTransform to convert Row to GenericRecord.
GenericRecordWriteConverter() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.avro.GenericRecordWriteConverter
 
get(JobInfo) - Method in interface org.apache.beam.runners.fnexecution.control.ExecutableStageContext.Factory
Get or create ExecutableStageContext for given JobInfo.
get(JobInfo) - Method in class org.apache.beam.runners.fnexecution.control.ReferenceCountingExecutableStageContextFactory
 
get(K, W) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.BagUserStateHandler
Returns an Iterable of values representing the bag user state for the given key and window.
get(W) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.IterableSideInputHandler
Returns an Iterable of values representing the side input for the given window.
get(W) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.MultimapSideInputHandler
Returns an Iterable of keys representing the side input for the given window.
get(K, W) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.MultimapSideInputHandler
Returns an Iterable of values representing the side input for the given key and window.
get() - Method in class org.apache.beam.runners.portability.CloseableResource
Gets the underlying resource.
get() - Static method in class org.apache.beam.runners.spark.io.EmptyCheckpointMark
 
get(PCollectionView<T>, BoundedWindow) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.SparkSideInputReader
 
get(PCollectionView<T>, BoundedWindow) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.utils.CachedSideInputReader
 
get(PCollectionView<T>, BoundedWindow) - Method in class org.apache.beam.runners.spark.util.CachedSideInputReader
 
get(Long) - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
Returns the Broadcast containing the GlobalWatermarkHolder.SparkWatermarks mapped to their sources.
get(PCollectionView<T>, BoundedWindow) - Method in class org.apache.beam.runners.spark.util.SparkSideInputReader
 
get(int) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedList
 
get(Object) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedMap
 
get(int) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedRow
 
get() - Method in class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
 
get() - Method in interface org.apache.beam.sdk.options.ValueProvider
Returns the runtime value wrapped by this ValueProvider in case it is ValueProvider.isAccessible(), otherwise fails.
get() - Method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
 
get() - Method in class org.apache.beam.sdk.options.ValueProvider.RuntimeValueProvider
 
get() - Method in class org.apache.beam.sdk.options.ValueProvider.StaticValueProvider
 
get(Class<?>) - Method in class org.apache.beam.sdk.schemas.AutoValueSchema.AbstractGetterTypeSupplier
 
get(ObjectT) - Method in interface org.apache.beam.sdk.schemas.FieldValueGetter
 
get(Class<?>) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema.GetterTypeSupplier
 
get(Class<?>) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema.SetterTypeSupplier
 
get(Class<?>) - Method in class org.apache.beam.sdk.schemas.JavaFieldSchema.JavaFieldTypeSupplier
 
get(Object) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
get(Class<?>) - Method in interface org.apache.beam.sdk.schemas.utils.FieldValueTypeSupplier
Return all the FieldValueTypeInformations.
get(Class<?>, Schema) - Method in interface org.apache.beam.sdk.schemas.utils.FieldValueTypeSupplier
Return all the FieldValueTypeInformations.
get(K) - Method in interface org.apache.beam.sdk.state.MapState
A deferred lookup.
get(String) - Method in interface org.apache.beam.sdk.state.TimerMap
 
get() - Method in interface org.apache.beam.sdk.testing.SerializableMatchers.SerializableSupplier
 
get(TupleTag<V>) - Method in class org.apache.beam.sdk.transforms.CombineFns.CoCombineResult
Returns the value represented by the given TupleTag.
get(TupleTag<T>) - Method in interface org.apache.beam.sdk.transforms.DoFn.MultiOutputReceiver
Returns an DoFn.OutputReceiver for the given tag.
get() - Method in interface org.apache.beam.sdk.transforms.Materializations.IterableView
Returns an iterable for all values.
get() - Method in interface org.apache.beam.sdk.transforms.Materializations.MultimapView
Returns an iterable of all keys.
get(K) - Method in interface org.apache.beam.sdk.transforms.Materializations.MultimapView
Returns an iterable of all the values for the specified key.
get(int) - Method in class org.apache.beam.sdk.values.PCollectionList
Returns the PCollection at the given index (origin zero).
get(TupleTag<T>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
Returns the PCollection associated with the given TupleTag in this PCollectionTuple.
get(String) - Method in class org.apache.beam.sdk.values.PCollectionTuple
Returns the PCollection associated with the given tag in this PCollectionTuple.
get(int) - Method in class org.apache.beam.sdk.values.TupleTagList
Returns the TupleTag at the given index (origin zero).
getAccum() - Method in interface org.apache.beam.sdk.state.CombiningState
Read the merged accumulator for this state cell.
getAccumulatorCoder(CoderRegistry, Coder) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.CountMinSketchFn
 
getAccumulatorCoder(CoderRegistry, Coder) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.TDigestQuantilesFn
 
getAccumulatorCoder(CoderRegistry, Coder<Row>) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CovarianceFn
 
getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.VarianceFn
 
getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
 
getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn
 
getAccumulatorCoder(CoderRegistry, Coder<Double>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
 
getAccumulatorCoder(CoderRegistry, Coder<V>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
 
getAccumulatorCoder(CoderRegistry, Coder<Integer>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
 
getAccumulatorCoder(CoderRegistry, Coder<Long>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
 
getAccumulatorCoder(CoderRegistry, Coder<InputT>) - Method in interface org.apache.beam.sdk.transforms.CombineFnBase.GlobalCombineFn
Returns the Coder to use for accumulator AccumT values, or null if it is not able to be inferred.
getAccumulatorCoder(CoderRegistry, Coder<DataT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
 
getAccumulatorCoder(CoderRegistry, Coder<DataT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
 
getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
 
getAccumulatorCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
 
getAdditionalInputs() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.ParDoSingle
 
getAdditionalInputs() - Method in class org.apache.beam.sdk.io.WriteFiles
 
getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
Returns the side inputs of this Combine, tagged with the tag of the PCollectionView.
getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
Returns the side inputs of this Combine, tagged with the tag of the PCollectionView.
getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
Returns the side inputs of this ParDo, tagged with the tag of the PCollectionView.
getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
Returns the side inputs of this ParDo, tagged with the tag of the PCollectionView.
getAdditionalInputs() - Method in class org.apache.beam.sdk.transforms.PTransform
Returns all PValues that are consumed as inputs to this PTransform that are independent of the expansion of the InputT within PTransform.expand(PInput).
getAdditionalOutputTags() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
 
getAddresses() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
 
getAlgorithm() - Method in enum org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType
Returns the string representation of this type.
getAll(TupleTag<V>) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
Returns the values from the table represented by the given TupleTag<V> as an Iterable<V> (which may be empty if there are no results).
getAll(String) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
Like CoGbkResult.getAll(TupleTag) but using a String instead of a TupleTag.
getAll() - Method in class org.apache.beam.sdk.values.PCollectionList
Returns an immutable List of all the PCollections in this PCollectionList.
getAll() - Method in class org.apache.beam.sdk.values.PCollectionTuple
Returns an immutable Map from TupleTag to corresponding PCollection, for all the members of this PCollectionTuple.
getAll() - Method in class org.apache.beam.sdk.values.TupleTagList
Returns an immutable List of all the TupleTags in this TupleTagList.
getAllFields() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
If true, all fields are being accessed.
getAllIds(String, String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
getAllJobs() - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
 
getAllMetadata() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
Deprecated.
use schema options instead.
getAllowedLateness() - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
getAllowedTimestampSkew() - Method in class org.apache.beam.sdk.transforms.DoFn
Deprecated.
This method permits a DoFn to emit elements behind the watermark. These elements are considered late, and if behind the allowed lateness of a downstream PCollection may be silently dropped. See https://issues.apache.org/jira/browse/BEAM-644 for details on a replacement.
getAllowedTimestampSkew() - Method in class org.apache.beam.sdk.transforms.WithTimestamps
Deprecated.
This method permits a to elements to be emitted behind the watermark. These elements are considered late, and if behind the allowed lateness of a downstream PCollection may be silently dropped. See https://issues.apache.org/jira/browse/BEAM-644 for details on a replacement.
getAllowNonRestoredState() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getAllRows(String, String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
getAllWorkerStatuses(long, TimeUnit) - Method in class org.apache.beam.runners.fnexecution.status.BeamWorkerStatusGrpcService
Get all the statuses from all connected SDK harnesses within specified timeout.
getAnnotatedConstructor(Class) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
 
getAnnotatedCreateMethod(Class) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
 
getApexDAG() - Method in class org.apache.beam.runners.apex.ApexRunnerResult
Return the DAG executed by the pipeline.
getApexLauncher() - Method in class org.apache.beam.runners.apex.ApexYarnLauncher.LaunchParams
 
getApiRootUrl() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
The root URL for the Dataflow API.
getApiServiceDescriptor() - Method in class org.apache.beam.runners.fnexecution.GrpcFnServer
Get an Endpoints.ApiServiceDescriptor describing the endpoint this GrpcFnServer is bound to.
getApplicationName() - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
 
getApplicationName() - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
 
getApplicationName() - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher.TableAndQuery
 
getAppliedFn(CoderRegistry, Coder<? extends KV<K, ? extends Iterable<InputT>>>, WindowingStrategy<?, ?>) - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
Returns the Combine.CombineFn bound to its coders.
getAppName() - Method in interface org.apache.beam.sdk.options.ApplicationNameOptions
Name of application, for display purposes.
getApproximateArrivalTimestamp() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
getArgument() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
 
getArgument() - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
 
getArgument() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
 
getArgument() - Method in class org.apache.beam.sdk.schemas.logicaltypes.PassThroughLogicalType
 
getArgument() - Method in interface org.apache.beam.sdk.schemas.Schema.LogicalType
An optional argument to configure the type.
getArgumentType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
 
getArgumentType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
 
getArgumentType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
 
getArgumentType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.PassThroughLogicalType
 
getArgumentType() - Method in interface org.apache.beam.sdk.schemas.Schema.LogicalType
A schema type representing how to interpret the argument.
getArgumentTypes(Method) - Method in class org.apache.beam.sdk.values.TypeDescriptor
Returns a list of argument types for the given method, which must be a part of the class.
getArray(String) - Method in class org.apache.beam.sdk.values.Row
Get an array value by field name, IllegalStateException is thrown if schema doesn't match.
getArray(int) - Method in class org.apache.beam.sdk.values.Row
Get an array value by field index, IllegalStateException is thrown if schema doesn't match.
getArtifact(ArtifactApi.LegacyGetArtifactRequest, StreamObserver<ArtifactApi.ArtifactChunk>) - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactRetrievalService
 
getArtifactManifestUri(String) - Static method in class org.apache.beam.runners.fnexecution.jobsubmission.PortablePipelineJarUtils
 
getArtifactPort() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver.ServerConfiguration
 
getArtifactStagingPath() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver.ServerConfiguration
 
getArtifactUri(String, String) - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactStagingService
 
getArtifactUri(String, String) - Method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService
 
getArtifactUri(String, String) - Method in class org.apache.beam.runners.fnexecution.artifact.JavaFilesystemLegacyArtifactStagingService
 
getATN() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
getATN() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
getAttempted() - Method in class org.apache.beam.sdk.metrics.MetricResult
Return the value of this metric across all attempts of executing all parts of the pipeline.
getAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
Returns the given attribute value.
getAttributeMap() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
Returns the full map of attributes.
getAutoCommit() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
getAutoscalingAlgorithm() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
[Experimental] The autoscaling algorithm to use for the workerpool.
getAutoWatermarkInterval() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getAwsCredentialsProvider() - Method in interface org.apache.beam.sdk.io.aws.options.AwsOptions
The credential instance that should be used to authenticate against AWS services.
getAwsCredentialsProvider() - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
The credential instance that should be used to authenticate against AWS services.
getAwsRegion() - Method in interface org.apache.beam.sdk.io.aws.options.AwsOptions
AWS region used by the AWS client.
getAwsServiceEndpoint() - Method in interface org.apache.beam.sdk.io.aws.options.AwsOptions
The AWS service endpoint used by the AWS client.
getBacklogCheckTime() - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicy.PartitionContext
The time at which latest offset for the partition was fetched in order to calculate backlog.
getBagUserStateSpecs() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
Get a mapping from PTransform id to user state input id to bag user states that are used during execution.
getBaseAutoValueClass(Class<?>) - Static method in class org.apache.beam.sdk.schemas.utils.AutoValueUtils
 
getBasePath() - Method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService.StagingSessionToken
Access is public for json conversion.
getBaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
 
getBaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
 
getBaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
 
getBaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.PassThroughLogicalType
 
getBaseType() - Method in interface org.apache.beam.sdk.schemas.Schema.LogicalType
The base Schema.FieldType used to store values of this type.
getBaseValue(String, Class<T>) - Method in class org.apache.beam.sdk.values.Row
Returns the base type for this field.
getBaseValue(String) - Method in class org.apache.beam.sdk.values.Row
Returns the base type for this field.
getBaseValue(int, Class<T>) - Method in class org.apache.beam.sdk.values.Row
Returns the base type for this field.
getBaseValue(int) - Method in class org.apache.beam.sdk.values.Row
Returns the base type for this field.
getBaseValues() - Method in class org.apache.beam.sdk.values.Row
Return a list of data values.
getBatchDuration() - Method in class org.apache.beam.runners.spark.io.CreateStream
 
getBatches() - Method in class org.apache.beam.runners.spark.io.CreateStream
Get the underlying queue representing the mock stream of micro-batches.
getBatchIntervalMillis() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
getBatchSize() - Method in class org.apache.beam.sdk.transforms.GroupIntoBatches
Returns the size of the batch.
getBeamRelInput(RelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils
 
getBeamSqlTable() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
 
getBeamSqlUdfs() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.UdfUdafProvider
For UDFs implement BeamSqlUdf.
getBigtableOptions() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Deprecated.
will be replaced by bigtable options configurator.
getBigtableOptions() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
Deprecated.
will be replaced by bigtable options configurator.
getBoolean(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getBoolean(Map<String, Object>, String, Boolean) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getBoolean(String) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.BOOLEAN value by field name, IllegalStateException is thrown if schema doesn't match.
getBoolean(int) - Method in class org.apache.beam.sdk.values.Row
Get a Boolean value by field index, ClassCastException is thrown if schema doesn't match.
getBootstrapServers() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
 
getBoundednessOfRelNode(RelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
This method returns the Boundedness of a RelNode.
getBroadcastSizeEstimate() - Method in class org.apache.beam.runners.spark.util.SideInputBroadcast
 
getBroadcastValue(String) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.SideInputBroadcast
 
getBucket() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
Returns the bucket name associated with this GCS path, or an empty string if this is a relative path component.
getBufferSize() - Method in class org.apache.beam.sdk.fn.stream.BufferingStreamObserver
 
getBuilderCreator(Class<?>, Schema, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.AutoValueUtils
Try to find an accessible builder class for creating an AutoValue class.
getBuiltinMethods() - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BeamBuiltinFunctionProvider
 
getBundle(OutputReceiverFactory, StateRequestHandler, BundleProgressHandler) - Method in interface org.apache.beam.runners.fnexecution.control.StageBundleFactory
Get a new bundle for processing the data in an executable stage.
getBundle(OutputReceiverFactory, TimerReceiverFactory, StateRequestHandler, BundleProgressHandler) - Method in interface org.apache.beam.runners.fnexecution.control.StageBundleFactory
Get a new bundle for processing the data in an executable stage.
getBundleSize() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
getByte(String) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.BYTE value by field name, IllegalStateException is thrown if schema doesn't match.
getByte(int) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.BYTE value by field index, ClassCastException is thrown if schema doesn't match.
getBytes(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getBytes(Map<String, Object>, String, byte[]) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getBytes() - Method in class org.apache.beam.sdk.io.range.ByteKey
Returns a newly-allocated byte[] representing this ByteKey.
getBytes(String) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.BYTES value by field name, IllegalStateException is thrown if schema doesn't match.
getBytes(int) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.BYTES value by field index, ClassCastException is thrown if schema doesn't match.
getBytesPerOffset() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
Returns approximately how many bytes of data correspond to a single offset in this source.
getCacheToken() - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.BagUserStateHandler
Returns the currently valid cache token.
getCacheTokens() - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandler
Retrieves a list of valid cache tokens.
getCaseEnumType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
Returns the EnumerationType that is used to represent the case type.
getCaseType() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType.Value
Returns the enumeration that specified which OneOf field is set.
getCastSpec(RelDataType) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamBigQuerySqlDialect
BigQuery data type reference: Bigquery Standard SQL Data Types.
getCatalog() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
getCause() - Method in class org.apache.beam.runners.jet.FailedRunningPipelineResults
 
getCause() - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
Returns the reason that this WindowFn is invalid.
getChannelFactory() - Method in class org.apache.beam.sdk.io.CompressedSource
 
getChannelNames() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
getCheckpointDir() - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
 
getCheckpointDurationMillis() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
getCheckpointingInterval() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getCheckpointingMode() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getCheckpointMark() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
 
getCheckpointMark() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
Returns a UnboundedSource.CheckpointMark representing the progress of this UnboundedReader.
getCheckpointMarkCoder() - Method in class org.apache.beam.runners.gearpump.translators.io.ValuesSource
 
getCheckpointMarkCoder() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
 
getCheckpointMarkCoder() - Method in class org.apache.beam.sdk.io.UnboundedSource
Returns a Coder for encoding and decoding the checkpoints for this source.
getCheckpointTimeoutMillis() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getChildRels(RelNode) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
 
getClasses() - Method in class org.apache.beam.sdk.values.TypeDescriptor
Returns a set of TypeDescriptors, one for each superclass (including this class).
getClassName() - Method in class org.apache.beam.runners.dataflow.util.CloudObject
Gets the name of the Java class that this CloudObject represents.
getClazz() - Method in class org.apache.beam.sdk.schemas.utils.ReflectUtils.ClassWithSchema
 
getClient(String) - Static method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
 
getClientConfiguration() - Method in interface org.apache.beam.sdk.io.aws.options.AwsOptions
The client configuration instance that should be used to configure AWS service clients.
getClientContext() - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
 
getClientContext() - Method in class org.apache.beam.runners.gearpump.GearpumpPipelineResult
 
getClientInfo(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
getClientInfo() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
getCloningBehavior() - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
getClosingBehavior() - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
getClosure() - Method in class org.apache.beam.sdk.transforms.Contextful
Returns the closure.
getCloudWatchClient() - Method in interface org.apache.beam.sdk.io.aws.dynamodb.AwsClientsProvider
 
getCloudWatchClient() - Method in class org.apache.beam.sdk.io.aws.dynamodb.BasicDynamoDBProvider
 
getCloudWatchClient() - Method in interface org.apache.beam.sdk.io.aws.sns.AwsClientsProvider
 
getCloudWatchClient() - Method in interface org.apache.beam.sdk.io.kinesis.AWSClientsProvider
 
getClustering() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
getClusterName() - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
 
getCmd() - Method in class org.apache.beam.runners.apex.ApexYarnLauncher.LaunchParams
 
getCodec(DestinationT) - Method in class org.apache.beam.sdk.io.DynamicAvroDestinations
Return an AVRO codec for a given destination.
getCodeJarPathname() - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
 
getCoder() - Method in class org.apache.beam.runners.fnexecution.control.RemoteOutputReceiver
 
getCoder() - Method in class org.apache.beam.runners.fnexecution.data.RemoteInputDestination
 
getCoder(String) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.SideInputBroadcast
 
getCoder(Class<T>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
Returns the Coder to use for values of the given class.
getCoder(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
Returns the Coder to use for values of the given type.
getCoder(TypeDescriptor<OutputT>, TypeDescriptor<InputT>, Coder<InputT>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
Deprecated.
This method is to change in an unknown backwards incompatible way once support for this functionality is refined.
getCoder(Class<? extends T>, Class<T>, Map<Type, ? extends Coder<?>>, TypeVariable<?>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
Deprecated.
This method is to change in an unknown backwards incompatible way once support for this functionality is refined.
getCoder() - Method in class org.apache.beam.sdk.coders.DelegateCoder
Returns the coder used to encode/decode the intermediate values produced/consumed by the coding functions of this DelegateCoder.
getCoder(CoderRegistry) - Method in class org.apache.beam.sdk.io.kafka.ConfluentSchemaRegistryDeserializerProvider
 
getCoder() - Static method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
Returns a Coder suitable for IntervalWindow.
getCoder() - Method in class org.apache.beam.sdk.values.PCollection
Returns the Coder used by this PCollection to encode and decode the values stored in it.
getCoderArguments() - Method in class org.apache.beam.sdk.coders.AtomicCoder
If this is a Coder for a parameterized type, returns the list of Coders being used for each of the parameters in the same order they appear within the parameterized type's type signature.
getCoderArguments() - Method in class org.apache.beam.sdk.coders.Coder
If this is a Coder for a parameterized type, returns the list of Coders being used for each of the parameters in the same order they appear within the parameterized type's type signature.
getCoderArguments() - Method in class org.apache.beam.sdk.coders.CustomCoder
If this is a Coder for a parameterized type, returns the list of Coders being used for each of the parameters in the same order they appear within the parameterized type's type signature.
getCoderArguments() - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.coders.KvCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.coders.MapCoder
If this is a Coder for a parameterized type, returns the list of Coders being used for each of the parameters in the same order they appear within the parameterized type's type signature.
getCoderArguments() - Method in class org.apache.beam.sdk.coders.NullableCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.coders.ShardedKeyCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.coders.SnappyCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.coders.VarLongCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.io.hadoop.WritableCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.testing.TestStream.TestStreamCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
 
getCoderArguments() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
 
getCoderArguments() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
 
getCoderArguments() - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
 
getCoderInternal() - Method in interface org.apache.beam.sdk.values.PCollectionView
Deprecated.
this method will be removed entirely. The PCollection underlying a side input, including its Coder, is part of the side input's specification with a ParDo transform, which will obtain that information via a package-private channel.
getCoderInternal() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
 
getCoderProvider() - Static method in class org.apache.beam.sdk.coders.AvroCoder
Returns a CoderProvider which uses the AvroCoder if possible for all types.
getCoderProvider() - Static method in class org.apache.beam.sdk.coders.SerializableCoder
Returns a CoderProvider which uses the SerializableCoder if possible for all types.
getCoderProvider() - Static method in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
Returns a CoderProvider which uses the DynamicProtoCoder for proto messages.
getCoderProvider() - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
Returns a CoderProvider which uses the ProtoCoder for proto messages.
getCoderProvider() - Static method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteResultCoder
 
getCoderProvider() - Static method in class org.apache.beam.sdk.io.hadoop.WritableCoder
Returns a CoderProvider which uses the WritableCoder for Hadoop writable types.
getCoderProviders() - Method in interface org.apache.beam.sdk.coders.CoderProviderRegistrar
Returns a list of coder providers which will be registered by default within each coder registry instance.
getCoderProviders() - Method in class org.apache.beam.sdk.coders.DefaultCoder.DefaultCoderProviderRegistrar
 
getCoderProviders() - Method in class org.apache.beam.sdk.coders.SerializableCoder.SerializableCoderProviderRegistrar
 
getCoderProviders() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtobufCoderProviderRegistrar
 
getCoderProviders() - Method in class org.apache.beam.sdk.io.amqp.AmqpMessageCoderProviderRegistrar
 
getCoderProviders() - Method in class org.apache.beam.sdk.io.aws.dynamodb.AttributeValueCoderProviderRegistrar
 
getCoderProviders() - Method in class org.apache.beam.sdk.io.aws.sns.SnsCoderProviderRegistrar
 
getCoderProviders() - Method in class org.apache.beam.sdk.io.aws2.sns.SnsCoderProviderRegistrar
 
getCoderProviders() - Method in class org.apache.beam.sdk.io.aws2.sqs.MessageCoderRegistrar
 
getCoderProviders() - Method in class org.apache.beam.sdk.io.aws2.sqs.SendMessageRequestCoderRegistrar
 
getCoderProviders() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryCoderProviderRegistrar
 
getCoderProviders() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubCoderProviderRegistrar
 
getCoderProviders() - Method in class org.apache.beam.sdk.io.hadoop.WritableCoder.WritableCoderProviderRegistrar
 
getCoderProviders() - Method in class org.apache.beam.sdk.io.hbase.HBaseCoderProviderRegistrar
 
getCoderRegistry() - Method in class org.apache.beam.sdk.Pipeline
Returns the CoderRegistry that this Pipeline uses.
getCoGbkResultSchema() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
Returns the CoGbkResultSchema associated with this KeyedPCollectionTuple.
getCollations() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
 
getCollection() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple.TaggedKeyedPCollection
Returns the underlying PCollection of this TaggedKeyedPCollection.
getCollectionElementType() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
 
getColumns() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
getCombineFn() - Method in interface org.apache.beam.runners.spark.aggregators.NamedAggregators.State
 
getCombineFn() - Method in interface org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregators.State
 
getCombineFn() - Method in class org.apache.beam.sdk.extensions.sql.impl.UdafImpl
 
getCombineFn() - Method in class org.apache.beam.sdk.transforms.Combine.GloballyAsSingletonView
 
getComment() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table
 
getCommitDeadline() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
getCommitted() - Method in class org.apache.beam.sdk.metrics.MetricResult
Return the value of this metric across all successfully completed parts of the pipeline.
getCommittedOrNull() - Method in class org.apache.beam.sdk.metrics.MetricResult
Return the value of this metric across all attempts of executing all parts of the pipeline.
getCompletedTimers() - Method in class org.apache.beam.runners.direct.WatermarkManager.TimerUpdate
 
getComponents() - Method in class org.apache.beam.sdk.coders.AtomicCoder
Returns the list of Coders that are components of this Coder.
getComponents() - Method in class org.apache.beam.sdk.coders.StructuredCoder
Returns the list of Coders that are components of this Coder.
getComponents() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Path
Hierarchy list of component paths making up the full path, starting with the top-level child component path.
getComponents() - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
 
getComponents() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
 
getComponents() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
 
getComponentType() - Method in class org.apache.beam.sdk.values.TypeDescriptor
Returns the component type if this type is an array type, otherwise returns null.
getCompression() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
Returns the method with which this file will be decompressed in FileIO.ReadableFile.open().
getCompression() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
 
getComputeNumShards() - Method in class org.apache.beam.sdk.io.WriteFiles
 
getConfigFile() - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
 
getConfigPropertiesAsMap(HiveConf) - Static method in class org.apache.beam.sdk.io.hcatalog.test.HCatalogIOTestUtils
Returns config params for the test datastore as a Map.
getConfiguration() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
 
getConfiguration() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
 
getConfiguration() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
 
getConfiguration() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
 
getConnectStringPrefix() - Method in class org.apache.beam.sdk.extensions.sql.impl.JdbcDriver
 
getConnectTimeout() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
 
getConstructorCreator(Class<?>, Schema, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.AutoValueUtils
Try to find an accessible constructor for creating an AutoValue class.
getConstructorCreator(Class, Constructor, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
 
getConstructorCreator(Class, Constructor, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
 
getContainerVersion() - Method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
Provides the container version that will be used for constructing harness image paths.
getContent() - Method in class org.apache.beam.sdk.io.tika.ParseResult
Returns the extracted text.
getContext() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
 
getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterAll
 
getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterEach
 
getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterFirst
 
getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
 
getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
 
getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterSynchronizedProcessingTime
 
getContinuationTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
 
getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
 
getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
 
getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.DefaultTrigger
 
getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.Never.NeverTrigger
 
getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
 
getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
 
getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.ReshuffleTrigger
Deprecated.
 
getContinuationTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
Return a trigger to use after a GroupByKey to preserve the intention of this trigger.
getContinuationTrigger(List<Trigger>) - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
Subclasses should override this to return the Trigger.getContinuationTrigger() of this Trigger.
getContinuationTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.Trigger.OnceTrigger
 
getConversionOptions(JSONObject) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryTableProvider
 
getConvertedSchemaInformation(Schema, TypeDescriptor<T>, SchemaRegistry) - Static method in class org.apache.beam.sdk.schemas.utils.ConvertHelpers
Get the coder used for converting from an inputSchema to a given type.
getConvertedType() - Method in class org.apache.beam.sdk.coders.AvroCoder.JodaTimestampConversion
 
getConvertPrimitive(Schema.FieldType, TypeDescriptor<?>, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.ConvertHelpers
Returns a function to convert a Row into a primitive type.
getCount() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSortRel
 
getCount() - Method in class org.apache.beam.sdk.metrics.DistributionResult
 
getCount() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
 
getCounter(MetricName) - Method in class org.apache.beam.runners.jet.metrics.JetMetricsContainer
 
getCounter(MetricName) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
Return the Counter that should be used for implementing the given metricName in this container.
getCounters(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
 
getCounters(MetricFilter) - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.WithMetricsSupport
 
getCounters() - Method in class org.apache.beam.sdk.metrics.MetricQueryResults
Return the metric results for the counters that matched the filter.
getCountryOfResidence() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
 
getCpu() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
getCpuRate() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
getCreateFromSnapshot() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
If set, the snapshot from which the job should be created.
getCreator(Class<T>, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Get an object creator for an AVRO-generated SpecificRecord.
getCredential() - Method in interface org.apache.beam.sdk.extensions.gcp.auth.CredentialFactory
 
getCredential() - Method in class org.apache.beam.sdk.extensions.gcp.auth.GcpCredentialFactory
Returns a default GCP Credentials or null when it fails.
getCredential() - Method in class org.apache.beam.sdk.extensions.gcp.auth.NoopCredentialFactory
 
getCredentialFactoryClass() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
The class of the credential factory that should be created and used to create credentials.
getCsvFormat() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTableProvider.CsvToRow
 
getCurrent() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
 
getCurrent() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
 
getCurrent() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
Gets the current record from the delegate reader.
getCurrent() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource.BigQueryStorageStreamReader
 
getCurrent() - Method in class org.apache.beam.sdk.io.Source.Reader
Returns the value of the data item that was read by the last Source.Reader.start() or Source.Reader.advance() call.
getCurrentBlock() - Method in class org.apache.beam.sdk.io.AvroSource.AvroReader
 
getCurrentBlock() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
Returns the current block (the block that was read by the last successful call to BlockBasedSource.BlockBasedReader.readNextBlock()).
getCurrentBlockOffset() - Method in class org.apache.beam.sdk.io.AvroSource.AvroReader
 
getCurrentBlockOffset() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
Returns the largest offset such that starting to read from that offset includes the current block.
getCurrentBlockSize() - Method in class org.apache.beam.sdk.io.AvroSource.AvroReader
 
getCurrentBlockSize() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
Returns the size of the current block in bytes as it is represented in the underlying file, if possible.
getCurrentContainer() - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
Return the MetricsContainer for the current thread.
getCurrentDirectory() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
 
getCurrentDirectory() - Method in interface org.apache.beam.sdk.io.fs.ResourceId
Returns the ResourceId that represents the current directory of this ResourceId.
getCurrentOffset() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
 
getCurrentOffset() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
 
getCurrentOffset() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
Returns the starting offset of the current record, which has been read by the last successful Source.Reader.start() or Source.Reader.advance() call.
getCurrentParent() - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
Gets the parent composite transform to the current transform, if one exists.
getCurrentRecord() - Method in class org.apache.beam.sdk.io.BlockBasedSource.Block
Returns the current record.
getCurrentRecordId() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
Returns a unique identifier for the current record.
getCurrentSchemaPlus() - Method in class org.apache.beam.sdk.extensions.sql.impl.JdbcConnection
Calcite-created SchemaPlus wrapper for the current schema.
getCurrentSource() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
 
getCurrentSource() - Method in class org.apache.beam.sdk.io.AvroSource.AvroReader
 
getCurrentSource() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
Returns a Source describing the same input that this Reader currently reads (including items already read).
getCurrentSource() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
 
getCurrentSource() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource.BigQueryStorageStreamReader
 
getCurrentSource() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
 
getCurrentSource() - Method in class org.apache.beam.sdk.io.Source.Reader
Returns a Source describing the same input that this Reader currently reads (including items already read).
getCurrentSource() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
Returns the UnboundedSource that created this reader.
getCurrentTimestamp() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
 
getCurrentTimestamp() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
By default, returns the minimum possible timestamp.
getCurrentTimestamp() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
 
getCurrentTimestamp() - Method in class org.apache.beam.sdk.io.Source.Reader
Returns the timestamp associated with the current data item.
getCurrentTransform() - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
 
getCurrentTransform() - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
 
getCurrentTransform() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
getCustomerId() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Order
 
getCustomError(HttpRequestWrapper, HttpResponseWrapper) - Method in class org.apache.beam.sdk.extensions.gcp.util.CustomHttpErrors
 
getCustomError() - Method in class org.apache.beam.sdk.extensions.gcp.util.CustomHttpErrors.MatcherAndError
 
getDanglingDataSets() - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.BatchTranslationContext
 
getData() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
getDataAsBytes() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
getDatabaseId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
getDataCatalogEndpoint() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogPipelineOptions
DataCatalog endpoint.
getDataCoder() - Method in class org.apache.beam.runners.dataflow.DataflowRunner.StreamingPCollectionViewWriterFn
 
getDataflowClient() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
An instance of the Dataflow client.
getDataflowEndpoint() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
Dataflow endpoint to use.
getDataflowJobFile() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
The path to write the translated Dataflow job specification out to at job submission time.
getDataflowKmsKey() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
GCP Cloud KMS key for Dataflow pipelines and buckets created by GcpTempLocationFactory.
getDataflowOptions() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
 
getDataflowRunnerInfo() - Static method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
Returns an instance of DataflowRunnerInfo.
getDataflowWorkerJar() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
 
getDataset(PValue) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
getDataset(String, String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
Gets the specified Dataset resource by dataset ID.
getDataset(String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
getDataSetOrThrow(String) - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.BatchTranslationContext
 
getDatasetService(BigQueryOptions) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices
Returns a real, mock, or fake BigQueryServices.DatasetService.
getDatasetService(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
 
getDataStreamOrThrow(String) - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext
 
getDateTime(String) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.DATETIME value by field name, IllegalStateException is thrown if schema doesn't match.
getDateTime(int) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.DATETIME value by field index, IllegalStateException is thrown if schema doesn't match.
getDayOfMonth() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
 
getDayOfMonth() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
 
getDebuggee() - Method in interface org.apache.beam.runners.dataflow.options.CloudDebuggerOptions
The Cloud Debugger debuggee to associate with.
getDecimal(String) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.DECIMAL value by field name, IllegalStateException is thrown if schema doesn't match.
getDecimal(int) - Method in class org.apache.beam.sdk.values.Row
Get a BigDecimal value by field index, ClassCastException is thrown if schema doesn't match.
getDef() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner.NonCumulativeCostImpl
 
getDef() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.RelMdNodeStats
 
getDefault() - Static method in class org.apache.beam.runners.fnexecution.logging.Slf4jLogWriter
 
getDefaultCoder(TypeDescriptor<?>, CoderRegistry) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
Returns the default coder for a given type descriptor.
getDefaultDestination() - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
Returns the default destination.
getDefaultEnvironmentConfig() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
getDefaultEnvironmentType() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
getDefaultJobName() - Static method in class org.apache.beam.runners.fnexecution.jobsubmission.PortablePipelineJarUtils
 
getDefaultOutputCoder() - Method in class org.apache.beam.runners.gearpump.translators.io.ValuesSource
 
getDefaultOutputCoder(CoderRegistry, Coder) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.TDigestQuantilesFn
 
getDefaultOutputCoder() - Method in class org.apache.beam.sdk.io.Source
Deprecated.
Override Source.getOutputCoder() instead.
getDefaultOutputCoder(CoderRegistry, Coder<Double>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
 
getDefaultOutputCoder(CoderRegistry, Coder<V>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
 
getDefaultOutputCoder(CoderRegistry, Coder<Integer>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
 
getDefaultOutputCoder(CoderRegistry, Coder<Long>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
 
getDefaultOutputCoder(CoderRegistry, Coder<InputT>) - Method in interface org.apache.beam.sdk.transforms.CombineFnBase.GlobalCombineFn
Returns the Coder to use by default for output OutputT values, or null if it is not able to be inferred.
getDefaultOutputCoder() - Method in class org.apache.beam.sdk.transforms.PTransform
Deprecated.
Instead, the PTransform should explicitly call PCollection.setCoder(org.apache.beam.sdk.coders.Coder<T>) on the returned PCollection.
getDefaultOutputCoder(InputT) - Method in class org.apache.beam.sdk.transforms.PTransform
Deprecated.
Instead, the PTransform should explicitly call PCollection.setCoder(org.apache.beam.sdk.coders.Coder<T>) on the returned PCollection.
getDefaultOutputCoder(InputT, PCollection<T>) - Method in class org.apache.beam.sdk.transforms.PTransform
Deprecated.
Instead, the PTransform should explicitly call PCollection.setCoder(org.apache.beam.sdk.coders.Coder<T>) on the returned PCollection.
getDefaultOutputCoder(CoderRegistry, Coder<T>) - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
 
getDefaultSdkHarnessLogLevel() - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
This option controls the default log level of all loggers without a log level override.
getDefaultValue() - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn
Returns the default value that was specified.
getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
 
getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
 
getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
 
getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
 
getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
Return a WindowMappingFn that returns the earliest window that contains the end of the main-input window.
getDefaultWindowMappingFn() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
Returns the default WindowMappingFn to use to map main input windows to side input windows.
getDefaultWorkerLogLevel() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
Deprecated.
This option controls the default log level of all loggers without a log level override.
getDelay() - Method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform.Delay
 
getDeletedTimers() - Method in class org.apache.beam.runners.direct.WatermarkManager.TimerUpdate
 
getDelimiters() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
 
getDescription() - Method in class org.apache.beam.sdk.schemas.Schema.Field
Returns the field's description.
getDescriptor(String) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
 
getDeserializer(Map<String, ?>, boolean) - Method in class org.apache.beam.sdk.io.kafka.ConfluentSchemaRegistryDeserializerProvider
 
getDestination(UserT) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
Returns an object that represents at a high level the destination being written to.
getDestination() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
 
getDestination() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
Return the user destination object for this writer.
getDestination(ValueInSingleWindow<T>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
Returns an object that represents at a high level which table is being written to.
getDestinationCoder() - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
Returns the coder for DestinationT.
getDestinationCoder() - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
Returns the coder for DestinationT.
getDestinationFile(boolean, FileBasedSink.DynamicDestinations<?, DestinationT, ?>, int, FileBasedSink.OutputFileHints) - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
 
getDictionary(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getDictionary(Map<String, Object>, String, Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getDirectoryTreatment() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
 
getDisableMetrics() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getDiskSizeGb() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
Remote worker disk size, in gigabytes, or 0 to use the default size.
getDistribution(MetricName) - Method in class org.apache.beam.runners.jet.metrics.JetMetricsContainer
 
getDistribution() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
 
getDistribution(MetricName) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
Return the Distribution that should be used for implementing the given metricName in this container.
getDistributions() - Method in class org.apache.beam.sdk.metrics.MetricQueryResults
Return the metric results for the distributions that matched the filter.
getDocumentCount() - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
 
getDoFnRunner(PipelineOptions, DoFn<InputT, OutputT>, SideInputReader, AbstractParDoP.JetOutputManager, TupleTag<OutputT>, List<TupleTag<?>>, Coder<InputT>, Map<TupleTag<?>, Coder<?>>, WindowingStrategy<?, ?>, DoFnSchemaInformation, Map<String, PCollectionView<?>>) - Method in class org.apache.beam.runners.jet.processors.ParDoP
 
getDoFnRunner(PipelineOptions, DoFn<KV<?, ?>, OutputT>, SideInputReader, AbstractParDoP.JetOutputManager, TupleTag<OutputT>, List<TupleTag<?>>, Coder<KV<?, ?>>, Map<TupleTag<?>, Coder<?>>, WindowingStrategy<?, ?>, DoFnSchemaInformation, Map<String, PCollectionView<?>>) - Method in class org.apache.beam.runners.jet.processors.StatefulParDoP
 
getDoFnSchemaInformation(DoFn<?, ?>, PCollection<?>) - Static method in class org.apache.beam.sdk.transforms.ParDo
Extract information on how the DoFn uses schemas.
getDouble(String) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.DOUBLE value by field name, IllegalStateException is thrown if schema doesn't match.
getDouble(int) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.DOUBLE value by field index, ClassCastException is thrown if schema doesn't match.
getDumpHeapOnOOM() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
If true, save a heap dump before killing a thread or process which is GC thrashing or out of memory.
getDynamicDestinations() - Method in class org.apache.beam.sdk.io.FileBasedSink
getDynamoDbClient() - Method in class org.apache.beam.sdk.io.aws2.dynamodb.BasicDynamoDbClientProvider
 
getDynamoDbClient() - Method in interface org.apache.beam.sdk.io.aws2.dynamodb.DynamoDbClientProvider
 
getEarlyTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
 
getElemCoder() - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
 
getElement() - Method in class org.apache.beam.sdk.io.gcp.bigquery.AvroWriteRequest
 
getElementCoders() - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
 
getElementConverters() - Method in class org.apache.beam.sdk.transforms.DoFnSchemaInformation
The schema of the @Element parameter.
getElementCount() - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
The number of elements after which this trigger may fire.
getElements() - Method in class org.apache.beam.sdk.testing.TestStream.ElementEvent
 
getElements() - Method in class org.apache.beam.sdk.transforms.Create.Values
 
getElementType() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
If the field is a container type, returns the element type.
getEmptyMatchTreatment() - Method in class org.apache.beam.sdk.io.FileBasedSource
 
getEmptyMatchTreatment() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
 
getEmptyTraitSet() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl
 
getEnableCloudDebugger() - Method in interface org.apache.beam.runners.dataflow.options.CloudDebuggerOptions
Whether to enable the Cloud Debugger snapshot agent for the current job.
getEnableSparkMetricSinks() - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
 
getEncodedElementByteSize(RandomAccessData) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
 
getEncodedElementByteSize(BigDecimal) - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
Returns the size in bytes of the encoded value using this coder.
getEncodedElementByteSize(Integer) - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
Returns the size in bytes of the encoded value using this coder.
getEncodedElementByteSize(Long) - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
Returns the size in bytes of the encoded value using this coder.
getEncodedElementByteSize(Short) - Method in class org.apache.beam.sdk.coders.BigEndianShortCoder
Returns the size in bytes of the encoded value using this coder.
getEncodedElementByteSize(BigInteger) - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
Returns the size in bytes of the encoded value using this coder.
getEncodedElementByteSize(Boolean) - Method in class org.apache.beam.sdk.coders.BooleanCoder
 
getEncodedElementByteSize(byte[]) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
 
getEncodedElementByteSize(Byte) - Method in class org.apache.beam.sdk.coders.ByteCoder
Returns the size in bytes of the encoded value using this coder.
getEncodedElementByteSize(T) - Method in class org.apache.beam.sdk.coders.Coder
Returns the size in bytes of the encoded value using this coder.
getEncodedElementByteSize(Double) - Method in class org.apache.beam.sdk.coders.DoubleCoder
Returns the size in bytes of the encoded value using this coder.
getEncodedElementByteSize(Float) - Method in class org.apache.beam.sdk.coders.FloatCoder
Returns the size in bytes of the encoded value using this coder.
getEncodedElementByteSize(Instant) - Method in class org.apache.beam.sdk.coders.InstantCoder
 
getEncodedElementByteSize(T) - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
Overridden to short-circuit the default StructuredCoder behavior of encoding and counting the bytes.
getEncodedElementByteSize(T) - Method in class org.apache.beam.sdk.coders.NullableCoder
Overridden to short-circuit the default StructuredCoder behavior of encoding and counting the bytes.
getEncodedElementByteSize(String) - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
Returns the size in bytes of the encoded value using this coder.
getEncodedElementByteSize(Integer) - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
 
getEncodedElementByteSize(Integer) - Method in class org.apache.beam.sdk.coders.VarIntCoder
 
getEncodedElementByteSize(Long) - Method in class org.apache.beam.sdk.coders.VarLongCoder
 
getEncodedElementByteSize(Void) - Method in class org.apache.beam.sdk.coders.VoidCoder
 
getEncodedElementByteSize(ByteString) - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
 
getEncodedElementByteSize(HyperLogLogPlus) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.HyperLogLogPlusCoder
 
getEncodedElementByteSize(EncodedBoundedWindow) - Method in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow.Coder
 
getEncodedElementByteSize(BigQueryInsertError) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertErrorCoder
 
getEncodedElementByteSize(TableRow) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
 
getEncodedElementByteSize(OffsetRange) - Method in class org.apache.beam.sdk.io.range.OffsetRange.Coder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.AvroCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.BigEndianShortCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.ByteCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.Coder
Returns the TypeDescriptor for the type encoded.
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.CollectionCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.DelegateCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.DoubleCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.DurationCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.FloatCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.InstantCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.IterableCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.KvCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.ListCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.MapCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.NullableCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.SerializableCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.SetCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.VarIntCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.VarLongCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.coders.VoidCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertErrorCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.io.range.OffsetRange.Coder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.io.xml.JAXBCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
 
getEncodedTypeDescriptor() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
 
getEncodedWindow() - Method in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow
 
getEncodingPositions() - Method in class org.apache.beam.sdk.schemas.Schema
Gets the encoding positions for this schema.
getEndKey() - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
Returns the ByteKey representing the upper bound of this ByteKeyRange.
getEndOffset() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
Returns the specified ending offset of the source.
getEndpoint() - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
The AWS service endpoint used by the AWS client.
getEnv() - Method in class org.apache.beam.runners.apex.ApexYarnLauncher.LaunchParams
 
getEnvironment() - Method in class org.apache.beam.runners.fnexecution.environment.ProcessEnvironment
 
getEnvironment() - Method in interface org.apache.beam.runners.fnexecution.environment.RemoteEnvironment
Return the environment that the remote handles.
getEnvironment() - Method in class org.apache.beam.runners.fnexecution.environment.RemoteEnvironment.SimpleRemoteEnvironment
 
getEnvironment() - Method in class org.apache.beam.runners.fnexecution.environment.StaticRemoteEnvironment
 
getEnvironmentCacheMillis() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
getEnvironmentExpirationMillis() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
getEquivalentFieldType(TableSchema.ColumnType) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema
Returns Beam equivalent of ClickHouse column type.
getEquivalentSchema(TableSchema) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema
Returns Beam equivalent of ClickHouse schema.
getError() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertError
 
getError() - Method in class org.apache.beam.sdk.io.tika.ParseResult
Returns the parse error, if the file was parsed unsuccessfully.
getErrorAsString() - Method in class org.apache.beam.sdk.io.tika.ParseResult
Same as ParseResult.getError(), but returns the complete stack trace of the error as a String.
getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.BoundedSource
An estimate of the total size (in bytes) of the data that would be read from this source.
getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
 
getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
 
getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageQuerySource
 
getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource
 
getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageTableSource
 
getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
 
getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
 
getEstimatedSizeBytes(PipelineOptions) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
 
getEvents() - Method in class org.apache.beam.sdk.testing.TestStream
Returns the sequence of Events in this TestStream.
getEx() - Method in class org.apache.beam.runners.jet.processors.FlattenP.Supplier
 
getExecutable() - Method in class org.apache.beam.runners.direct.WatermarkManager.FiredTimers
 
getExecutables() - Method in interface org.apache.beam.runners.direct.ExecutableGraph
 
getExecutionEnvironment() - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.BatchTranslationContext
 
getExecutionEnvironment() - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext
 
getExecutionModeForBatch() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getExecutionRetryDelay() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getExecutorService() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
The ExecutorService instance to use to create threads, can be overridden to specify an ExecutorService that is compatible with the user's environment.
getExpansionPort() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver.ServerConfiguration
 
getExpectedAssertions() - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
 
getExpectedRecords(int) - Static method in class org.apache.beam.sdk.io.hcatalog.test.HCatalogIOTestUtils
Returns a list of strings containing 'expected' test data for verification.
getExpectedRecordsAsKV(int) - Static method in class org.apache.beam.sdk.io.hcatalog.test.HCatalogIOTestUtils
Returns a list of KV<String, Integer> containing 'expected' test data for verification.
getExperiments() - Method in interface org.apache.beam.sdk.options.ExperimentalOptions
 
getExperimentValue(PipelineOptions, String) - Static method in interface org.apache.beam.sdk.options.ExperimentalOptions
Return the value for the specified experiment or null if not present.
getExplicitHashKey(byte[]) - Method in interface org.apache.beam.sdk.io.kinesis.KinesisPartitioner
 
getExpression(SchemaPlus, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
getExtendedSequenceNumber() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
getExtensionHosts() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
 
getExtensionRegistry() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
Returns the ExtensionRegistry listing all known Protocol Buffers extension messages to T registered with this ProtoCoder.
getExternalSorterType() - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
Returns the external sorter type.
getFactory() - Method in class org.apache.beam.sdk.schemas.utils.AvroUtils.AvroConvertValueForGetter
 
getFactory() - Method in class org.apache.beam.sdk.schemas.utils.AvroUtils.AvroConvertValueForSetter
 
getFactory() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
 
getFactory() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
 
getFailedInserts() - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
Returns a PCollection containing the TableRows that didn't made it to BQ.
getFailedInsertsWithErr() - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
Returns a PCollection containing the BigQueryInsertErrors with detailed error information.
getFailedMutations() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteResult
 
getFailOnCheckpointingErrors() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getFanout() - Method in class org.apache.beam.sdk.transforms.Combine.GloballyAsSingletonView
 
getField() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
 
getField(int) - Method in class org.apache.beam.sdk.schemas.Schema
Return a field by index.
getField(String) - Method in class org.apache.beam.sdk.schemas.Schema
 
getFieldCount() - Method in class org.apache.beam.sdk.schemas.Schema
Return the count of fields.
getFieldCount() - Method in class org.apache.beam.sdk.values.Row
Return the size of data fields.
getFieldCount() - Method in class org.apache.beam.sdk.values.RowWithGetters
 
getFieldCount() - Method in class org.apache.beam.sdk.values.RowWithStorage
 
getFieldId() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor
 
getFieldName() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor
 
getFieldNames() - Method in class org.apache.beam.sdk.schemas.Schema
Return the list of all field names.
getFieldOptionById(int) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
 
getFieldRename() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor
 
getFields() - Method in class org.apache.beam.sdk.schemas.Schema
 
getFields(Class<?>) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
 
getFieldsAccessed() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
 
getFieldType(OneOfType.Value) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
 
getFieldTypes(Class<T>, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Get field types for an AVRO-generated SpecificRecord or a POJO.
getFieldTypes(Class<?>, Schema, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
 
getFieldTypes(Class<?>, Schema, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
 
getFileDescriptor(String) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
 
getFileLocation() - Method in class org.apache.beam.sdk.io.tika.ParseResult
Returns the absolute path to the input file.
getFilename() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
 
getFileName() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
getFilename(BoundedWindow, PaneInfo, int, int, Compression) - Method in interface org.apache.beam.sdk.io.FileIO.Write.FileNaming
Generates the filename.
getFilename() - Method in interface org.apache.beam.sdk.io.fs.ResourceId
Returns the name of the file or directory denoted by this ResourceId.
getFilenamePolicy(DestinationT) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
Converts a destination into a FileBasedSink.FilenamePolicy.
getFileOrPatternSpec() - Method in class org.apache.beam.sdk.io.FileBasedSource
 
getFileOrPatternSpecProvider() - Method in class org.apache.beam.sdk.io.FileBasedSource
 
getFilePattern() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTable
 
getFilePattern() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
 
getFilesToStage() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
List of local files to make available to workers.
getFilesToStage() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
List of local files to make available to workers.
getFilesToStage() - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
List of local files to make available to workers.
getFilesToStage() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
List of local files to make available to workers.
getFileSystem() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
getFlatJsonRows(Schema) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery
Loads rows from BigQuery into Rows with given Schema.
getFlexRSGoal() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
This option controls Flexible Resource Scheduling mode.
getFlinkMaster() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
The url of the Flink JobManager on which to execute pipelines.
getFloat(String) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.FLOAT value by field name, IllegalStateException is thrown if schema doesn't match.
getFloat(int) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.FLOAT value by field index, ClassCastException is thrown if schema doesn't match.
getFn() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.ParDoSingle
 
getFn() - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
 
getFn() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
Returns the CombineFnBase.GlobalCombineFn used by this Combine operation.
getFn() - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
Returns the CombineFnBase.GlobalCombineFn used by this Combine operation.
getFn() - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
Returns the CombineFnBase.GlobalCombineFn used by this Combine operation.
getFn() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
 
getFn() - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
 
getFnApiEnvironmentMajorVersion() - Method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
Provides the FnAPI environment's major version number.
getFractionConsumed() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
 
getFractionConsumed() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
Returns a value in [0, 1] representing approximately what fraction of the current source this reader has read so far, or null if such an estimate is not available.
getFractionConsumed() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource.BigQueryStorageStreamReader
 
getFractionConsumed() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
 
getFractionConsumed() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
 
getFractionConsumed() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
 
getFractionConsumed() - Method in interface org.apache.beam.sdk.io.range.RangeTracker
Returns the approximate fraction of positions in the source that have been consumed by successful RangeTracker.tryReturnRecordAt(boolean, PositionT) calls, or 0.0 if no such calls have happened.
getFractionOfBlockConsumed() - Method in class org.apache.beam.sdk.io.BlockBasedSource.Block
Returns the fraction of the block already consumed, if possible, as a value in [0, 1].
getFrom() - Method in class org.apache.beam.sdk.io.range.OffsetRange
 
getFromRowFunction() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDynamicMessageSchema
 
getFromRowFunction() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
Returns the toRow conversion function.
getFromRowFunction(Class<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Retrieve the function that converts a Row object to the specified type.
getFromRowFunction(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Retrieve the function that converts a Row object to the specified type.
getFromRowFunction(Class<T>) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
 
getFromRowFunction() - Method in class org.apache.beam.sdk.values.PCollection
Returns the attached schema's fromRowFunction.
getFullName(PTransform<?, ?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
Returns the full name of the currently being translated transform.
getFunctionNames() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
getFunctions(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
getGapDuration() - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
 
getGauge(MetricName) - Method in class org.apache.beam.runners.jet.metrics.JetMetricsContainer
 
getGauge(MetricName) - Method in interface org.apache.beam.sdk.metrics.MetricsContainer
Return the Gauge that should be used for implementing the given metricName in this container.
getGauges(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
 
getGauges(MetricFilter) - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.WithMetricsSupport
 
getGauges() - Method in class org.apache.beam.sdk.metrics.MetricQueryResults
Return the metric results for the gauges that matched the filter.
getGcloudCancelCommand(DataflowPipelineOptions, String) - Static method in class org.apache.beam.runners.dataflow.util.MonitoringUtil
 
getGcpCredential() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
The credential instance that should be used to authenticate against GCP services.
getGcpTempLocation() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
A GCS path for storing temporary files in GCP.
getGcsEndpoint() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
GCS endpoint to use.
getGcsPerformanceMetrics() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
If true, reports metrics of certain operations, such as batch copies.
getGcsUploadBufferSizeBytes() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
The buffer size (in bytes) to use when uploading files to GCS.
getGcsUtil() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
The GcsUtil instance that should be used to communicate with Google Cloud Storage.
getGenericRecordToRowFunction(Schema) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
getGetters(Class<T>, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Get generated getters for an AVRO-generated SpecificRecord or a POJO.
getGetters(Class<?>, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
Return the list of FieldValueGetters for a Java Bean class
getGetters(Class<?>, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
 
getGetters() - Method in class org.apache.beam.sdk.values.RowWithGetters
 
getGetterTarget() - Method in class org.apache.beam.sdk.values.RowWithGetters
 
getGoogleApiTrace() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions
This option enables tracing of API calls to Google services used within the Apache Beam SDK.
getGrammarFileName() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
getGrammarFileName() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
getHdfsConfiguration() - Method in interface org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptions
 
getHeaderAccessor() - Static method in class org.apache.beam.runners.fnexecution.GrpcContextHeaderAccessorProvider
 
getHeaders() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
 
getHighWatermark() - Method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.SparkWatermarks
 
getHistograms(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
 
getHistograms(MetricFilter) - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.WithMetricsSupport
 
getHiveConf() - Method in class org.apache.beam.sdk.io.hcatalog.test.EmbeddedMetastoreService
Returns the HiveConf object for the embedded metastore.
getHiveConfAsMap() - Method in class org.apache.beam.sdk.io.hcatalog.test.EmbeddedMetastoreService
 
getHoldability() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
getHost() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver.ServerConfiguration
 
getHost() - Method in class org.apache.beam.sdk.expansion.service.ExpansionServer
Get the host that this ExpansionServer is bound to.
getHost() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
getHTTPWriteTimeout() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
 
getId() - Method in interface org.apache.beam.runners.fnexecution.control.RemoteBundle
Get an id used to represent this bundle.
getId() - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor.ActiveBundle
Returns an id used to represent this bundle.
getId() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation
 
getId() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
 
getId() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
 
getId() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Order
 
getId() - Method in interface org.apache.beam.sdk.fn.IdGenerator
 
getId() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.ProjectPath
 
getId() - Method in class org.apache.beam.sdk.values.TupleTag
Returns the id of this TupleTag.
getId() - Method in class org.apache.beam.sdk.values.ValueWithRecordId
 
getIdAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
Get the id attribute.
getIdAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
Get the id attribute.
getIdentifier() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
 
getIdentifier() - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
 
getIdentifier() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
 
getIdentifier() - Method in class org.apache.beam.sdk.schemas.logicaltypes.PassThroughLogicalType
 
getIdentifier() - Method in interface org.apache.beam.sdk.schemas.Schema.LogicalType
The unique identifier for this type.
getImplementor() - Method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
 
getImplementor(boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.UdafImpl
 
getImplementor() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLCastFunctionImpl
 
getInboundObserver() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer
 
getIncompatibleGlobalWindowErrorMessage() - Method in interface org.apache.beam.sdk.transforms.CombineFnBase.GlobalCombineFn
Returns the error message for not supported default values in Combine.globally().
getIncompatibleGlobalWindowErrorMessage() - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
 
getIndex() - Method in class org.apache.beam.sdk.extensions.sql.impl.utils.SerializableRexInputRef
 
getIndex() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
 
getIndex() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
getIndex(TupleTag<?>) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
Returns the index for the given tuple tag, if the tag is present in this schema, -1 if it isn't.
getIndex() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
The zero-based index of this trigger firing that produced this pane.
getIndexes() - Method in class org.apache.beam.sdk.extensions.sql.impl.utils.SerializableRexFieldAccess
 
getInput(PTransform<InputT, ?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
 
getInput() - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
 
getInput() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
getInput(RelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils
 
getInputFile() - Method in interface org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.WordCountOptions
 
getinputFormatClass() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
 
getinputFormatKeyClass() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
 
getinputFormatValueClass() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
 
getInputReceivers() - Method in interface org.apache.beam.runners.fnexecution.control.RemoteBundle
Get a map of PCollection ids to receivers which consume input elements, forwarding them to the remote environment.
getInputReceivers() - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor.ActiveBundle
Get a map of PCollection ids to receivers which consume input elements, forwarding them to the remote environment.
getInputs(PTransform<InputT, ?>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
Returns the input of the currently being translated transform.
getInputs() - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
 
getInputs() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
getInputStream(PValue) - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
 
getInputType() - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
Returns a TypeDescriptor capturing what is known statically about the input type of this CombineFn instance's most-derived class.
getInputTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.DoFn
Returns a TypeDescriptor capturing what is known statically about the input type of this DoFn instance's most-derived class.
getInputTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.InferableFunction
Returns a TypeDescriptor capturing what is known statically about the input type of this InferableFunction instance's most-derived class.
getInputValueCoder(Coder<KV<K, V>>) - Static method in class org.apache.beam.sdk.transforms.GroupByKey
Returns the Coder of the values of the input to this transform.
getInputWatermark() - Method in class org.apache.beam.runners.direct.WatermarkManager.TransformWatermarks
Returns the input watermark of the AppliedPTransform.
getInsertBundleParallelism() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
 
getInsertDefault() - Method in class org.apache.beam.sdk.transforms.Combine.GloballyAsSingletonView
 
getInstance() - Static method in class org.apache.beam.runners.spark.aggregators.AggregatorsAccumulator
 
getInstance() - Static method in class org.apache.beam.runners.spark.metrics.MetricsAccumulator
 
getInstance() - Static method in class org.apache.beam.runners.spark.structuredstreaming.aggregators.AggregatorsAccumulator
 
getInstance() - Static method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsAccumulator
 
getInstanceId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
getInstructionId() - Method in class org.apache.beam.sdk.fn.data.LogicalEndpoint
 
getInstructionRequestHandler() - Method in class org.apache.beam.runners.fnexecution.environment.ProcessEnvironment
 
getInstructionRequestHandler() - Method in interface org.apache.beam.runners.fnexecution.environment.RemoteEnvironment
Return an InstructionRequestHandler which can communicate with the environment.
getInstructionRequestHandler() - Method in class org.apache.beam.runners.fnexecution.environment.RemoteEnvironment.SimpleRemoteEnvironment
 
getInstructionRequestHandler() - Method in class org.apache.beam.runners.fnexecution.environment.StaticRemoteEnvironment
 
getInt(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getInt(Map<String, Object>, String, Integer) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getInt16(String) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.INT16 value by field name, IllegalStateException is thrown if schema doesn't match.
getInt16(int) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.INT16 value by field index, ClassCastException is thrown if schema doesn't match.
getInt32(String) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.INT32 value by field name, IllegalStateException is thrown if schema doesn't match.
getInt32(int) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.INT32 value by field index, ClassCastException is thrown if schema doesn't match.
getInt64(String) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.INT64 value by field name, IllegalStateException is thrown if schema doesn't match.
getInt64(int) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.INT64 value by field index, ClassCastException is thrown if schema doesn't match.
getInterface() - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
 
getInterfaces() - Method in class org.apache.beam.sdk.values.TypeDescriptor
Returns a set of TypeDescriptors, one for each interface implemented by this class.
getIo() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
getIterable(String) - Method in class org.apache.beam.sdk.values.Row
Get an iterable value by field name, IllegalStateException is thrown if schema doesn't match.
getIterable(int) - Method in class org.apache.beam.sdk.values.Row
Get an iterable value by field index, IllegalStateException is thrown if schema doesn't match.
getIterableComponentType(TypeDescriptor) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
For an array T[] or a subclass of Iterable, return a TypeDescriptor describing T.
getJavaClass(RelDataType) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamJavaTypeFactory
 
getJAXBClass() - Method in class org.apache.beam.sdk.io.xml.JAXBCoder
 
getJetDefaultParallelism() - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
 
getJetLocalMode() - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
 
getJetProcessorsCooperative() - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
 
getJetServers() - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
 
getJmsCorrelationID() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
 
getJmsDeliveryMode() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
 
getJmsDestination() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
 
getJmsExpiration() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
 
getJmsMessageID() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
 
getJmsPriority() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
 
getJmsRedelivered() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
 
getJmsReplyTo() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
 
getJmsTimestamp() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
 
getJmsType() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
 
getJob(String) - Method in class org.apache.beam.runners.dataflow.DataflowClient
Gets the Dataflow Job with the given jobId.
getJob() - Method in exception org.apache.beam.runners.dataflow.DataflowJobException
Returns the failed job.
getJob() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator.JobSpecification
 
getJob(JobReference) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.JobService
Gets the specified Job by the given JobReference.
getJob(JobReference) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
 
getJobEndpoint() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
getJobId() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
Get the id of this job.
getJobId() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
The identity of the Dataflow job.
getJobId() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
 
getJobInfo() - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.BatchTranslationContext
 
getJobInfo() - Method in interface org.apache.beam.runners.flink.FlinkPortablePipelineTranslator.TranslationContext
 
getJobInfo() - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext
 
getJobMessages(String, long) - Method in class org.apache.beam.runners.dataflow.util.MonitoringUtil
Return job messages sorted in ascending order by timestamp.
getJobMetrics(String) - Method in class org.apache.beam.runners.dataflow.DataflowClient
Gets the JobMetrics with the given jobId.
getJobMetrics(JobApi.GetJobMetricsRequest, StreamObserver<JobApi.GetJobMetricsResponse>) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService
 
getJobMetrics(JobApi.GetJobMetricsRequest, StreamObserver<JobApi.GetJobMetricsResponse>) - Method in class org.apache.beam.runners.portability.testing.TestJobService
 
getJobMonitoringPageURL(String, String) - Static method in class org.apache.beam.runners.dataflow.util.MonitoringUtil
Deprecated.
this method defaults the region to "us-central1". Prefer using the overload with an explicit regionId parameter.
getJobMonitoringPageURL(String, String, String) - Static method in class org.apache.beam.runners.dataflow.util.MonitoringUtil
 
getJobName() - Method in interface org.apache.beam.sdk.options.PipelineOptions
 
getJobs(JobApi.GetJobsRequest, StreamObserver<JobApi.GetJobsResponse>) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService
 
getJobServerConfig() - Method in interface org.apache.beam.runners.portability.testing.TestPortablePipelineOptions
 
getJobServerDriver() - Method in interface org.apache.beam.runners.portability.testing.TestPortablePipelineOptions
 
getJobServerUrl() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver
 
getJobService(BigQueryOptions) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices
Returns a real, mock, or fake BigQueryServices.JobService.
getJobService(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
 
getJoinColumns(boolean, List<Pair<RexNode, RexNode>>, int, Schema) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamJoinTransforms
 
getJsonClustering() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
getJsonFactory() - Static method in class org.apache.beam.sdk.extensions.gcp.util.Transport
 
getJsonTimePartitioning() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
getKey() - Method in class org.apache.beam.runners.direct.WatermarkManager.FiredTimers
 
getKey() - Method in interface org.apache.beam.runners.local.Bundle
Returns the key that was output in the most recent GroupByKey in the execution of this bundle.
getKey() - Method in class org.apache.beam.runners.local.StructuralKey
Returns the key that this StructuralKey was created from.
getKey() - Method in class org.apache.beam.sdk.metrics.MetricResult
 
getKey() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
 
getKey() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
The key for the display item.
getKey() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
The key for the display item.
getKey() - Method in class org.apache.beam.sdk.values.KV
Returns the key of this KV.
getKey() - Method in class org.apache.beam.sdk.values.ShardedKey
 
getKeyCoder() - Method in class org.apache.beam.sdk.coders.KvCoder
 
getKeyCoder() - Method in class org.apache.beam.sdk.coders.MapCoder
 
getKeyCoder(Coder<KV<K, V>>) - Static method in class org.apache.beam.sdk.transforms.GroupByKey
Returns the Coder of the keys of the input to this transform, which is also used as the Coder of the keys of the output of this transform.
getKeyCoder() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
Returns the key Coder for all PCollections in this KeyedPCollectionTuple.
getKeyedCollections() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
Returns a list of TaggedKeyedPCollections for the PCollections contained in this KeyedPCollectionTuple.
getKeyRange() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
Returns the range of keys that will be read from the table.
getKeyRanges() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Returns the range of keys that will be read from the table.
getKeySet() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
getKeystorePassword() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
 
getKeystorePath() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
 
getKeyTranslationFunction() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
 
getKeyTypeDescriptor() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
 
getKind() - Method in class org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters
 
getKind() - Method in interface org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.PubsubClientFactory
Return the display name for this factory.
getKind() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier
 
getKindString() - Method in class org.apache.beam.sdk.io.BoundedReadFromUnboundedSource
 
getKindString() - Method in class org.apache.beam.sdk.io.Read.Bounded
 
getKindString() - Method in class org.apache.beam.sdk.io.Read.Unbounded
 
getKindString() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
 
getKindString() - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
 
getKindString() - Method in class org.apache.beam.sdk.transforms.Combine.PerKeyWithHotKeyFanout
 
getKindString() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
 
getKindString() - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
 
getKindString() - Method in class org.apache.beam.sdk.transforms.PTransform
Returns the name to use by default for this PTransform (not including the names of any enclosing PTransforms).
getKindString() - Method in class org.apache.beam.sdk.transforms.windowing.Window
 
getKindString() - Method in class org.apache.beam.sdk.values.PValueBase
Returns a String capturing the kind of this PValueBase.
getKinesisClient() - Method in interface org.apache.beam.sdk.io.kinesis.AWSClientsProvider
 
getKV() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
 
getLabel() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
Retrieve the optional label for an item.
getLabel() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
The optional label for an item.
getLabels() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
Labels that will be applied to the billing records for this job.
getLanguageOptions() - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl
 
getLastEmitted() - Method in class org.apache.beam.sdk.io.CountingSource.CounterMark
Returns the last value emitted by the reader.
getLastFieldId() - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
getLastWatermarkedBatchTime() - Static method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
 
getLatencyTrackingInterval() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getLateTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
 
getLegacyEnvironmentMajorVersion() - Method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
Provides the legacy environment's major version number.
getLength() - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
 
getLimitCountOfSortRel() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.AbstractBeamCalcRel
 
getLinkUrl() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
Retrieve the optional link URL for an item.
getLinkUrl() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
The optional link URL for an item.
getList() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier
 
getListeners() - Method in interface org.apache.beam.runners.spark.SparkContextOptions
 
getListOfMaps(Map<String, Object>, String, List<Map<String, Object>>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getLiteralGqlQuery() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
 
getLoadBalanceBundles() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
getLocalhost() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
 
getLocalValue() - Method in class org.apache.beam.runners.flink.metrics.MetricsAccumulator
 
getLocation() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table
 
getLogicalType() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
 
getLogicalType(Class<LogicalTypeT>) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
Helper function for retrieving the concrete logical type subclass.
getLogicalTypeName() - Method in class org.apache.beam.sdk.coders.AvroCoder.JodaTimestampConversion
 
getLogicalTypeValue(String, Class<T>) - Method in class org.apache.beam.sdk.values.Row
 
getLogicalTypeValue(int, Class<T>) - Method in class org.apache.beam.sdk.values.Row
Returns the Logical Type input type for this field.
getLong(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getLong(Map<String, Object>, String, Long) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getLowWatermark() - Method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.SparkWatermarks
 
getMainOutputTag() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.ParDoSingle
 
getMainOutputTag() - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
getMainOutputTag() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
 
getMainTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
The main trigger, which will continue firing until the "until" trigger fires.
getManifest(ArtifactApi.GetManifestRequest, StreamObserver<ArtifactApi.GetManifestResponse>) - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactRetrievalService
 
getManifestProxy(String) - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactRetrievalService
 
getMap() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier
 
getMap(String) - Method in class org.apache.beam.sdk.values.Row
Get a MAP value by field name, IllegalStateException is thrown if schema doesn't match.
getMap(int) - Method in class org.apache.beam.sdk.values.Row
Get a MAP value by field index, IllegalStateException is thrown if schema doesn't match.
getMapKeyType() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
If the field is a map type, returns the key type.
getMapKeyType() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
 
getMapType(TypeDescriptor, int) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
 
getMapValueType() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
If the field is a map type, returns the key type.
getMapValueType() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
 
getMatcher() - Method in class org.apache.beam.sdk.extensions.gcp.util.CustomHttpErrors.MatcherAndError
 
getMaterialization() - Method in class org.apache.beam.sdk.transforms.ViewFn
Gets the materialization of this ViewFn.
getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn
 
getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.ListViewFn
 
getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.MapViewFn
 
getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.MultimapViewFn
 
getMaterialization() - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn
 
getMax() - Method in class org.apache.beam.sdk.metrics.DistributionResult
 
getMaxBundleSize() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getMaxBundleTimeMills() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getMaxConditionCost() - Method in interface org.apache.beam.runners.dataflow.options.CloudDebuggerOptions
The maximum cost (as a ratio of CPU time) allowed for evaluating conditional snapshots.
getMaxCumulativeBackoff() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
getMaxEndOffset(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
 
getMaxEndOffset(PipelineOptions) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
Returns the actual ending offset of the current source.
getMaxInvocationHistory() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver.ServerConfiguration
 
getMaxNumericPrecision() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelDataTypeSystem
 
getMaxNumericScale() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelDataTypeSystem
 
getMaxNumWorkers() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
The maximum number of workers to use for the workerpool.
getMaxParallelism() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getMaxRecordsPerBatch() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
getMaxStreamingBatchSize() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
 
getMaxStreamingRowsToBatch() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
 
getMean() - Method in class org.apache.beam.sdk.metrics.DistributionResult
 
getMean() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
 
getMemoryMB() - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
Returns the configured size of the memory buffer.
getMemoryMB() - Method in class org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options
Returns the configured size of the memory buffer.
getMessage() - Method in exception org.apache.beam.sdk.coders.Coder.NonDeterministicException
 
getMessage() - Method in exception org.apache.beam.sdk.transforms.windowing.IncompatibleWindowException
 
getMessageBacklog() - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicy.PartitionContext
Current backlog in messages (latest offset of the partition - last processed record offset).
getMessageId() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
Returns the messageId of the message populated by Cloud Pub/Sub.
getMessageStream(JobApi.JobMessagesRequest, StreamObserver<JobApi.JobMessagesResponse>) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService
 
getMessageType() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
Returns the Protocol Buffers Message type this ProtoCoder supports.
getMetaData() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
getMetadata(DestinationT) - Method in class org.apache.beam.sdk.io.DynamicAvroDestinations
Return AVRO file metadata for a given destination.
getMetadata() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
Returns the MatchResult.Metadata of the file.
getMetadata() - Method in class org.apache.beam.sdk.io.tika.ParseResult
Returns the extracted metadata.
getMetadata(String) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
Deprecated.
use schema options instead.
getMetadataQuery() - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
 
getMetadataString(String) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
Deprecated.
use schema options instead.
getMetaStore() - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlCli
 
getMeters(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
 
getMeters(MetricFilter) - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.WithMetricsSupport
 
getMethod() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
 
getMethods(Class) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
Returns the list of non private/protected, non-static methods in the class, caching the results.
getMethodsMap(Class) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
 
getMetrics() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation
 
getMetricsContainer(String) - Method in class org.apache.beam.runners.flink.metrics.FlinkMetricContainer
 
getMetricsGraphiteHost() - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
 
getMetricsGraphitePort() - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
 
getMetricsHttpSinkUrl() - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
 
getMetricsMapName(long) - Static method in class org.apache.beam.runners.jet.metrics.JetMetricsContainer
 
getMetricsPushPeriod() - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
 
getMetricsSink() - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
 
getMimeType() - Method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
Deprecated.
 
getMimeType() - Method in interface org.apache.beam.sdk.io.FileBasedSink.OutputFileHints
Returns the MIME type that should be used for the files that will hold the output data.
getMin() - Method in class org.apache.beam.sdk.metrics.DistributionResult
 
getMinBundleSize() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
Returns the minimum bundle size that should be used when splitting the source into sub-sources.
getMinCpuPlatform() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
Specifies a Minimum CPU platform for VM instances.
getMinimumTimestamp() - Method in interface org.apache.beam.runners.local.Bundle
Return the minimum timestamp among elements in this bundle.
getMinPauseBetweenCheckpoints() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getMinReadTimeMillis() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
getMode() - Method in class org.apache.beam.sdk.io.FileBasedSource
 
getMode() - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
getModeNames() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
getModifiableCollection() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteTable
 
getMonthOfYear() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
 
getMutableOutput(TupleTag<T>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
getName(int) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
getName() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
 
getName() - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
 
getName() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table
 
getName() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.SubscriptionPath
 
getName() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.TopicPath
 
getName() - Method in class org.apache.beam.sdk.metrics.DelegatingCounter
 
getName() - Method in interface org.apache.beam.sdk.metrics.Metric
The MetricName given to this metric.
getName() - Method in class org.apache.beam.sdk.metrics.MetricName
The name of this metric.
getName() - Method in class org.apache.beam.sdk.metrics.MetricNameFilter
If set, the metric must have this name to match this MetricNameFilter.
getName() - Method in class org.apache.beam.sdk.metrics.MetricResult
Return the name of the metric.
getName() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
Returns the field name.
getName() - Method in class org.apache.beam.sdk.schemas.Schema.Field
Returns the field name.
getName() - Method in class org.apache.beam.sdk.transforms.PTransform
Returns the transform name.
getName() - Method in class org.apache.beam.sdk.values.PCollection
Returns the name of this PCollection.
getName() - Method in interface org.apache.beam.sdk.values.PValue
Returns the name of this PValue.
getName() - Method in class org.apache.beam.sdk.values.PValueBase
Returns the name of this PValueBase.
getNameCount() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
getNameOverride() - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
 
getNameOverride() - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
 
getNamespace() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
 
getNamespace() - Method in class org.apache.beam.sdk.metrics.MetricName
The namespace associated with this metric.
getNamespace() - Method in class org.apache.beam.sdk.metrics.MetricNameFilter
The inNamespace that a metric must be in to match this MetricNameFilter.
getNamespace() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
 
getNamespace() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
The namespace for the display item.
getNamespace() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
The namespace for the display item.
getNeedsAttributes() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
 
getNeedsMessageId() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
 
getNestedFieldsAccessed() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
 
getNetwork() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
GCE network for launching workers.
getNetworkTimeout() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
getNewBigquerryClient(String) - Static method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
 
getNextOffset() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
 
getNodeStats() - Method in interface org.apache.beam.sdk.extensions.sql.impl.planner.NodeStatsMetadata
 
getNodeStats(RelNode, RelMetadataQuery) - Method in interface org.apache.beam.sdk.extensions.sql.impl.planner.NodeStatsMetadata.Handler
 
getNodeStats(RelNode, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.RelMdNodeStats
 
getNodeStats(RelNode, RelMetadataQuery) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils
 
getNonCumulativeCost(RelNode, RelMetadataQuery) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner.NonCumulativeCostImpl
 
getNonSpeculativeIndex() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
The zero-based index of this trigger firing among non-speculative panes.
getNonWildcardPrefix(String) - Static method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
Returns the prefix portion of the glob that doesn't contain wildcards.
getNotSupported() - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTableFilter
Identify parts of a predicate that are not supported by the IO push-down capabilities to be preserved in a Calc following BeamIOSourceRel.
getNotSupported() - Method in class org.apache.beam.sdk.extensions.sql.meta.DefaultTableFilter
Since predicate push-down is assumed not to be supported by default - return an unchanged list of filters to be preserved.
getNotSupported() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryFilter
 
getNotSupported() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableFilter
 
getNullable() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
 
getNum() - Method in class org.apache.beam.runners.spark.io.ConsoleIO.Write.Unbound
 
getNumber() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
 
getNumber() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
 
getNumber() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
 
getNumberOfExecutionRetries() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getNumberOfWorkerHarnessThreads() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
Number of threads to use on the Dataflow worker harness.
getNumConcurrentCheckpoints() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getNumEntities(PipelineOptions, String, String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
Returns Number of entities available for reading.
getNumExtractJobCalls() - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
 
getNumQuerySplits() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
 
getNumRows(BigQueryOptions, TableReference) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
It returns the number of rows for a given table.
getNumSampledBytesPerFile() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
 
getNumShardsProvider() - Method in class org.apache.beam.sdk.io.WriteFiles
 
getNumStreamingKeys() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
 
getNumWorkers() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
Number of workers to use when executing the Dataflow job.
getObject(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getObject(Map<String, Object>, String, Map<String, Object>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getObject() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
Returns the object name associated with this GCS path, or an empty string if no object is specified.
getObject(GcsPath) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
Returns the StorageObject for the given GcsPath.
getObjectReuse() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getObjects(List<GcsPath>) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
Returns StorageObjectOrIOExceptions for the given GcsPaths.
getOffset() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
 
getOffset() - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
 
getOffset() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
 
getOffset() - Method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform.AlignTo
 
getOnCreateMatcher() - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
 
getOneOfSchema() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
Returns the schema of the underlying Row that is used to represent the union.
getOneOfTypes() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
 
getOnly(TupleTag<V>) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
If there is a singleton value for the given tag, returns it.
getOnly(String) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
Like CoGbkResult.getOnly(TupleTag) but using a String instead of a TupleTag.
getOnly(TupleTag<V>, V) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
If there is a singleton value for the given tag, returns it.
getOnly(String, V) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
Like CoGbkResult.getOnly(TupleTag, Object) but using a String instead of a TupleTag.
getOnSuccessMatcher() - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
 
getOnTimeBehavior() - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
getOperand0() - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
 
getOperandList() - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlCheckConstraint
 
getOperandList() - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlColumnDeclaration
 
getOperandList() - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlCreateExternalTable
 
getOperator() - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlCheckConstraint
 
getOperator() - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlColumnDeclaration
 
getOptionNames() - Method in class org.apache.beam.sdk.schemas.Schema.Options
 
getOptions() - Method in class org.apache.beam.sdk.Pipeline
 
getOptions() - Method in class org.apache.beam.sdk.schemas.Schema.Field
Returns the fields Schema.Options.
getOptions() - Method in class org.apache.beam.sdk.schemas.Schema
 
getOptions() - Method in class org.apache.beam.sdk.testing.TestPipeline
 
getOptionsId() - Method in interface org.apache.beam.sdk.options.PipelineOptions
Provides a process wide unique ID for this PipelineOptions object, assigned at graph construction time.
getOrCreateReader(PipelineOptions, CheckpointMarkT) - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
 
getOriginalWindowFn() - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
Returns the original windowFn that this InvalidWindows replaced.
getOutboundObserver() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer
 
getOutName(int) - Method in class org.apache.beam.sdk.values.TupleTag
If this TupleTag is tagging output outputIndex of a PTransform, returns the name that should be used by default for the output.
getOutput(PTransform<?, OutputT>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
 
getOutput() - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
 
getOutput() - Method in interface org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.WordCountOptions
 
getOutput() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
getOutput() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteResult
 
getOutputCoder() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
 
getOutputCoder(SerializableFunction<InputT, OutputT>, Coder<InputT>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
Deprecated.
This method is to change in an unknown backwards incompatible way once support for this functionality is refined.
getOutputCoder() - Method in class org.apache.beam.sdk.io.AvroSource
 
getOutputCoder() - Method in class org.apache.beam.sdk.io.CompressedSource
Returns the delegate source's output coder.
getOutputCoder() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
 
getOutputCoder() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource
 
getOutputCoder() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
 
getOutputCoder() - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
 
getOutputCoder() - Method in class org.apache.beam.sdk.io.Source
Returns the Coder to use for the data read from this source.
getOutputCoder() - Method in class org.apache.beam.sdk.io.xml.XmlSource
 
getOutputCoders() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
getOutputExecutablePath() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
getOutputFile() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
 
getOutputFilePrefix() - Method in interface org.apache.beam.sdk.extensions.sql.example.BeamSqlDataCatalogExample.DCExamplePipelineOptions
Output file prefix.
getOutputKvCoder(Coder<KV<K, V>>) - Static method in class org.apache.beam.sdk.transforms.GroupByKey
Returns the Coder of the output of this transform.
getOutputs(PTransform<?, OutputT>) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
Returns the output of the currently being translated transform.
getOutputs() - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
 
getOutputs() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
getOutputSchema(Schema, FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.utils.SelectHelpers
Get the output schema resulting from selecting the given FieldAccessDescriptor from the given schema.
getOutputStrategyInternal(WindowingStrategy<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.Window
Get the output strategy of this Window PTransform.
getOutputTime(Instant, GlobalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
 
getOutputTime(Instant, W) - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
 
getOutputTime(Instant, W) - Method in class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
 
getOutputTime(Instant, IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
Ensures that later sliding windows have an output time that is past the end of earlier windows.
getOutputTime(Instant, W) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
Returns the output timestamp to use for data depending on the given inputTimestamp in the specified window.
getOutputType() - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
Returns a TypeDescriptor capturing what is known statically about the output type of this CombineFn instance's most-derived class.
getOutputTypeDescriptor() - Method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides.BatchStatefulDoFn
 
getOutputTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.DoFn
Returns a TypeDescriptor capturing what is known statically about the output type of this DoFn instance's most-derived class.
getOutputTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.InferableFunction
Returns a TypeDescriptor capturing what is known statically about the output type of this InferableFunction instance's most-derived class.
getOutputWatermark() - Method in class org.apache.beam.runners.direct.WatermarkManager.TransformWatermarks
Returns the output watermark of the AppliedPTransform.
getOverrides() - Method in class org.apache.beam.runners.apex.ApexRunner
 
getOverrideWindmillBinary() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
Custom windmill_main binary to use with the streaming runner.
getPane() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow
Returns the pane of this ValueInSingleWindow in its window.
getPaneInfo() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
 
getParallelism() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getParallelism() - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
 
getParameters() - Method in class org.apache.beam.sdk.extensions.sql.impl.UdafImpl
 
getParameters() - Method in class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase
Returns the parameters of this function.
getParameters() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLCastFunctionImpl
 
getParent() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
Returns the parent path, or null if this path does not have a parent.
getParents() - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
 
getParser() - Method in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
Get the memoized Parser, possibly initializing it lazily.
getParser() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
Get the memoized Parser, possibly initializing it lazily.
getPartition() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
 
getPartition() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
 
getPartitionKey(byte[]) - Method in interface org.apache.beam.sdk.io.kinesis.KinesisPartitioner
 
getPartitionKey() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
getPartitions() - Method in class org.apache.beam.runners.spark.io.SourceRDD.Bounded
 
getPartitions() - Method in class org.apache.beam.runners.spark.io.SourceRDD.Unbounded
 
getPartitions() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark
 
getPassword() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
 
getPath() - Method in class org.apache.beam.sdk.extensions.sql.impl.TableName
Table path up to the leaf table name.
getPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.ProjectPath
 
getPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.SubscriptionPath
 
getPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.TopicPath
 
getPath() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
 
getPath() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
The path for the display item within a component hierarchy.
getPathValidator() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
The path validator instance that should be used to validate paths.
getPathValidatorClass() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
The class of the validator that should be created and used to validate paths.
getPayload() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
Returns the main PubSub message.
getPayload() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
 
getPCollection() - Method in interface org.apache.beam.runners.local.Bundle
Returns the PCollection that the elements of this bundle belong to.
getPCollection() - Method in interface org.apache.beam.sdk.values.PCollectionView
For internal use only.
getPCollection() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
 
getPCollectionInputs() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
 
getPCollectionInputs() - Method in interface org.apache.beam.sdk.extensions.sql.impl.rel.BeamRelNode
 
getPerDestinationOutputFilenames() - Method in class org.apache.beam.sdk.io.WriteFilesResult
Returns a PCollection of all output filenames generated by this WriteFiles organized by user destination type.
getPerElementConsumers(CollectionT) - Method in interface org.apache.beam.runners.direct.ExecutableGraph
 
getPerElementInputs(ExecutableT) - Method in interface org.apache.beam.runners.direct.ExecutableGraph
 
getPeriod() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
 
getPeriod() - Method in class org.apache.beam.sdk.transforms.windowing.TimestampTransform.AlignTo
 
getPipeline(JobApi.GetJobPipelineRequest, StreamObserver<JobApi.GetJobPipelineResponse>) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService
 
getPipeline() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation
Retrieve the job's pipeline.
getPipeline() - Method in class org.apache.beam.sdk.io.gcp.bigquery.WriteResult
 
getPipeline() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteResult
 
getPipeline() - Method in class org.apache.beam.sdk.io.WriteFilesResult
 
getPipeline() - Method in class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
 
getPipeline() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
 
getPipeline() - Method in class org.apache.beam.sdk.transforms.WithFailures.Result
 
getPipeline() - Method in class org.apache.beam.sdk.values.PBegin
 
getPipeline() - Method in class org.apache.beam.sdk.values.PCollectionList
 
getPipeline() - Method in class org.apache.beam.sdk.values.PCollectionTuple
 
getPipeline() - Method in class org.apache.beam.sdk.values.PDone
 
getPipeline() - Method in interface org.apache.beam.sdk.values.PInput
Returns the owning Pipeline of this PInput.
getPipeline() - Method in interface org.apache.beam.sdk.values.POutput
Returns the owning Pipeline of this POutput.
getPipeline() - Method in class org.apache.beam.sdk.values.PValueBase
 
getPipelineFromClasspath(String) - Static method in class org.apache.beam.runners.fnexecution.jobsubmission.PortablePipelineJarUtils
 
getPipelineOptions() - Method in class org.apache.beam.runners.apex.ApexRunnerRegistrar.Options
 
getPipelineOptions() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineRegistrar.Options
 
getPipelineOptions() - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
Returns the configured pipeline options.
getPipelineOptions() - Method in class org.apache.beam.runners.direct.DirectRegistrar.Options
 
getPipelineOptions() - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.BatchTranslationContext
 
getPipelineOptions() - Method in interface org.apache.beam.runners.flink.FlinkPortablePipelineTranslator.TranslationContext
 
getPipelineOptions() - Method in class org.apache.beam.runners.flink.FlinkRunner
For testing.
getPipelineOptions() - Method in class org.apache.beam.runners.flink.FlinkRunnerRegistrar.Options
 
getPipelineOptions() - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext
 
getPipelineOptions() - Method in class org.apache.beam.runners.flink.TestFlinkRunner
 
getPipelineOptions() - Method in class org.apache.beam.runners.gearpump.GearpumpRunnerRegistrar.Options
 
getPipelineOptions() - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
 
getPipelineOptions() - Method in class org.apache.beam.runners.jet.JetRunnerRegistrar.Options
 
getPipelineOptions() - Method in class org.apache.beam.runners.portability.testing.TestPortablePipelineOptions.TestPortablePipelineOptionsRegistrar
 
getPipelineOptions() - Method in class org.apache.beam.runners.spark.SparkRunnerRegistrar.Options
 
getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.gcp.options.GcpPipelineOptionsRegistrar
 
getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
 
getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlPipelineOptionsRegistrar
 
getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.JdbcConnection
 
getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
 
getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
 
getPipelineOptions() - Method in interface org.apache.beam.sdk.extensions.sql.impl.rel.BeamRelNode
Perform a DFS(Depth-First-Search) to find the PipelineOptions config.
getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamValuesRel
 
getPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogPipelineOptionsRegistrar
 
getPipelineOptions() - Method in class org.apache.beam.sdk.io.aws.options.AwsPipelineOptionsRegistrar
 
getPipelineOptions() - Method in class org.apache.beam.sdk.io.aws2.options.AwsPipelineOptionsRegistrar
 
getPipelineOptions() - Method in class org.apache.beam.sdk.io.gcp.common.GcpIoPipelineOptionsRegistrar
 
getPipelineOptions() - Method in class org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptionsRegistrar
 
getPipelineOptions() - Method in class org.apache.beam.sdk.options.DefaultPipelineOptionsRegistrar
 
getPipelineOptions() - Method in class org.apache.beam.sdk.options.ManualDockerEnvironmentOptions.Options
 
getPipelineOptions() - Method in interface org.apache.beam.sdk.options.PipelineOptionsRegistrar
 
getPipelineOptions() - Method in class org.apache.beam.sdk.options.RemoteEnvironmentOptions.Options
 
getPipelineOptions() - Method in interface org.apache.beam.sdk.state.StateContext
Returns the PipelineOptions specified with the PipelineRunner.
getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.CombineWithContext.Context
Returns the PipelineOptions specified with the PipelineRunner invoking this KeyedCombineFn.
getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.DoFn.FinishBundleContext
Returns the PipelineOptions specified with the PipelineRunner invoking this DoFn.
getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.DoFn.StartBundleContext
Returns the PipelineOptions specified with the PipelineRunner invoking this DoFn.
getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
Returns the PipelineOptions specified with the PipelineRunner invoking this DoFn.
getPipelineOptions() - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
getPipelineOptionsFromClasspath(String) - Static method in class org.apache.beam.runners.fnexecution.jobsubmission.PortablePipelineJarUtils
 
getPipelineProto() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator.JobSpecification
 
getPipelineRunners() - Method in class org.apache.beam.runners.apex.ApexRunnerRegistrar.Runner
 
getPipelineRunners() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineRegistrar.Runner
 
getPipelineRunners() - Method in class org.apache.beam.runners.direct.DirectRegistrar.Runner
 
getPipelineRunners() - Method in class org.apache.beam.runners.flink.FlinkRunnerRegistrar.Runner
 
getPipelineRunners() - Method in class org.apache.beam.runners.gearpump.GearpumpRunnerRegistrar.Runner
 
getPipelineRunners() - Method in class org.apache.beam.runners.jet.JetRunnerRegistrar.Runner
 
getPipelineRunners() - Method in class org.apache.beam.runners.portability.PortableRunnerRegistrar
 
getPipelineRunners() - Method in class org.apache.beam.runners.spark.SparkRunnerRegistrar.Runner
 
getPipelineUrl() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
The URL of the staged portable pipeline.
getPlanner() - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
 
getPlannerName() - Method in interface org.apache.beam.sdk.extensions.sql.impl.BeamSqlPipelineOptions
 
getPort() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver.ServerConfiguration
 
getPort() - Method in class org.apache.beam.sdk.expansion.service.ExpansionServer
Get the port that this ExpansionServer is bound to.
getPort() - Method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortRead
 
getPort() - Method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortWrite
 
getPositionForFractionConsumed(double) - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
Returns a position P such that the range [start, P) represents approximately the given fraction of the range [start, end).
getPrefix() - Method in class org.apache.beam.sdk.extensions.sql.impl.TableName
First element in the path.
getPrimary() - Method in class org.apache.beam.sdk.transforms.splittabledofn.SplitResult
Returns the primary restriction.
getProcessBundleDescriptor(String) - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClient
 
getProcessBundleDescriptor(BeamFnApi.GetProcessBundleDescriptorRequest, StreamObserver<BeamFnApi.ProcessBundleDescriptor>) - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService
 
getProcessBundleDescriptor() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
 
getProcessBundleDescriptor() - Method in interface org.apache.beam.runners.fnexecution.control.StageBundleFactory
 
getProcessingTimeAdvance() - Method in class org.apache.beam.sdk.testing.TestStream.ProcessingTimeEvent
 
getProcessor(BeamFnApi.ProcessBundleDescriptor, List<RemoteInputDestination>) - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient
Provides SdkHarnessClient.BundleProcessor that is capable of processing bundles not containing timers or state accesses such as: Side inputs User state Remote references
getProcessor(BeamFnApi.ProcessBundleDescriptor, List<RemoteInputDestination>, StateDelegator) - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient
Provides SdkHarnessClient.BundleProcessor that is capable of processing bundles not containing timers.
getProcessor(BeamFnApi.ProcessBundleDescriptor, List<RemoteInputDestination>, StateDelegator, Map<String, Map<String, ProcessBundleDescriptors.TimerSpec>>) - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient
Provides SdkHarnessClient.BundleProcessor that is capable of processing bundles containing timers and state accesses such as: Side inputs User state Remote references
getProduced(ExecutableT) - Method in interface org.apache.beam.runners.direct.ExecutableGraph
 
getProducer(PValue) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
Get the AppliedPTransform that produced the provided PValue.
getProducer(CollectionT) - Method in interface org.apache.beam.runners.direct.ExecutableGraph
 
getProfilingAgentConfiguration() - Method in interface org.apache.beam.runners.dataflow.options.DataflowProfilingOptions
 
getProject() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
 
getProject() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
Project id to use when launching jobs.
getProject() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
Get the project path.
getProjectId() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
Get the project this job exists in.
getProjectId() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
 
getProjectId() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
 
getProjectId() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
getProjectId() - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher.TableAndQuery
 
getProperties() - Method in class org.apache.beam.runners.dataflow.DataflowRunnerInfo
 
getProperties() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
getProperties() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table
 
getProperties() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
 
getProvidedSparkContext() - Method in interface org.apache.beam.runners.spark.SparkContextOptions
 
getProviderRuntimeValues() - Method in interface org.apache.beam.sdk.testing.TestPipeline.TestValueProviderOptions
 
getProvisionInfo(ProvisionApi.GetProvisionInfoRequest, StreamObserver<ProvisionApi.GetProvisionInfoResponse>) - Method in class org.apache.beam.runners.fnexecution.provisioning.StaticGrpcProvisionService
 
getProxyConfiguration() - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
The client configuration instance that should be used to configure AWS service clients.
getPTransformForInput() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaCSVTable
 
getPTransformForInput() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
 
getPTransformForOutput() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaCSVTable
 
getPTransformForOutput() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
 
getPTransformId() - Method in class org.apache.beam.runners.fnexecution.data.RemoteInputDestination
 
getPubsubRootUrl() - Method in interface org.apache.beam.sdk.io.gcp.pubsub.PubsubOptions
Root URL for use with the Google Cloud Pub/Sub API.
getQualifiers() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor
 
getQuery() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
 
getQuery() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
getQuery() - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher.TableAndQuery
 
getQueryString() - Method in interface org.apache.beam.sdk.extensions.sql.example.BeamSqlDataCatalogExample.DCExamplePipelineOptions
SQL Query.
getRange() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
Returns the current range.
getRate() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
 
getRate() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
 
getRateLimitPolicy() - Method in interface org.apache.beam.sdk.io.kinesis.RateLimitPolicyFactory
 
getRawType() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
Returns the raw class type.
getRawType() - Method in class org.apache.beam.sdk.values.TypeDescriptor
Returns the Class underlying the Type represented by this TypeDescriptor.
getReaderContext(Map<String, String>) - Static method in class org.apache.beam.sdk.io.hcatalog.test.HCatalogIOTestUtils
Returns a ReaderContext instance for the passed datastore config params.
getReadTime() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
getReadTimePercentage() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
getReason() - Method in exception org.apache.beam.sdk.coders.CannotProvideCoderException
 
getReasons() - Method in exception org.apache.beam.sdk.coders.Coder.NonDeterministicException
 
getReceiver() - Method in class org.apache.beam.runners.fnexecution.control.RemoteOutputReceiver
 
getRecord() - Method in class org.apache.beam.sdk.io.gcp.bigquery.SchemaAndRecord
 
getRecordType() - Method in class org.apache.beam.sdk.coders.SerializableCoder
 
getReferentialConstraints() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
 
getRegion() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
Get the region this job exists in.
getRegion() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
The Google Compute Engine region for creating Dataflow jobs.
getRegion() - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
AWS region used by the AWS client.
getRegisteredOptions() - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
 
getRegistrationFuture() - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor
 
getRelList() - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
 
getRemote() - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
 
getRemoteInputDestinations() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
Get RemoteInputDestinations that input data are sent to the BeamFnApi.ProcessBundleDescriptor over.
getRemoteOutputCoders() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
Get all of the transforms materialized by this ProcessBundleDescriptors.ExecutableProcessBundleDescriptor and the Java Coder for the wire format of that transform.
getRepeatedTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
 
getReplacedByJob() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
Returns a new DataflowPipelineJob for the job that replaced this one, if applicable.
getReplacedByJob() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
 
getReplacementTransform(AppliedPTransform<PCollection<? extends InputT>, PCollection<OutputT>, ParDo.SingleOutput<InputT, OutputT>>) - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory
 
getReplacementTransform(AppliedPTransform<PCollection<? extends InputT>, PCollectionTuple, PTransform<PCollection<? extends InputT>, PCollectionTuple>>) - Method in class org.apache.beam.runners.direct.ParDoMultiOverrideFactory
 
getRequirements() - Method in class org.apache.beam.sdk.transforms.Contextful
Returns the requirements needed to run the closure.
getResidual() - Method in class org.apache.beam.sdk.transforms.splittabledofn.SplitResult
Returns the residual restriction.
getRetainDockerContainers() - Method in interface org.apache.beam.sdk.options.ManualDockerEnvironmentOptions
 
getRetainExternalizedCheckpointsOnCancellation() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getRetrievalServiceType() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
getRetrievalToken(String) - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactStagingService
 
getRetrievalToken(String) - Method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService
 
getRetrievalToken(String) - Method in class org.apache.beam.runners.fnexecution.artifact.JavaFilesystemLegacyArtifactStagingService
 
getReturnType(RelDataTypeFactory) - Method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
 
getReturnType(RelDataTypeFactory, SqlOperatorBinding) - Method in class org.apache.beam.sdk.extensions.sql.impl.ScalarFunctionImpl
 
getReturnType(RelDataTypeFactory) - Method in class org.apache.beam.sdk.extensions.sql.impl.UdafImpl
 
getRoot() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
getRootCause() - Method in exception org.apache.beam.sdk.coders.CannotProvideCoderException
Returns the inner-most CannotProvideCoderException when they are deeply nested.
getRootSchema() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
getRootTransforms() - Method in interface org.apache.beam.runners.direct.ExecutableGraph
 
getRow() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertError
 
getRow(String) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.ROW value by field name, IllegalStateException is thrown if schema doesn't match.
getRow(int) - Method in class org.apache.beam.sdk.values.Row
Get a Row value by field index, IllegalStateException is thrown if schema doesn't match.
getRowCount() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
 
getRowCount() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
 
getRowReceiver(TupleTag<T>) - Method in interface org.apache.beam.sdk.transforms.DoFn.MultiOutputReceiver
Returns a DoFn.OutputReceiver for publishing Row objects to the given tag.
getRows() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
getRows() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider.TableWithRows
 
getRowSchema() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
 
getRowSelector(Schema, FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.utils.SelectHelpers
 
getRowSelectorOptimized(Schema, FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.utils.SelectHelpers
 
getRowsWritten() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteResult
The number of rows written in this batch.
getRowToGenericRecordFunction(Schema) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
getRowType(RelDataTypeFactory) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteTable
 
getRule() - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
 
getRuleIndex() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierContext
 
getRuleIndex() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionComponentContext
 
getRuleIndex() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.DotExpressionContext
 
getRuleIndex() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.FieldSpecifierContext
 
getRuleIndex() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierContext
 
getRuleIndex() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifiedComponentContext
 
getRuleIndex() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifierListContext
 
getRuleNames() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
getRuleNames() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
getRuleSets() - Static method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRuleSets
 
getRunMillis() - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
 
getRunner() - Method in interface org.apache.beam.sdk.options.PipelineOptions
The pipeline runner that will be used to execute the pipeline.
getS3ClientFactoryClass() - Method in interface org.apache.beam.sdk.io.aws.options.S3Options
 
getS3StorageClass() - Method in interface org.apache.beam.sdk.io.aws.options.S3Options
 
getS3ThreadPoolSize() - Method in interface org.apache.beam.sdk.io.aws.options.S3Options
 
getS3UploadBufferSizeBytes() - Method in interface org.apache.beam.sdk.io.aws.options.S3Options
 
getSamplingStrategy() - Method in class org.apache.beam.sdk.io.TextRowCountEstimator
 
getSaveHeapDumpsToGcsPath() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
CAUTION: This option implies dumpHeapOnOOM, and has similar caveats.
getSavepointPath() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
getSaveProfilesToGcs() - Method in interface org.apache.beam.runners.dataflow.options.DataflowProfilingOptions
 
getScan() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
 
getSchema() - Method in class org.apache.beam.sdk.coders.AvroCoder
Returns the schema used by this coder.
getSchema() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDynamicMessageSchema
 
getSchema() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
getSchema() - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTable
Get the schema info of the table.
getSchema() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
 
getSchema() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonTableProvider.PubsubIOTableConfiguration
Table schema, describes Pubsub message schema.
getSchema() - Method in class org.apache.beam.sdk.extensions.sql.meta.SchemaBaseBeamTable
 
getSchema() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table
 
getSchema(DestinationT) - Method in class org.apache.beam.sdk.io.DynamicAvroDestinations
Return an AVRO schema for a given destination.
getSchema() - Method in class org.apache.beam.sdk.io.gcp.bigquery.AvroWriteRequest
 
getSchema(DestinationT) - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
Returns the table schema for the destination.
getSchema() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
Returns the schema associated with this type.
getSchema(Class<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Retrieve a Schema for a given Class type.
getSchema(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Retrieve a Schema for a given TypeDescriptor type.
getSchema(Class<T>, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
 
getSchema() - Method in class org.apache.beam.sdk.schemas.utils.ReflectUtils.ClassWithSchema
 
getSchema() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
 
getSchema() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
Returns the schema used by this CoGbkResult.
getSchema() - Method in class org.apache.beam.sdk.values.PCollection
Returns the attached schema.
getSchema() - Method in class org.apache.beam.sdk.values.Row.Builder
Return the schema for the row being built.
getSchema() - Method in class org.apache.beam.sdk.values.Row.FieldValueBuilder
 
getSchema() - Method in class org.apache.beam.sdk.values.Row
Return Schema which describes the fields.
getSchemaCoder(Class<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Retrieve a SchemaCoder for a given Class type.
getSchemaCoder(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Retrieve a SchemaCoder for a given TypeDescriptor type.
getSchemaProviders() - Method in class org.apache.beam.sdk.schemas.annotations.DefaultSchema.DefaultSchemaProviderRegistrar
 
getSchemaProviders() - Method in interface org.apache.beam.sdk.schemas.SchemaProviderRegistrar
Returns a list of schema providers which will be registered by default within each schema registry instance.
getSchemaRegistry() - Method in class org.apache.beam.sdk.Pipeline
 
getScheme() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
 
getScheme() - Method in class org.apache.beam.sdk.io.FileSystem
Get the URI scheme which defines the namespace of the FileSystem.
getScheme() - Method in interface org.apache.beam.sdk.io.fs.ResourceId
Get the scheme which defines the namespace of the ResourceId.
getSdkComponents() - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
 
getSdkHarnessLogLevelOverrides() - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
This option controls the log levels for specifically named loggers.
getSdkWorkerId() - Method in interface org.apache.beam.runners.fnexecution.HeaderAccessor
This method should be called from the request method.
getSdkWorkerParallelism() - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
getSemiPersistDir() - Method in interface org.apache.beam.sdk.options.RemoteEnvironmentOptions
 
getSequenceNumber() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
getSerializableFunctionUdfs() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.UdfUdafProvider
For UDFs implement SerializableFunction.
getSerializableOptions() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
getSerializedATN() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
getSerializedATN() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
getSerializers() - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
 
getServer() - Method in class org.apache.beam.runners.fnexecution.GrpcFnServer
Get the underlying Server contained by this GrpcFnServer.
getServerFactory() - Method in class org.apache.beam.runners.fnexecution.environment.DockerEnvironmentFactory.Provider
 
getServerFactory() - Method in class org.apache.beam.runners.fnexecution.environment.EmbeddedEnvironmentFactory.Provider
 
getServerFactory() - Method in interface org.apache.beam.runners.fnexecution.environment.EnvironmentFactory.Provider
Create the ServerFactory applicable to this environment.
getServerFactory() - Method in class org.apache.beam.runners.fnexecution.environment.ExternalEnvironmentFactory.Provider
 
getService() - Method in class org.apache.beam.runners.fnexecution.GrpcFnServer
Get the service exposed by this GrpcFnServer.
getServiceAccount() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
Run the job as a specific service account, instead of the default GCE robot.
getSessionId() - Method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService.StagingSessionToken
Access is public for json conversion.
getSetFieldCreator(Class<T>, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
 
getSetters(Class<?>, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
Return the list of FieldValueSetters for a Java Bean class
getSetters(Class<?>, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
 
getSetTimers() - Method in class org.apache.beam.runners.direct.WatermarkManager.TimerUpdate
 
getShard() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
 
getShardId() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
getShardingFunction() - Method in class org.apache.beam.sdk.io.WriteFiles
 
getShardNumber() - Method in class org.apache.beam.sdk.values.ShardedKey
 
getShortValue() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
Return the optional short value for an item, or null if none is provided.
getShortValue() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
The optional short value for an item, or null if none is provided.
getSideInput(String) - Method in interface org.apache.beam.runners.fnexecution.translation.BatchSideInputHandlerFactory.SideInputGetter
 
getSideInputDataSet(PCollectionView<?>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
getSideInputs() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.ParDoSingle
 
getSideInputs(ExecutableStage) - Static method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors
 
getSideInputs() - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
Override to specify that this object needs access to one or more side inputs.
getSideInputs() - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
Specifies that this object needs access to one or more side inputs.
getSideInputs() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
Returns the side inputs used by this Combine operation.
getSideInputs() - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
 
getSideInputs() - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
Returns the side inputs used by this Combine operation.
getSideInputs() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
 
getSideInputs() - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
 
getSideInputs() - Method in class org.apache.beam.sdk.transforms.Requirements
The side inputs that this Contextful needs access to.
getSideInputSpecs() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
Get a mapping from PTransform id to side input id to side inputs that are used during execution.
getSideInputWindow(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.WindowMappingFn
Returns the window of the side input corresponding to the given window of the main input.
getSingleFileMetadata() - Method in class org.apache.beam.sdk.io.FileBasedSource
Returns the information about the single file that this source is reading from.
getSingleWorkerStatus(String, long, TimeUnit) - Method in class org.apache.beam.runners.fnexecution.status.BeamWorkerStatusGrpcService
Get the latest SDK worker status from the client's corresponding SDK harness.
getSink() - Method in interface org.apache.beam.runners.fnexecution.control.ControlClientPool
Sink for control clients.
getSink() - Method in class org.apache.beam.runners.fnexecution.control.MapControlClientPool
 
getSink() - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
Returns the FileBasedSink for this write operation.
getSink() - Method in class org.apache.beam.sdk.io.WriteFiles
 
getSize() - Method in class org.apache.beam.sdk.schemas.utils.AvroUtils.FixedBytesField
Get the size.
getSize() - Method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
 
getSize() - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
 
getSize() - Method in interface org.apache.beam.sdk.transforms.splittabledofn.Sizes.HasSize
A representation for the amount of known work represented as a size.
getSize() - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
 
getSize() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
 
getSizeLimit(PipelineOptions) - Static method in interface org.apache.beam.sdk.fn.data.BeamFnDataBufferingOutboundObserver
 
getSketchFromByteBuffer(ByteBuffer) - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount
Converts the passed-in sketch from ByteBuffer to byte[], mapping null ByteBuffers (representing empty sketches) to empty byte[]s.
getSnsClient() - Method in interface org.apache.beam.sdk.io.aws2.sns.SnsClientProvider
 
getSocketTimeout() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
 
getSorterType() - Method in class org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options
Returns the sorter type.
getSource() - Method in interface org.apache.beam.runners.fnexecution.control.ControlClientPool
Source of control clients.
getSource() - Method in class org.apache.beam.runners.fnexecution.control.MapControlClientPool
 
getSource() - Method in class org.apache.beam.sdk.io.Read.Bounded
Returns the BoundedSource used to create this Read PTransform.
getSource() - Method in class org.apache.beam.sdk.io.Read.Unbounded
Returns the UnboundedSource used to create this Read PTransform.
getSource() - Method in class org.apache.beam.sdk.io.TextIO.Read
 
getSource() - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
 
getSourceStream(DataSource) - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
 
getSparkMaster() - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
 
getSparkSession() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
getSplit() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.SerializableSplit
 
getSplitBacklogBytes() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
Returns the size of the backlog of unread data in the underlying data source represented by this split of this source.
getSplitPointsConsumed() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
Returns the total amount of parallelism in the consumed (returned and processed) range of this reader's current BoundedSource (as would be returned by BoundedSource.BoundedReader.getCurrentSource()).
getSplitPointsConsumed() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
 
getSplitPointsConsumed() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
 
getSplitPointsConsumed() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
 
getSplitPointsProcessed() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
Returns the total number of split points that have been processed.
getSplitPointsRemaining() - Method in class org.apache.beam.sdk.io.AvroSource.AvroReader
 
getSplitPointsRemaining() - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
Returns the total amount of parallelism in the unprocessed part of this reader's current BoundedSource (as would be returned by BoundedSource.BoundedReader.getCurrentSource()).
getSplitPointsRemaining() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
 
getSplitPointsRemaining() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
 
getSqsClient() - Method in interface org.apache.beam.sdk.io.aws2.sqs.SqsClientProvider
 
getSSEAlgorithm() - Method in interface org.apache.beam.sdk.io.aws.options.S3Options
 
getSSEAwsKeyManagementParams() - Method in interface org.apache.beam.sdk.io.aws.options.S3Options
 
getSSECustomerKey() - Method in interface org.apache.beam.sdk.io.aws.options.S3Options
 
getStableUniqueNames() - Method in interface org.apache.beam.sdk.options.PipelineOptions
Whether to check for stable unique names on each transform.
getStageBundleFactory(ExecutableStage) - Method in class org.apache.beam.runners.fnexecution.control.DefaultExecutableStageContext
 
getStageBundleFactory(ExecutableStage) - Method in interface org.apache.beam.runners.fnexecution.control.ExecutableStageContext
 
getStager() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
The resource stager instance that should be used to stage resources.
getStagerClass() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
The class responsible for staging resources to be accessible by workers during job execution.
getStagingLocation() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
GCS path for staging local files, e.g.
getStartDate() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
 
getStartDate() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
 
getStartDate() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
 
getStartKey() - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
Returns the ByteKey representing the lower bound of this ByteKeyRange.
getStartOffset() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
Returns the starting offset of the source.
getStartPosition() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
 
getStartPosition() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
 
getStartPosition() - Method in interface org.apache.beam.sdk.io.range.RangeTracker
Returns the starting position of the current range, inclusive.
getStartTime() - Method in class org.apache.beam.sdk.io.CountingSource.CounterMark
Returns the time the reader was started.
getState() - Method in class org.apache.beam.runners.apex.ApexRunnerResult
 
getState() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
 
getState() - Method in class org.apache.beam.runners.dataflow.util.DataflowTemplateJob
 
getState() - Method in class org.apache.beam.runners.direct.DirectRunner.DirectPipelineResult
 
getState() - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
 
getState() - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
 
getState(JobApi.GetJobStateRequest, StreamObserver<JobApi.JobStateEvent>) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService
 
getState() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation
Retrieve the job's current state.
getState() - Method in class org.apache.beam.runners.gearpump.GearpumpPipelineResult
 
getState() - Method in class org.apache.beam.runners.jet.FailedRunningPipelineResults
 
getState() - Method in class org.apache.beam.runners.jet.JetPipelineResult
 
getState(JobApi.GetJobStateRequest, StreamObserver<JobApi.JobStateEvent>) - Method in class org.apache.beam.runners.portability.testing.TestJobService
 
getState() - Method in class org.apache.beam.runners.spark.SparkPipelineResult
 
getState() - Method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineResult
 
getState() - Method in interface org.apache.beam.sdk.PipelineResult
Retrieves the current state of the pipeline execution.
getState() - Method in interface org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimator
Get current state of the WatermarkEstimator instance, which can be used to recreate the WatermarkEstimator when processing the restriction.
getState() - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.Manual
 
getState() - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.MonotonicallyIncreasing
 
getState() - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.WallTime
 
getStateBackendFactory() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
State backend to store Beam's state during computation.
getStateCoder() - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
Used to encode the state of this Watch.Growth.TerminationCondition.
getStateEvent() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation
Retrieve the job's current state.
getStateStream(JobApi.GetJobStateRequest, StreamObserver<JobApi.JobStateEvent>) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService
 
getStaticCreator(Class, Method, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
 
getStaticCreator(Class, Method, Schema, FieldValueTypeSupplier, ByteBuddyUtils.TypeConversionsFactory) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
 
getStatistic() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteTable
 
getStepNames() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator.JobSpecification
Returns the mapping of AppliedPTransforms to the internal step name for that AppliedPTransform.
getStopPipelineWatermark() - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
 
getStopPosition() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
 
getStopPosition() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
 
getStopPosition() - Method in interface org.apache.beam.sdk.io.range.RangeTracker
Returns the ending position of the current range, exclusive.
getStorageClient(BigQueryOptions) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices
Returns a real, mock, or fake BigQueryServices.StorageClient.
getStorageClient(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
 
getStorageLevel() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
getStreamName() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
getString(Map<String, Object>, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getString(Map<String, Object>, String, String) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getString(String) - Method in class org.apache.beam.sdk.values.Row
Get a Schema.TypeName.STRING value by field name, IllegalStateException is thrown if schema doesn't match.
getString(int) - Method in class org.apache.beam.sdk.values.Row
Get a String value by field index, ClassCastException is thrown if schema doesn't match.
getStrings(Map<String, Object>, String, List<String>) - Static method in class org.apache.beam.runners.dataflow.util.Structs
 
getSubnetwork() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
GCE subnetwork for launching workers.
getSubProvider(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.FullNameTableProvider
 
getSubProvider(String) - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.TableProvider
Returns a sub-provider, e.g.
getSubProviders() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.TableProvider
Returns all sub-providers, e.g.
getSubSchema(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
getSubSchemaNames() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
getSubscription() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
Get the subscription being read from.
getSubscriptionProvider() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
Get the ValueProvider for the subscription being read from.
getSubSequenceNumber() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
getSuggestedFilenameSuffix() - Method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
Deprecated.
 
getSuggestedFilenameSuffix() - Method in interface org.apache.beam.sdk.io.FileBasedSink.OutputFileHints
 
getSuggestedSuffix() - Method in enum org.apache.beam.sdk.io.Compression
 
getSum() - Method in class org.apache.beam.sdk.metrics.DistributionResult
 
getSum() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
 
getSumAndReset() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
 
getSupertype(Class<? super T>) - Method in class org.apache.beam.sdk.values.TypeDescriptor
Returns the generic form of a supertype.
getSupported() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryFilter
 
getSupported() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableFilter
 
getSupportedClass() - Method in interface org.apache.beam.runners.dataflow.util.CloudObjectTranslator
Gets the class this CloudObjectTranslator is capable of converting.
getSupportedClass() - Method in class org.apache.beam.runners.dataflow.util.SchemaCoderCloudObjectTranslator
 
getSynchronizedProcessingInputTime() - Method in class org.apache.beam.runners.direct.WatermarkManager.TransformWatermarks
Returns the synchronized processing input time of the AppliedPTransform.
getSynchronizedProcessingOutputTime() - Method in class org.apache.beam.runners.direct.WatermarkManager.TransformWatermarks
Returns the synchronized processing output time of the AppliedPTransform.
getSynchronizedProcessingOutputWatermark() - Method in interface org.apache.beam.runners.local.Bundle
Returns the processing time output watermark at the time the producing Executable committed this bundle.
getSynchronizedProcessingTime() - Method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.SparkWatermarks
 
getTable(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
getTable(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
 
getTable(String) - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.TableProvider
Get a specific table from this provider it is present, or null if it is not present.
getTable() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertError
 
getTable() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
Returns the table to read, or null if reading from a query instead.
getTable() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
getTable() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Returns the table reference, or null.
getTable(TableReference) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
Gets the specified Table resource by table ID.
getTable(TableReference, List<String>) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
 
getTable(DestinationT) - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
Returns a TableDestination object for the destination.
getTable() - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
getTable(TableReference) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
getTable(TableReference, List<String>) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
getTableByFullName(TableName) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
 
getTableByFullName(TableName) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.FullNameTableProvider
 
getTableDescription() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
getTableId() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Returns the table being read from.
getTableId() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
 
getTableId() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
 
getTableName() - Method in class org.apache.beam.sdk.extensions.sql.impl.TableName
Table name, the last element of the fully-specified table name with path.
getTableNames() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
getTablePath(Table) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
Returns a full table path (exlucding top-level schema) for a given ZetaSQL Table.
getTableProvider() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
getTableProvider() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
Returns the table to read, or null if reading from a query instead.
getTableProvider() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
getTableReference() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
getTableResource(String, String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
 
getTables() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
 
getTables() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.InMemoryMetaTableProvider
 
getTables() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.ReadOnlyTableProvider
 
getTables() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.TableProvider
Get all tables from this provider.
getTables() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
 
getTables() - Method in class org.apache.beam.sdk.extensions.sql.meta.store.InMemoryMetaStore
 
getTableSchema(String, String) - Static method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO
Returns TableSchema for a given table.
getTableSchema() - Method in class org.apache.beam.sdk.io.gcp.bigquery.SchemaAndRecord
 
getTableSchema(String, String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogBeamSchema
Gets the table schema, or absent optional if the table doesn't exist in the database.
getTableSpec() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamPCollectionTable
 
getTableStatistics(PipelineOptions) - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTable
Estimates the number of rows or the rate for unbounded Tables.
getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.avro.AvroTable
 
getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
 
getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
 
getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable
 
getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.parquet.ParquetTable
 
getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestBoundedTable
 
getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestUnboundedTable
 
getTableStatistics(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTable
 
getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.avro.AvroTableProvider
 
getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryTableProvider
 
getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogTableProvider
 
getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.datastore.DataStoreV1TableProvider
 
getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.KafkaTableProvider
 
getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTableProvider
 
getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.parquet.ParquetTableProvider
 
getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonTableProvider
 
getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.ReadOnlyTableProvider
 
getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.seqgen.GenerateSequenceTableProvider
 
getTableType() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.TableProvider
Gets the table type this provider handles.
getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
 
getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTableProvider
 
getTableType() - Method in class org.apache.beam.sdk.extensions.sql.meta.store.InMemoryMetaStore
 
getTag(int) - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
Returns the tuple tag at the given index.
getTag() - Method in class org.apache.beam.sdk.values.TaggedPValue
Returns the local tag associated with the PValue.
getTagInternal() - Method in interface org.apache.beam.sdk.values.PCollectionView
Deprecated.
this method will be removed entirely. The PCollection underlying a side input, is part of the side input's specification with a ParDo transform, which will obtain that information via a package-private channel.
getTagInternal() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
Returns a unique TupleTag identifying this PCollectionView.
getTagsToSideInputs(Collection<PCollectionView<?>>) - Static method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils
 
getTargetDataset() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.TestBigQueryOptions
 
getTargetParallelism() - Method in interface org.apache.beam.runners.direct.DirectOptions
 
getTargetTable(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageQuerySource
 
getTargetTable(BigQueryOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageTableSource
 
getTempDatasetId() - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
 
getTempDirectoryProvider() - Method in class org.apache.beam.sdk.io.FileBasedSink
Returns the directory inside which temprary files will be written according to the configured FileBasedSink.FilenamePolicy.
getTempFilename() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
 
getTemplateLocation() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
Where the runner should generate a template file.
getTempLocation() - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
Returns the configured temporary location.
getTempLocation() - Method in class org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options
Returns the configured temporary location.
getTempLocation() - Method in interface org.apache.beam.sdk.options.PipelineOptions
A pipeline level default location for storing temporary files.
getTempRoot() - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
 
GetterBasedSchemaProvider - Class in org.apache.beam.sdk.schemas
A SchemaProvider base class that vends schemas and rows based on FieldValueGetters.
GetterBasedSchemaProvider() - Constructor for class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
 
GetterTypeSupplier() - Constructor for class org.apache.beam.sdk.schemas.JavaBeanSchema.GetterTypeSupplier
 
getTestMode() - Method in interface org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineOptions
Set to true to run the job in test mode.
getTestTimeoutSeconds() - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
 
getTimeDomain() - Method in interface org.apache.beam.sdk.state.TimerSpec
 
getTimeLimit(PipelineOptions) - Static method in interface org.apache.beam.sdk.fn.data.BeamFnDataBufferingOutboundObserver
 
getTimePartitioning() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
getTimerFamilyId() - Method in class org.apache.beam.sdk.fn.data.LogicalEndpoint
 
getTimerReceivers() - Method in interface org.apache.beam.runners.fnexecution.control.RemoteBundle
Get a map of (transform id, timer id) to receivers which consume timers, forwarding them to the remote environment.
getTimerReceivers() - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor.ActiveBundle
 
getTimers() - Method in class org.apache.beam.runners.direct.WatermarkManager.FiredTimers
Gets all of the timers that have fired within the provided TimeDomain.
getTimers(MetricFilter) - Method in class org.apache.beam.runners.spark.metrics.WithMetricsSupport
 
getTimers() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
 
getTimers(MetricFilter) - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.WithMetricsSupport
 
getTimerSpec() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.TimerSpec
 
getTimerSpecs() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
Get a mapping from PTransform id to timer id to timer specs that are used during execution.
getTimes() - Method in class org.apache.beam.runners.spark.io.CreateStream
Get times so they can be pushed into the GlobalWatermarkHolder.
getTimestamp(T, Instant) - Method in interface org.apache.beam.sdk.io.kafka.KafkaPublishTimestampFunction
Returns timestamp for element being published to Kafka.
getTimestamp() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
 
getTimestamp() - Method in class org.apache.beam.sdk.metrics.GaugeResult.EmptyGaugeResult
 
getTimestamp() - Method in class org.apache.beam.sdk.metrics.GaugeResult
 
getTimestamp() - Method in class org.apache.beam.sdk.values.TimestampedValue
 
getTimestamp() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow
Returns the timestamp of this ValueInSingleWindow.
getTimestampAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
Get the timestamp attribute.
getTimestampAttribute() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
Get the timestamp attribute.
getTimestampCombiner() - Method in interface org.apache.beam.sdk.state.WatermarkHoldState
Return the TimestampCombiner which will be used to determine a watermark hold time given an element timestamp, and to combine watermarks from windows which are about to be merged.
getTimestampCombiner() - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
getTimestampForRecord(TimestampPolicy.PartitionContext, KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.CustomTimestampPolicyWithLimitedDelay
 
getTimestampForRecord(TimestampPolicy.PartitionContext, KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicy
Returns record timestamp (aka event time).
getTimestampForRecord(TimestampPolicy.PartitionContext, KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.LogAppendTimePolicy
 
getTimestampForRecord(TimestampPolicy.PartitionContext, KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.ProcessingTimePolicy
 
getTimestampForRecord(TimestampPolicy.PartitionContext, KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.TimestampFnPolicy
 
getTimestampTransforms() - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
The transforms applied to the arrival time of an element to determine when this trigger allows output.
getTimestampType() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
 
getTimeZone() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
 
getTimeZone() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
 
getTimeZone() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
 
getTiming() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
Return the timing of this pane.
getTo() - Method in class org.apache.beam.sdk.io.range.OffsetRange
 
getTokenNames() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
Deprecated.
getTokenNames() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
Deprecated.
getTopic() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
Get the topic being written to.
getTopic() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
Get the topic being read from.
getTopic() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
 
getTopic() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
 
getTopicProvider() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
Get the ValueProvider for the topic being written to.
getTopicProvider() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
Get the ValueProvider for the topic being read from.
getTopics() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
 
getToRowFunction() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDynamicMessageSchema
 
getToRowFunction() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
Returns the fromRow conversion function.
getToRowFunction(Class<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Retrieve the function that converts an object of the specified type to a Row object.
getToRowFunction(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Retrieve the function that converts an object of the specified type to a Row object.
getToRowFunction(Class<T>, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
 
getToRowFunction() - Method in class org.apache.beam.sdk.values.PCollection
Returns the attached schema's toRowFunction.
getTotalBacklogBytes() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
Returns the size of the backlog of unread data in the underlying data source represented by all splits of this source.
getTraitDef() - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
 
getTransactionIsolation() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
getTransform(RunnerApi.FunctionSpec) - Method in interface org.apache.beam.sdk.expansion.service.ExpansionService.TransformProvider
 
getTransformId() - Method in class org.apache.beam.sdk.fn.data.LogicalEndpoint
 
getTransformingMap(Map<K1, V1>, Function<K1, K2>, Function<V1, V2>) - Static method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils
 
getTransformNameMapping() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
Mapping of old PTranform names to new ones, specified as JSON {"oldName":"newName",...} .
getTransformPayloadTranslators() - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.Registrar
 
getTransformTranslator(Class<TransformT>) - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator
Returns the TransformTranslator to use for instances of the specified PTransform class, or null if none registered.
getTransformTranslator(TransformHierarchy.Node) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.PipelineTranslatorBatch
Returns a translator for the given node, if it is possible, otherwise null.
getTransformTranslator(TransformHierarchy.Node) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
Get a TransformTranslator for the given TransformHierarchy.Node.
getTransformTranslator(TransformHierarchy.Node) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.streaming.PipelineTranslatorStreaming
Returns a translator for the given node, if it is possible, otherwise null.
getTranslationContext() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
 
getTranslator() - Method in class org.apache.beam.runners.dataflow.DataflowRunner
Returns the DataflowPipelineTranslator associated with this object.
getTransport() - Static method in class org.apache.beam.sdk.extensions.gcp.util.Transport
 
getTrigger() - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
getTruncateTimestamps() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogPipelineOptions
Whether to truncate timestamps in tables described by Data Catalog.
getTruncateTimestamps() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.ConversionOptions
 
getTupleTag() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple.TaggedKeyedPCollection
Returns the TupleTag of this TaggedKeyedPCollection.
getTupleTagId(PValue) - Static method in class org.apache.beam.runners.jet.Utils
 
getTupleTagList() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
Returns the TupleTagList tuple associated with this schema.
getType() - Method in class org.apache.beam.sdk.coders.AvroCoder
Returns the type this coder encodes/decodes.
getType(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
getType() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table
type of the table.
getType() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
 
getType() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
Returns the field type.
getType() - Method in class org.apache.beam.sdk.schemas.Schema.Field
Returns the fields Schema.FieldType.
getType(String) - Method in class org.apache.beam.sdk.schemas.Schema.Options
Get the type of an option.
getType() - Method in interface org.apache.beam.sdk.testing.TestStream.Event
 
getType() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
Retrieve the DisplayData.Type of display data.
getType() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
The DisplayData.Type of display data.
getType() - Method in class org.apache.beam.sdk.values.TypeDescriptor
Returns the Type represented by this TypeDescriptor.
getTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.ViewFn
Return the TypeDescriptor describing the output of this fn.
getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollection
Returns a TypeDescriptor<T> with some reflective information about T, if possible.
getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn
 
getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.ListViewFn
 
getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.MapViewFn
 
getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.MultimapViewFn
 
getTypeDescriptor() - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn
 
getTypeDescriptor() - Method in class org.apache.beam.sdk.values.TupleTag
Returns a TypeDescriptor capturing what is known statically about the type of this TupleTag instance's most-derived class.
getTypeFactory() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
getTypeFactory() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl
 
getTypeMap() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
getTypeName() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
 
getTypeNames() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
getTypeParameter(String) - Method in class org.apache.beam.sdk.values.TypeDescriptor
Returns a TypeVariable for the named type parameter.
getTypes() - Method in class org.apache.beam.sdk.values.TypeDescriptor
Returns a set of TypeDescriptor, one for each superclass as well as each interface implemented by this class.
getUdafs() - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.UdfUdafProvider
 
getUnderlyingDoFn() - Method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides.BatchStatefulDoFn
 
getUnionCoder() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
 
getUnionTag() - Method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils.RawUnionValue
 
getUnionTag() - Method in class org.apache.beam.sdk.transforms.join.RawUnionValue
 
getUniqueId() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
getUntilTrigger() - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
The trigger that signals termination of this trigger.
getUrn(PrimitiveParDoSingleFactory.ParDoSingle<?, ?>) - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.PayloadTranslator
 
getUrn() - Method in interface org.apache.beam.sdk.transforms.Materialization
Gets the URN describing this Materialization.
getUsePublicIps() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
Specifies whether worker pools should be started with public IP addresses.
getUserAgent() - Method in interface org.apache.beam.sdk.options.PipelineOptions
A user agent string as per RFC2616, describing the pipeline to external services.
getUsername() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
 
getUsesProvidedSparkContext() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
getUsingStandardSql() - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher.TableAndQuery
 
getUUID() - Method in class org.apache.beam.sdk.schemas.Schema
Get this schema's UUID.
getV1Beta1Path() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.SubscriptionPath
 
getV1Beta1Path() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.TopicPath
 
getValue() - Method in class org.apache.beam.runners.flink.metrics.FlinkMetricContainer.FlinkDistributionGauge
 
getValue() - Method in class org.apache.beam.runners.flink.metrics.FlinkMetricContainer.FlinkGauge
 
getValue() - Method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils.RawUnionValue
 
getValue(String, Class<T>) - Method in class org.apache.beam.runners.spark.aggregators.NamedAggregators
 
getValue(String, Class<T>) - Method in class org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregators
 
getValue() - Method in class org.apache.beam.runners.spark.util.ByteArray
 
getValue() - Method in class org.apache.beam.runners.spark.util.SideInputBroadcast
 
getValue() - Method in class org.apache.beam.sdk.coders.StructuralByteArray
 
getValue() - Method in class org.apache.beam.sdk.io.range.ByteKey
Returns a read-only ByteBuffer representing this ByteKey.
getValue() - Method in class org.apache.beam.sdk.metrics.GaugeResult.EmptyGaugeResult
 
getValue() - Method in class org.apache.beam.sdk.metrics.GaugeResult
 
getValue() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType.Value
Return the integer enum value.
getValue(Class<T>) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType.Value
Returns the current value of the OneOf as the destination type.
getValue() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType.Value
Returns the current value of the OneOf.
getValue(String) - Method in class org.apache.beam.sdk.schemas.Schema.Options
Get the value of an option.
getValue(String, Class<T>) - Method in class org.apache.beam.sdk.schemas.Schema.Options
Get the value of an option.
getValue() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
Retrieve the value of the display item.
getValue() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
The value of the display item.
getValue() - Method in class org.apache.beam.sdk.transforms.join.RawUnionValue
 
getValue() - Method in class org.apache.beam.sdk.values.KV
Returns the value of this KV.
getValue(int) - Method in class org.apache.beam.sdk.values.Row
Get value by field index, ClassCastException is thrown if schema doesn't match.
getValue(String) - Method in class org.apache.beam.sdk.values.Row
Get value by field name, ClassCastException is thrown if type doesn't match.
getValue(int) - Method in class org.apache.beam.sdk.values.RowWithGetters
 
getValue(int) - Method in class org.apache.beam.sdk.values.RowWithStorage
 
getValue() - Method in class org.apache.beam.sdk.values.TaggedPValue
Returns the PValue.
getValue() - Method in class org.apache.beam.sdk.values.TimestampedValue
 
getValue() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow
Returns the value of this ValueInSingleWindow.
getValue() - Method in class org.apache.beam.sdk.values.ValueWithRecordId
 
getValueCoder() - Method in class org.apache.beam.sdk.coders.KvCoder
 
getValueCoder() - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
Gets the value coder that will be prefixed by the length.
getValueCoder() - Method in class org.apache.beam.sdk.coders.MapCoder
 
getValueCoder() - Method in class org.apache.beam.sdk.coders.NullableCoder
Returns the inner Coder wrapped by this NullableCoder instance.
getValueCoder() - Method in class org.apache.beam.sdk.testing.TestStream
 
getValueCoder() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
 
getValueCoder() - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
 
getValueOrDefault(String, T) - Method in class org.apache.beam.sdk.schemas.Schema.Options
Get the value of an option.
getValues() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
 
getValues() - Method in class org.apache.beam.sdk.values.Row
Return the list of data values.
getValues() - Method in class org.apache.beam.sdk.values.RowWithGetters
 
getValues() - Method in class org.apache.beam.sdk.values.RowWithStorage
 
getValuesMap() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
 
getValueTranslationFunction() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
 
getValueTypeDescriptor() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
 
getVerifyRowValues() - Method in interface org.apache.beam.sdk.extensions.sql.impl.BeamSqlPipelineOptions
 
getView() - Method in class org.apache.beam.runners.apex.ApexRunner.CreateApexPCollectionView
 
getView() - Method in class org.apache.beam.runners.dataflow.CreateDataflowView
 
getView() - Method in class org.apache.beam.runners.dataflow.DataflowRunner.StreamingPCollectionViewWriterFn
 
getView() - Method in class org.apache.beam.sdk.transforms.View.CreatePCollectionView
Deprecated.
This should not be used to obtain the output of any given application of this PTransform. That should be obtained by inspecting the TransformHierarchy.Node that contains this View.CreatePCollectionView, as this view may have been replaced within pipeline surgery.
getViewFn() - Method in interface org.apache.beam.sdk.values.PCollectionView
Deprecated.
this method will be removed entirely. The ViewFn for a side input is an attribute of the side input's specification with a ParDo transform, which will obtain this specification via a package-private channel.
getViewFn() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
 
getVocabulary() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
getVocabulary() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
getWarnings() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
getWatermark() - Method in class org.apache.beam.runners.gearpump.translators.io.GearpumpSource
 
getWatermark() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
 
getWatermark(TimestampPolicy.PartitionContext) - Method in class org.apache.beam.sdk.io.kafka.CustomTimestampPolicyWithLimitedDelay
 
getWatermark(TimestampPolicy.PartitionContext) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicy
Returns watermark for the partition.
getWatermark(TimestampPolicy.PartitionContext) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.LogAppendTimePolicy
 
getWatermark(TimestampPolicy.PartitionContext) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.ProcessingTimePolicy
 
getWatermark(TimestampPolicy.PartitionContext) - Method in class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.TimestampFnPolicy
 
getWatermark() - Method in interface org.apache.beam.sdk.io.kinesis.WatermarkPolicy
 
getWatermark() - Method in class org.apache.beam.sdk.io.kinesis.WatermarkPolicyFactory.ArrivalTimeWatermarkPolicy
 
getWatermark() - Method in class org.apache.beam.sdk.io.kinesis.WatermarkPolicyFactory.CustomWatermarkPolicy
 
getWatermark() - Method in class org.apache.beam.sdk.io.kinesis.WatermarkPolicyFactory.ProcessingTimeWatermarkPolicy
 
getWatermark() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
Returns a timestamp before or at the timestamps of all future elements read by this reader.
getWatermark() - Method in class org.apache.beam.sdk.testing.TestStream.WatermarkEvent
 
getWatermarkAndState() - Method in interface org.apache.beam.sdk.fn.splittabledofn.WatermarkEstimators.WatermarkAndStateObserver
 
getWatermarkMillis() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
 
getWatermarks(ExecutableT) - Method in class org.apache.beam.runners.direct.WatermarkManager
Gets the input and output watermarks for an AppliedPTransform.
getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterAll
 
getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterEach
 
getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterFirst
 
getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
 
getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
 
getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterSynchronizedProcessingTime
 
getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
 
getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
 
getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.DefaultTrigger
 
getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.Never.NeverTrigger
 
getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
 
getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
 
getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.ReshuffleTrigger
Deprecated.
 
getWatermarkThatGuaranteesFiring(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
For internal use only; no backwards-compatibility guarantees.
getWindmillServiceEndpoint() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
Custom windmill service endpoint.
getWindmillServicePort() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
 
getWindow() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
This method returns the number of tuples in each window.
getWindow() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
 
getWindow() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow
Returns the window of this ValueInSingleWindow.
getWindowedWrites() - Method in class org.apache.beam.sdk.io.WriteFiles
 
getWindowFn() - Method in class org.apache.beam.sdk.transforms.windowing.Window.Assign
 
getWindowFn() - Method in class org.apache.beam.sdk.transforms.windowing.Window
 
getWindowFn() - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
getWindowingStrategy(String, RunnerApi.Components) - Static method in class org.apache.beam.runners.fnexecution.translation.PipelineTranslatorUtils
 
getWindowingStrategy() - Method in class org.apache.beam.sdk.values.PCollection
Returns the WindowingStrategy of this PCollection.
getWindowingStrategyInternal() - Method in interface org.apache.beam.sdk.values.PCollectionView
Deprecated.
this method will be removed entirely. The PCollection underlying a side input, including its WindowingStrategy, is part of the side input's specification with a ParDo transform, which will obtain that information via a package-private channel.
getWindowingStrategyInternal() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
Returns the WindowingStrategy of this PCollectionView, which should be that of the underlying PCollection.
getWindowMappingFn() - Method in interface org.apache.beam.sdk.values.PCollectionView
For internal use only.
getWindowMappingFn() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
 
getWindowTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
 
getWindowTypeDescriptor() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
Returns a TypeDescriptor capturing what is known statically about the window type of this WindowFn instance's most-derived class.
getWorkerCacheMb() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
The size of the worker's in-memory cache, in megabytes.
getWorkerDiskType() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
Specifies what type of persistent disk is used.
getWorkerHarnessContainerImage() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
Docker container image that executes Dataflow worker harness, residing in Google Container Registry.
getWorkerId() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
The identity of the worker running this pipeline.
getWorkerId() - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClient
 
getWorkerLogLevelOverrides() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
Deprecated.
This option controls the log levels for specifically named loggers.
getWorkerMachineType() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
Machine type to create Dataflow worker VMs as.
getWorkerRegion() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
The Compute Engine region (https://cloud.google.com/compute/docs/regions-zones/regions-zones) in which worker processing should occur, e.g.
getWorkerSystemErrMessageLevel() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
Deprecated.
Controls the log level given to messages printed to System.err.
getWorkerSystemOutMessageLevel() - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
Deprecated.
Controls the log level given to messages printed to System.out.
getWorkerZone() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
The Compute Engine zone (https://cloud.google.com/compute/docs/regions-zones/regions-zones) in which worker processing should occur, e.g.
getWritableByteChannelFactory() - Method in class org.apache.beam.sdk.io.FileBasedSink
getWriteOperation() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
Return the WriteOperation that this Writer belongs to.
getYarnDeployDependencies() - Static method in class org.apache.beam.runners.apex.ApexYarnLauncher
From the current classpath, find the jar files that need to be deployed with the application to run on YARN.
getZetaSqlRuleSets() - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
 
getZone() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
Deprecated.
global(Map<Integer, GlobalWatermarkHolder.SparkWatermarks>) - Static method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
Build a global TimerInternals for all feeding streams.
Global() - Constructor for class org.apache.beam.sdk.schemas.transforms.Group.Global
 
globalDefault() - Static method in class org.apache.beam.sdk.values.WindowingStrategy
Return a fully specified, default windowing strategy.
GlobalDigest() - Constructor for class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.GlobalDigest
 
globally() - Static method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct
Computes the approximate number of distinct elements in the input PCollection<InputT> and returns a PCollection<Long>.
globally() - Static method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies
Create the PTransform that will build a Count-min sketch for keeping track of the frequency of the elements in the whole stream.
globally() - Static method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles
Compute the stream in order to build a T-Digest structure (MergingDigest) for keeping track of the stream distribution and returns a PCollection<MergingDigest>.
globally() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Extract
Returns a PTransform that takes an input PCollection<byte[]> of HLL++ sketches and returns a PCollection<Long> of the estimated count of distinct elements extracted from each sketch.
globally() - Method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Init.Builder
Returns a Combine.Globally PTransform that takes an input PCollection<InputT> and returns a PCollection<byte[]> which consists of the HLL++ sketch computed from the elements in the input PCollection.
globally() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.MergePartial
Returns a Combine.Globally PTransform that takes an input PCollection<byte[]> of HLL++ sketches and returns a PCollection<byte[]> of a new sketch merged from the input sketches.
globally() - Static method in class org.apache.beam.sdk.schemas.transforms.Group
Returns a transform that groups all elements in the input PCollection.
globally(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles
Returns a PTransform that takes a PCollection<T> and returns a PCollection<List<T>> whose single value is a List of the approximate N-tiles of the elements of the input PCollection.
globally(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles
Like ApproximateQuantiles.globally(int, Comparator), but sorts using the elements' natural ordering.
globally(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateUnique
Returns a PTransform that takes a PCollection<T> and returns a PCollection<Long> containing a single value that is an estimate of the number of distinct elements in the input PCollection.
globally(double) - Static method in class org.apache.beam.sdk.transforms.ApproximateUnique
Like ApproximateUnique.globally(int), but specifies the desired maximum estimation error instead of the sample size.
Globally(int) - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique.Globally
 
Globally(double) - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique.Globally
 
globally(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
Returns a Combine.Globally PTransform that uses the given SerializableFunction to combine all the elements in each window of the input PCollection into a single value in the output PCollection.
globally(SerializableBiFunction<V, V, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
Returns a Combine.Globally PTransform that uses the given SerializableBiFunction to combine all the elements in each window of the input PCollection into a single value in the output PCollection.
globally(CombineFnBase.GlobalCombineFn<? super InputT, ?, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Combine
Returns a Combine.Globally PTransform that uses the given GloballyCombineFn to combine all the elements in each window of the input PCollection into a single value in the output PCollection.
globally() - Static method in class org.apache.beam.sdk.transforms.Count
Returns a PTransform that counts the number of elements in its input PCollection.
globally() - Static method in class org.apache.beam.sdk.transforms.Latest
Returns a PTransform that takes as input a PCollection<T> and returns a PCollection<T> whose contents is the latest element according to its event time, or null if there are no elements.
globally() - Static method in class org.apache.beam.sdk.transforms.Max
Returns a PTransform that takes an input PCollection<T> and returns a PCollection<T> whose contents is the maximum according to the natural ordering of T of the input PCollection's elements, or null if there are no elements.
globally(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Max
Returns a PTransform that takes an input PCollection<T> and returns a PCollection<T> whose contents is the maximum of the input PCollection's elements, or null if there are no elements.
globally() - Static method in class org.apache.beam.sdk.transforms.Mean
Returns a PTransform that takes an input PCollection<NumT> and returns a PCollection<Double> whose contents is the mean of the input PCollection's elements, or 0 if there are no elements.
globally() - Static method in class org.apache.beam.sdk.transforms.Min
Returns a PTransform that takes an input PCollection<T> and returns a PCollection<T> whose contents is the minimum according to the natural ordering of T of the input PCollection's elements, or null if there are no elements.
globally(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Min
Returns a PTransform that takes an input PCollection<T> and returns a PCollection<T> whose contents is the minimum of the input PCollection's elements, or null if there are no elements.
GloballyDistinct() - Constructor for class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.GloballyDistinct
 
GlobalSketch() - Constructor for class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.GlobalSketch
 
GlobalWatermarkHolder - Class in org.apache.beam.runners.spark.util
A store to hold the global watermarks for a micro-batch.
GlobalWatermarkHolder() - Constructor for class org.apache.beam.runners.spark.util.GlobalWatermarkHolder
 
GlobalWatermarkHolder.SparkWatermarks - Class in org.apache.beam.runners.spark.util
A GlobalWatermarkHolder.SparkWatermarks holds the watermarks and batch time relevant to a micro-batch input from a specific source.
GlobalWatermarkHolder.WatermarkAdvancingStreamingListener - Class in org.apache.beam.runners.spark.util
Advance the WMs onBatchCompleted event.
GlobalWindow - Class in org.apache.beam.sdk.transforms.windowing
The default window into which all data is placed (via GlobalWindows).
GlobalWindow.Coder - Class in org.apache.beam.sdk.transforms.windowing
GlobalWindow.Coder for encoding and decoding GlobalWindows.
GlobalWindows - Class in org.apache.beam.sdk.transforms.windowing
A WindowFn that assigns all data to the same window.
GlobalWindows() - Constructor for class org.apache.beam.sdk.transforms.windowing.GlobalWindows
 
GoogleApiDebugOptions - Interface in org.apache.beam.sdk.extensions.gcp.options
These options configure debug settings for Google API clients created within the Apache Beam SDK.
GoogleApiDebugOptions.GoogleApiTracer - Class in org.apache.beam.sdk.extensions.gcp.options
A GoogleClientRequestInitializer that adds the trace destination to Google API calls.
GoogleApiTracer() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions.GoogleApiTracer
 
GraphiteSink - Class in org.apache.beam.runners.spark.metrics.sink
A Spark Sink that is tailored to report AggregatorMetric metrics to Graphite.
GraphiteSink(Properties, MetricRegistry, SecurityManager) - Constructor for class org.apache.beam.runners.spark.metrics.sink.GraphiteSink
 
greaterThan(T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
greaterThan(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
greaterThan(T) - Static method in class org.apache.beam.sdk.transforms.Filter
Returns a PTransform that takes an input PCollection<T> and returns a PCollection<T> with elements that are greater than a given value, based on the elements' natural ordering.
greaterThanEq(T) - Static method in class org.apache.beam.sdk.transforms.Filter
Returns a PTransform that takes an input PCollection<T> and returns a PCollection<T> with elements that are greater than or equal to a given value, based on the elements' natural ordering.
greaterThanOrEqualTo(T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
greaterThanOrEqualTo(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
Group - Class in org.apache.beam.sdk.schemas.transforms
A generic grouping transform for schema PCollections.
Group() - Constructor for class org.apache.beam.sdk.schemas.transforms.Group
 
Group.ByFields<InputT> - Class in org.apache.beam.sdk.schemas.transforms
a PTransform that groups schema elements based on the given fields.
Group.CombineFieldsByFields<InputT> - Class in org.apache.beam.sdk.schemas.transforms
a PTransform that does a per-key combine using an aggregation built up by calls to aggregateField and aggregateFields.
Group.CombineFieldsGlobally<InputT> - Class in org.apache.beam.sdk.schemas.transforms
a PTransform that does a global combine using an aggregation built up by calls to aggregateField and aggregateFields.
Group.CombineGlobally<InputT,OutputT> - Class in org.apache.beam.sdk.schemas.transforms
a PTransform that does a global combine using a provider Combine.CombineFn.
Group.Global<InputT> - Class in org.apache.beam.sdk.schemas.transforms
A PTransform for doing global aggregations on schema PCollections.
GroupAlsoByWindowViaOutputBufferFn<K,InputT,W extends BoundedWindow> - Class in org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions
A FlatMap function that groups by windows in batch mode using ReduceFnRunner.
GroupAlsoByWindowViaOutputBufferFn(WindowingStrategy<?, W>, StateInternalsFactory<K>, SystemReduceFn<K, InputT, Iterable<InputT>, Iterable<InputT>, W>, SerializablePipelineOptions) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.GroupAlsoByWindowViaOutputBufferFn
 
groupBy(WindowedValue<KV<K, V>>) - Method in class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator.GroupByFn
 
GroupByKey<K,V> - Class in org.apache.beam.sdk.transforms
GroupByKey<K, V> takes a PCollection<KV<K, V>>, groups the values by key and windows, and returns a PCollection<KV<K, Iterable<V>>> representing a map from each distinct key and window of the input PCollection to an Iterable over all the values associated with that key in the input per window.
groupByKeyAndWindow(JavaDStream<WindowedValue<KV<K, InputT>>>, Coder<K>, Coder<WindowedValue<InputT>>, WindowingStrategy<?, W>, SerializablePipelineOptions, List<Integer>, String) - Static method in class org.apache.beam.runners.spark.stateful.SparkGroupAlsoByWindowViaWindowSet
 
GroupByKeyTranslator<K,V> - Class in org.apache.beam.runners.gearpump.translators
GroupByKey is translated to Gearpump groupBy function.
GroupByKeyTranslator() - Constructor for class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator
 
GroupByKeyTranslator.GearpumpWindowFn<T,W extends BoundedWindow> - Class in org.apache.beam.runners.gearpump.translators
A transform used internally to translate Beam's Window to Gearpump's Window.
GroupByKeyTranslator.GroupByFn<K,V> - Class in org.apache.beam.runners.gearpump.translators
A transform used internally to group KV message by its key.
GroupByKeyTranslator.KeyedByTimestamp<K,V> - Class in org.apache.beam.runners.gearpump.translators
A transform used internally to transform WindowedValue to KV.
GroupByKeyTranslator.Merge<K,V> - Class in org.apache.beam.runners.gearpump.translators
A transform used internally by Gearpump which encapsulates the merge logic.
grouped() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Same transform but can be applied to PCollection of MutationGroup.
groupedValues(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
Returns a Combine.GroupedValues PTransform that takes a PCollection of KVs where a key maps to an Iterable of values, e.g., the result of a GroupByKey, then uses the given SerializableFunction to combine all the values associated with a key, ignoring the key.
groupedValues(SerializableBiFunction<V, V, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
Returns a Combine.GroupedValues PTransform that takes a PCollection of KVs where a key maps to an Iterable of values, e.g., the result of a GroupByKey, then uses the given SerializableFunction to combine all the values associated with a key, ignoring the key.
groupedValues(CombineFnBase.GlobalCombineFn<? super InputT, ?, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Combine
Returns a Combine.GroupedValues PTransform that takes a PCollection of KVs where a key maps to an Iterable of values, e.g., the result of a GroupByKey, then uses the given CombineFn to combine all the values associated with a key, ignoring the key.
GroupingState<InputT,OutputT> - Interface in org.apache.beam.sdk.state
A ReadableState cell that combines multiple input values and outputs a single value of a different type.
GroupIntoBatches<K,InputT> - Class in org.apache.beam.sdk.transforms
A PTransform that batches inputs to a desired batch size.
Growth() - Constructor for class org.apache.beam.sdk.transforms.Watch.Growth
 
growthOf(Watch.Growth.PollFn<InputT, OutputT>, Requirements) - Static method in class org.apache.beam.sdk.transforms.Watch
Watches the growth of the given poll function.
growthOf(Watch.Growth.PollFn<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Watch
Watches the growth of the given poll function.
growthOf(Contextful<Watch.Growth.PollFn<InputT, OutputT>>, SerializableFunction<OutputT, KeyT>) - Static method in class org.apache.beam.sdk.transforms.Watch
Watches the growth of the given poll function, using the given "key function" to deduplicate outputs.
GrpcContextHeaderAccessorProvider - Class in org.apache.beam.runners.fnexecution
A HeaderAccessorProvider which intercept the header in a GRPC request and expose the relevant fields.
GrpcContextHeaderAccessorProvider() - Constructor for class org.apache.beam.runners.fnexecution.GrpcContextHeaderAccessorProvider
 
GrpcDataService - Class in org.apache.beam.runners.fnexecution.data
A FnDataService implemented via gRPC.
GrpcDataService() - Constructor for class org.apache.beam.runners.fnexecution.data.GrpcDataService
Deprecated.
This constructor is for migrating Dataflow purpose only.
GrpcFnServer<ServiceT extends FnService> - Class in org.apache.beam.runners.fnexecution
A gRPC Server which manages a single FnService.
GrpcLoggingService - Class in org.apache.beam.runners.fnexecution.logging
An implementation of the Beam Fn Logging Service over gRPC.
GrpcStateService - Class in org.apache.beam.runners.fnexecution.state
An implementation of the Beam Fn State service.

H

HadoopFileSystemModule - Class in org.apache.beam.sdk.io.hdfs
A Jackson Module that registers a JsonSerializer and JsonDeserializer for a Hadoop Configuration.
HadoopFileSystemModule() - Constructor for class org.apache.beam.sdk.io.hdfs.HadoopFileSystemModule
 
HadoopFileSystemOptions - Interface in org.apache.beam.sdk.io.hdfs
PipelineOptions which encapsulate Hadoop Configuration for the HadoopFileSystem.
HadoopFileSystemOptions.ConfigurationLocator - Class in org.apache.beam.sdk.io.hdfs
A DefaultValueFactory which locates a Hadoop Configuration.
HadoopFileSystemOptionsRegistrar - Class in org.apache.beam.sdk.io.hdfs
AutoService registrar for HadoopFileSystemOptions.
HadoopFileSystemOptionsRegistrar() - Constructor for class org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptionsRegistrar
 
HadoopFileSystemRegistrar - Class in org.apache.beam.sdk.io.hdfs
AutoService registrar for the HadoopFileSystem.
HadoopFileSystemRegistrar() - Constructor for class org.apache.beam.sdk.io.hdfs.HadoopFileSystemRegistrar
 
HadoopFormatIO - Class in org.apache.beam.sdk.io.hadoop.format
A HadoopFormatIO is a Transform for reading data from any source or writing data to any sink which implements Hadoop InputFormat or OutputFormat.
HadoopFormatIO() - Constructor for class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
 
HadoopFormatIO.HadoopInputFormatBoundedSource<K,V> - Class in org.apache.beam.sdk.io.hadoop.format
Bounded source implementation for HadoopFormatIO.
HadoopFormatIO.Read<K,V> - Class in org.apache.beam.sdk.io.hadoop.format
A PTransform that reads from any data source which implements Hadoop InputFormat.
HadoopFormatIO.SerializableSplit - Class in org.apache.beam.sdk.io.hadoop.format
A wrapper to allow Hadoop InputSplit to be serialized using Java's standard serialization mechanisms.
HadoopFormatIO.Write<KeyT,ValueT> - Class in org.apache.beam.sdk.io.hadoop.format
A PTransform that writes to any data sink which implements Hadoop OutputFormat.
HadoopFormatIO.Write.ExternalSynchronizationBuilder<KeyT,ValueT> - Interface in org.apache.beam.sdk.io.hadoop.format
Builder for External Synchronization defining.
HadoopFormatIO.Write.PartitionedWriterBuilder<KeyT,ValueT> - Interface in org.apache.beam.sdk.io.hadoop.format
Builder for partitioning determining.
HadoopFormatIO.Write.WriteBuilder<KeyT,ValueT> - Interface in org.apache.beam.sdk.io.hadoop.format
Main builder of Write transformation.
HadoopInputFormatBoundedSource(SerializableConfiguration, Coder<K>, Coder<V>, SimpleFunction<?, K>, SimpleFunction<?, V>, HadoopFormatIO.SerializableSplit) - Constructor for class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
 
handle(BeamFnApi.InstructionRequest) - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClient
 
handle(BeamFnApi.InstructionRequest) - Method in interface org.apache.beam.runners.fnexecution.control.InstructionRequestHandler
 
handle(BeamFnApi.StateRequest) - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandler
Handle a BeamFnApi.StateRequest asynchronously.
has(TupleTag<T>) - Method in class org.apache.beam.sdk.values.PCollectionTuple
Returns whether this PCollectionTuple contains a PCollection with the given tag.
has(String) - Method in class org.apache.beam.sdk.values.PCollectionTuple
Returns whether this PCollectionTuple contains a PCollection with the given tag.
hasCommitted() - Method in class org.apache.beam.sdk.metrics.MetricResult
 
hasDatabase(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogBeamSchema
Checks if metastore client has the specified database.
hasDefault() - Method in class org.apache.beam.sdk.values.PCollectionViews.SingletonViewFn
Returns if a default value was specified.
HasDefaultTracker<RestrictionT extends HasDefaultTracker<RestrictionT,TrackerT>,TrackerT extends RestrictionTracker<RestrictionT,?>> - Interface in org.apache.beam.sdk.transforms.splittabledofn
Interface for restrictions for which a default implementation of DoFn.NewTracker is available, depending only on the restriction itself.
hasDefaultValue() - Method in class org.apache.beam.sdk.transforms.View.AsSingleton
Returns whether this transform has a default value.
HasDefaultWatermarkEstimator<WatermarkEstimatorStateT,WatermarkEstimatorT extends WatermarkEstimator<WatermarkEstimatorStateT>> - Interface in org.apache.beam.sdk.transforms.splittabledofn
Interface for watermark estimator state for which a default implementation of DoFn.NewWatermarkEstimator is available, depending only on the watermark estimator state itself.
HasDisplayData - Interface in org.apache.beam.sdk.transforms.display
Marker interface for PTransforms and components to specify display data used within UIs and diagnostic tools.
hasExperiment(DataflowPipelineDebugOptions, String) - Static method in class org.apache.beam.runners.dataflow.DataflowRunner
Returns true if the specified experiment is enabled, handling null experiments.
hasExperiment(PipelineOptions, String) - Static method in interface org.apache.beam.sdk.options.ExperimentalOptions
Returns true iff the provided pipeline options has the specified experiment enabled.
hasField(String) - Method in class org.apache.beam.sdk.schemas.Schema
Returns true if fieldName exists in the schema, false otherwise.
hasGlobWildcard(String) - Static method in class org.apache.beam.sdk.io.FileSystems
Checks whether the given spec contains a glob wildcard character.
hashCode() - Method in class org.apache.beam.runners.dataflow.util.CloudObject
 
hashCode() - Method in class org.apache.beam.runners.dataflow.util.OutputReference
 
hashCode() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
 
hashCode() - Method in class org.apache.beam.runners.direct.WatermarkManager.TimerUpdate
 
hashCode() - Method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils.RawUnionValue
 
hashCode() - Method in class org.apache.beam.runners.jet.Utils.ByteArrayKey
 
hashCode() - Method in class org.apache.beam.runners.spark.io.EmptyCheckpointMark
 
hashCode() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
 
hashCode() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.DecodeUsingBeamCoder
 
hashCode() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.EncodeUsingBeamCoder
 
hashCode() - Method in class org.apache.beam.runners.spark.util.ByteArray
 
hashCode() - Method in class org.apache.beam.sdk.coders.AtomicCoder
.
hashCode() - Method in class org.apache.beam.sdk.coders.AvroCoder
 
hashCode() - Method in class org.apache.beam.sdk.coders.Coder.Context
Deprecated.
 
hashCode() - Method in class org.apache.beam.sdk.coders.DelegateCoder
 
hashCode() - Method in class org.apache.beam.sdk.coders.RowCoder
 
hashCode() - Method in class org.apache.beam.sdk.coders.SerializableCoder
 
hashCode() - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
 
hashCode() - Method in class org.apache.beam.sdk.coders.StructuralByteArray
 
hashCode() - Method in class org.apache.beam.sdk.coders.StructuredCoder
 
hashCode() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
 
hashCode() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
hashCode() - Method in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
 
hashCode() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
 
hashCode() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
 
hashCode() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
 
hashCode() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Order
 
hashCode() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
hashCode() - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
 
hashCode() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
 
hashCode() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertError
 
hashCode() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
hashCode() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.ProjectPath
 
hashCode() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.SubscriptionPath
 
hashCode() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.TopicPath
 
hashCode() - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
 
hashCode() - Method in class org.apache.beam.sdk.io.hadoop.WritableCoder
 
hashCode() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
 
hashCode() - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
 
hashCode() - Method in class org.apache.beam.sdk.io.jms.JmsRecord
 
hashCode() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecord
 
hashCode() - Method in class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
hashCode() - Method in class org.apache.beam.sdk.io.range.ByteKey
 
hashCode() - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
 
hashCode() - Method in class org.apache.beam.sdk.io.range.OffsetRange
 
hashCode() - Method in class org.apache.beam.sdk.io.tika.ParseResult
 
hashCode() - Method in class org.apache.beam.sdk.io.xml.JAXBCoder
 
hashCode() - Method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
 
hashCode() - Method in class org.apache.beam.sdk.options.ValueProvider.RuntimeValueProvider
 
hashCode() - Method in class org.apache.beam.sdk.options.ValueProvider.StaticValueProvider
 
hashCode() - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
 
hashCode() - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema.GetterTypeSupplier
 
hashCode() - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema
 
hashCode() - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema.SetterTypeSupplier
 
hashCode() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType.Value
 
hashCode() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType.Value
 
hashCode() - Method in class org.apache.beam.sdk.schemas.Schema.Field
 
hashCode() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
 
hashCode() - Method in class org.apache.beam.sdk.schemas.Schema
 
hashCode() - Method in class org.apache.beam.sdk.schemas.Schema.Options
 
hashCode() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
 
hashCode() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
hashCode() - Method in class org.apache.beam.sdk.testing.PAssert.PAssertionSite
 
hashCode() - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
Deprecated.
Object.hashCode() is not supported on PAssert objects.
hashCode() - Method in class org.apache.beam.sdk.testing.SuccessOrFailure
 
hashCode() - Method in class org.apache.beam.sdk.testing.TestStream
 
hashCode() - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn.LargestUnique
 
hashCode() - Method in class org.apache.beam.sdk.transforms.CombineFns.CoCombineResult
 
hashCode() - Method in class org.apache.beam.sdk.transforms.display.DisplayData
 
hashCode() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Path
 
hashCode() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
 
hashCode() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
 
hashCode() - Method in class org.apache.beam.sdk.transforms.join.RawUnionValue
 
hashCode() - Method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
 
hashCode() - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
 
hashCode() - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
 
hashCode() - Method in class org.apache.beam.sdk.transforms.windowing.AfterSynchronizedProcessingTime
 
hashCode() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
 
hashCode() - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
 
hashCode() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow
 
hashCode() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
 
hashCode() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
 
hashCode() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
 
hashCode() - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
 
hashCode() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
 
hashCode() - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
 
hashCode() - Method in class org.apache.beam.sdk.values.KV
 
hashCode() - Method in class org.apache.beam.sdk.values.PCollectionList
 
hashCode() - Method in class org.apache.beam.sdk.values.PCollectionTuple
 
hashCode() - Method in class org.apache.beam.sdk.values.PCollectionViews.ListViewFn
 
hashCode() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
 
hashCode() - Method in class org.apache.beam.sdk.values.Row
 
hashCode() - Method in class org.apache.beam.sdk.values.RowWithGetters
 
hashCode() - Method in class org.apache.beam.sdk.values.ShardedKey
 
hashCode() - Method in class org.apache.beam.sdk.values.TimestampedValue
 
hashCode() - Method in class org.apache.beam.sdk.values.TupleTag
 
hashCode() - Method in class org.apache.beam.sdk.values.TypeDescriptor
 
hashCode() - Method in class org.apache.beam.sdk.values.TypeParameter
 
hashCode() - Method in class org.apache.beam.sdk.values.ValueWithRecordId
 
hashCode() - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
hasItem(T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.hasItem(Object).
hasItem(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.hasItem(Object).
hasItem(SerializableMatcher<? super T>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.hasItem(Matcher).
hasNext() - Method in class org.apache.beam.sdk.fn.stream.DataStreams.BlockingQueueIterator
 
hasNext() - Method in class org.apache.beam.sdk.fn.stream.DataStreams.DataStreamDecoder
 
hasOption(String) - Method in class org.apache.beam.sdk.schemas.Schema.Options
 
hasOptions() - Method in class org.apache.beam.sdk.schemas.Schema.Options
 
hasReplacementJob() - Method in enum org.apache.beam.sdk.PipelineResult.State
 
hasSchema() - Method in class org.apache.beam.sdk.values.PCollection
Returns whether this PCollection has an attached schema.
hasSize(int) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.hasSize(int).
hasSize(SerializableMatcher<? super Integer>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.hasSize(Matcher).
hasUnboundedPCollections(RunnerApi.Pipeline) - Static method in class org.apache.beam.runners.fnexecution.translation.PipelineTranslatorUtils
Indicates whether the given pipeline has any unbounded PCollections.
hasUnresolvedParameters() - Method in class org.apache.beam.sdk.values.TypeDescriptor
Returns whether this TypeDescriptor has any unresolved type parameters, as opposed to being a concrete type.
HBaseCoderProviderRegistrar - Class in org.apache.beam.sdk.io.hbase
A CoderProviderRegistrar for standard types used with HBaseIO.
HBaseCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.hbase.HBaseCoderProviderRegistrar
 
HBaseIO - Class in org.apache.beam.sdk.io.hbase
A bounded source and sink for HBase.
HBaseIO.Read - Class in org.apache.beam.sdk.io.hbase
A PTransform that reads from HBase.
HBaseIO.ReadAll - Class in org.apache.beam.sdk.io.hbase
Implementation of HBaseIO.readAll().
HBaseIO.Write - Class in org.apache.beam.sdk.io.hbase
A PTransform that writes to HBase.
HCatalogBeamSchema - Class in org.apache.beam.sdk.io.hcatalog
Adapter from HCatalog table schema to Beam Schema.
HCatalogIO - Class in org.apache.beam.sdk.io.hcatalog
IO to read and write data using HCatalog.
HCatalogIO.Read - Class in org.apache.beam.sdk.io.hcatalog
A PTransform to read data using HCatalog.
HCatalogIO.Write - Class in org.apache.beam.sdk.io.hcatalog
A PTransform to write to a HCatalog managed source.
HCatalogIOTestUtils - Class in org.apache.beam.sdk.io.hcatalog.test
Utility class for HCatalogIOTest and HCatalogIOIT.
HCatalogIOTestUtils() - Constructor for class org.apache.beam.sdk.io.hcatalog.test.HCatalogIOTestUtils
 
HCatalogTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog
Beam SQL table that wraps HCatalogIO.
HCatalogTable() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
 
HCatalogUtils - Class in org.apache.beam.sdk.io.hcatalog
Utility classes to enable meta store conf/client creation.
HCatalogUtils() - Constructor for class org.apache.beam.sdk.io.hcatalog.HCatalogUtils
 
HCatToRow - Class in org.apache.beam.sdk.io.hcatalog
Utilities to convert HCatRecords to Rows.
HCatToRow() - Constructor for class org.apache.beam.sdk.io.hcatalog.HCatToRow
 
HDFSSynchronization - Class in org.apache.beam.sdk.io.hadoop.format
Implementation of ExternalSynchronization which registers locks in the HDFS.
HDFSSynchronization(String) - Constructor for class org.apache.beam.sdk.io.hadoop.format.HDFSSynchronization
Creates instance of HDFSSynchronization.
HeaderAccessor - Interface in org.apache.beam.runners.fnexecution
Interface to access headers in the client request.
Hidden - Annotation Type in org.apache.beam.sdk.options
Methods and/or interfaces annotated with @Hidden will be suppressed from being output when --help is specified on the command-line.
HllCount - Class in org.apache.beam.sdk.extensions.zetasketch
PTransforms to compute HyperLogLogPlusPlus (HLL++) sketches on data streams based on the ZetaSketch implementation.
HllCount.Extract - Class in org.apache.beam.sdk.extensions.zetasketch
Provides PTransforms to extract the estimated count of distinct elements (as Longs) from each HLL++ sketch.
HllCount.Init - Class in org.apache.beam.sdk.extensions.zetasketch
Provides PTransforms to aggregate inputs into HLL++ sketches.
HllCount.Init.Builder<InputT> - Class in org.apache.beam.sdk.extensions.zetasketch
Builder for the HllCount.Init combining PTransform.
HllCount.MergePartial - Class in org.apache.beam.sdk.extensions.zetasketch
Provides PTransforms to merge HLL++ sketches into a new sketch.
HyperLogLogPlusCoder() - Constructor for class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.HyperLogLogPlusCoder
 

I

id() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobPreparation
 
id - Variable in enum org.apache.beam.sdk.io.kafka.KafkaTimestampType
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.Fixed32
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.Fixed64
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.SFixed32
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.SFixed64
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.SInt32
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.SInt64
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.UInt32
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.UInt64
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.CharType
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.DateType
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.TimestampWithLocalTzType
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.TimeType
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.TimeWithLocalTzType
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.NanosDuration
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.NanosInstant
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
IDENTIFIER - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
IDENTIFIER() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifiedComponentContext
 
IDENTIFIER() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.SimpleIdentifierContext
 
Identifier() - Constructor for class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
 
identity() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
Returns the identity element of this operation, i.e.
identity() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
Returns the value that should be used for the combine of the empty set.
identity() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
Returns the identity element of this operation, i.e.
identity() - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
Returns the identity element of this operation, i.e.
identity() - Static method in class org.apache.beam.sdk.transforms.SerializableFunctions
 
IDENTITY_ELEMENT - Static variable in class org.apache.beam.sdk.metrics.DistributionResult
The IDENTITY_ELEMENT is used to start accumulating distributions.
IdGenerator - Interface in org.apache.beam.sdk.fn
A generator of unique IDs.
IdGenerators - Class in org.apache.beam.sdk.fn
Common IdGenerator implementations.
IdGenerators() - Constructor for class org.apache.beam.sdk.fn.IdGenerators
 
ignored() - Static method in interface org.apache.beam.runners.fnexecution.control.BundleProgressHandler
Returns a handler that ignores metrics.
ignoreInput(Watch.Growth.TerminationCondition<?, StateT>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
Wraps a given input-independent Watch.Growth.TerminationCondition as an equivalent condition with a given input type, passing null to the original condition as input.
ignoreInsertIds() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Setting this option to true disables insertId based data deduplication offered by BigQuery.
ignoreUnknownValues() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Accept rows that contain values that do not match the schema.
immediate(T) - Static method in class org.apache.beam.sdk.state.ReadableStates
A ReadableState constructed from a constant value, hence immediately available.
immutableNames() - Method in class org.apache.beam.sdk.metrics.MetricsFilter
 
immutableNamesBuilder() - Method in class org.apache.beam.sdk.metrics.MetricsFilter.Builder
 
immutableSteps() - Method in class org.apache.beam.sdk.metrics.MetricsFilter
 
immutableStepsBuilder() - Method in class org.apache.beam.sdk.metrics.MetricsFilter.Builder
 
implement(EnumerableRelImplementor, EnumerableRel.Prefer) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter
 
Impulse - Class in org.apache.beam.sdk.transforms
For internal use only; no backwards-compatibility guarantees.
IMPULSE_ELEMENT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
ImpulseP - Class in org.apache.beam.runners.jet.processors
/** * Jet Processor implementation for Beam's Impulse primitive.
in(Pipeline) - Static method in class org.apache.beam.sdk.values.PBegin
Returns a PBegin in the given Pipeline.
in(Pipeline) - Static method in class org.apache.beam.sdk.values.PDone
Creates a PDone in the given Pipeline.
inbound(Iterator<ByteString>) - Static method in class org.apache.beam.sdk.fn.stream.DataStreams
Converts multiple ByteStrings into a single InputStream.
InboundDataClient - Interface in org.apache.beam.sdk.fn.data
A client representing some stream of inbound data.
inc() - Method in class org.apache.beam.runners.jet.metrics.CounterImpl
 
inc(long) - Method in class org.apache.beam.runners.jet.metrics.CounterImpl
 
inc() - Method in interface org.apache.beam.sdk.metrics.Counter
Increment the counter.
inc(long) - Method in interface org.apache.beam.sdk.metrics.Counter
Increment the counter by the given amount.
inc() - Method in class org.apache.beam.sdk.metrics.DelegatingCounter
Increment the counter.
inc(long) - Method in class org.apache.beam.sdk.metrics.DelegatingCounter
Increment the counter by the given amount.
include(String, HasDisplayData) - Method in interface org.apache.beam.sdk.transforms.display.DisplayData.Builder
Register display data from the specified subcomponent at the given path.
inCombinedNonLatePanes(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
Creates a new PAssert.IterableAssert like this one, but with the assertion restricted to only run on the provided window across all panes that were not produced by the arrival of late data.
inCombinedNonLatePanes(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
 
IncompatibleWindowException - Exception in org.apache.beam.sdk.transforms.windowing
Exception thrown by WindowFn.verifyCompatibility(WindowFn) if two compared WindowFns are not compatible, including the explanation of incompatibility.
IncompatibleWindowException(WindowFn<?, ?>, String) - Constructor for exception org.apache.beam.sdk.transforms.windowing.IncompatibleWindowException
 
incomplete(List<TimestampedValue<OutputT>>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
Constructs a Watch.Growth.PollResult with the given outputs and declares that new outputs might appear for the current input.
incomplete(Instant, List<OutputT>) - Static method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
Like Watch.Growth.PollResult.incomplete(List), but assigns the same timestamp to all new outputs.
increment() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
Returns a RandomAccessData that is the smallest value of same length which is strictly greater than this.
incrementingLongs() - Static method in class org.apache.beam.sdk.fn.IdGenerators
Returns an IdGenerator which provides successive incrementing longs.
INDEX_OF_MAX - Static variable in class org.apache.beam.sdk.io.ShardNameTemplate
Shard name containing the index and max.
indexOf(String) - Method in class org.apache.beam.sdk.schemas.Schema
Find the index of a given field.
indexOfProjectionColumnRef(long, List<ResolvedColumn>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
Return an index of the projection column reference.
inEarlyGlobalWindowPanes() - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
Creates a new PAssert.IterableAssert like this one, but with the assertion restricted to only run on panes in the GlobalWindow that were emitted before the GlobalWindow closed.
inEarlyGlobalWindowPanes() - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
 
inEarlyPane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
Creates a new PAssert.IterableAssert like this one, but with the assertion restricted to only run on the provided window across all panes that were produced by the arrival of early data.
inEarlyPane(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
 
inEarlyPane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
Creates a new PAssert.SingletonAssert like this one, but with the assertion restricted to only run on the provided window, running the checker only on early panes for each key.
InferableFunction<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
A ProcessFunction which is not a functional interface.
InferableFunction() - Constructor for class org.apache.beam.sdk.transforms.InferableFunction
 
InferableFunction(ProcessFunction<InputT, OutputT>) - Constructor for class org.apache.beam.sdk.transforms.InferableFunction
 
inferType(Object) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
Infer the DisplayData.Type for the given object.
inFinalPane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
Creates a new PAssert.IterableAssert like this one, but with the assertion restricted to only run on the provided window, running the checker only on the final pane for each key.
inFinalPane(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
 
inFinalPane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
Creates a new PAssert.SingletonAssert like this one, but with the assertion restricted to only run on the provided window, running the checker only on the final pane for each key.
INHERIT_IO_FILE - Static variable in class org.apache.beam.runners.fnexecution.environment.ProcessManager
A symbolic file to indicate that we want to inherit I/O of parent process.
init() - Method in class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator.Merge
 
init(Processor.Context) - Method in class org.apache.beam.runners.jet.processors.BoundedSourceP
 
init(Processor.Context) - Method in class org.apache.beam.runners.jet.processors.UnboundedSourceP
 
init(SparkPipelineOptions, JavaSparkContext) - Static method in class org.apache.beam.runners.spark.aggregators.AggregatorsAccumulator
Init aggregators accumulator if it has not been initiated.
init(SparkPipelineOptions, JavaSparkContext) - Static method in class org.apache.beam.runners.spark.metrics.MetricsAccumulator
Init metrics accumulator if it has not been initiated.
init(JavaSparkContext) - Static method in class org.apache.beam.runners.spark.structuredstreaming.aggregators.AggregatorsAccumulator
Init aggregators accumulator if it has not been initiated.
init(JavaSparkContext) - Static method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsAccumulator
Init metrics accumulator if it has not been initiated.
initAccumulators(SparkPipelineOptions, JavaSparkContext) - Static method in class org.apache.beam.runners.spark.SparkRunner
Init Metrics/Aggregators accumulators.
initAccumulators(SparkStructuredStreamingPipelineOptions, JavaSparkContext) - Static method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingRunner
Init Metrics/Aggregators accumulators.
initialBackoff() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
 
initialize(Map<ExecutableT, ? extends Iterable<Bundle<?, CollectionT>>>) - Method in class org.apache.beam.runners.direct.WatermarkManager
 
initialize(HttpRequest) - Method in class org.apache.beam.sdk.extensions.gcp.auth.NullCredentialInitializer
 
initialize(AbstractGoogleClientRequest<?>) - Method in class org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions.GoogleApiTracer
 
initialize(HttpRequest) - Method in class org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer
 
initialSystemTimeAt(Instant) - Method in class org.apache.beam.runners.spark.io.CreateStream
Set the initial synchronized processing time.
InjectPackageStrategy(Class<?>) - Constructor for class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.InjectPackageStrategy
 
inLatePane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
Creates a new PAssert.IterableAssert with the assertion restricted to only run on the provided window across all panes that were produced by the arrival of late data.
inLatePane(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
 
inLatePane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
Creates a new PAssert.SingletonAssert with the assertion restricted to only run on the provided window, running the checker only on late panes for each key.
inMemory(TableProvider...) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
This method creates BeamSqlEnv using empty * Pipeline Options.
InMemoryBagUserStateFactory<K,V,W extends BoundedWindow> - Class in org.apache.beam.runners.fnexecution.state
Holds user state in memory.
InMemoryBagUserStateFactory() - Constructor for class org.apache.beam.runners.fnexecution.state.InMemoryBagUserStateFactory
 
inMemoryFinalizer(InstructionRequestHandler) - Static method in class org.apache.beam.runners.fnexecution.control.BundleFinalizationHandlers
A bundle finalizer that stores all bundle finalization requests in memory.
InMemoryJobService - Class in org.apache.beam.runners.fnexecution.jobsubmission
A InMemoryJobService that prepares and runs jobs on behalf of a client using a JobInvoker.
InMemoryMetaStore - Class in org.apache.beam.sdk.extensions.sql.meta.store
A MetaStore which stores the meta info in memory.
InMemoryMetaStore() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.store.InMemoryMetaStore
 
InMemoryMetaTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider
A InMemoryMetaTableProvider is an abstract TableProvider for in-memory types.
InMemoryMetaTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.InMemoryMetaTableProvider
 
inNamespace(String) - Static method in class org.apache.beam.sdk.metrics.MetricNameFilter
 
inNamespace(Class<?>) - Static method in class org.apache.beam.sdk.metrics.MetricNameFilter
 
Inner() - Constructor for class org.apache.beam.sdk.schemas.transforms.Filter.Inner
 
innerBroadcastJoin(PCollection<RhsT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Join
Perform an inner join, broadcasting the right side.
innerJoin(PCollection<KV<K, V1>>, PCollection<KV<K, V2>>) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
Inner join of two collections of KV elements.
innerJoin(String, PCollection<KV<K, V1>>, PCollection<KV<K, V2>>) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
Inner join of two collections of KV elements.
innerJoin(PCollection<RhsT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Join
Perform an inner join.
inOnlyPane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
Creates a new PAssert.SingletonAssert like this one, but with the assertion restricted to only run on the provided window.
inOnTimePane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
Creates a new PAssert.IterableAssert like this one, but with the assertion restricted to only run on the provided window.
inOnTimePane(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
 
inOnTimePane(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
Creates a new PAssert.SingletonAssert like this one, but with the assertion restricted to only run on the provided window, running the checker only on the on-time pane for each key.
inOrder(Trigger...) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterEach
Returns an AfterEach Trigger with the given subtriggers.
inOrder(List<Trigger>) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterEach
Returns an AfterEach Trigger with the given subtriggers.
InProcessManagedChannelFactory - Class in org.apache.beam.sdk.fn.test
A ManagedChannelFactory that uses in-process channels.
InProcessServerFactory - Class in org.apache.beam.runners.fnexecution
A ServerFactory which creates servers with the InProcessServerBuilder.
inputOf(ProcessFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
Returns a type descriptor for the input of the given ProcessFunction, subject to Java type erasure: may contain unresolved type variables if the type was erased.
inputOf(SerializableFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
Binary compatibility adapter for TypeDescriptors.inputOf(ProcessFunction).
inputOf(Contextful.Fn<InputT, OutputT>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
Like TypeDescriptors.inputOf(ProcessFunction) but for Contextful.Fn.
INPUTS - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
insertAll(TableReference, List<ValueInSingleWindow<TableRow>>, List<String>, InsertRetryPolicy, List<ValueInSingleWindow<T>>, ErrorContainer<T>, boolean, boolean, boolean) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
Inserts TableRows with the specified insertIds if not null.
insertAll(TableReference, List<TableRow>, List<String>) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
insertAll(TableReference, List<ValueInSingleWindow<TableRow>>, List<String>, InsertRetryPolicy, List<ValueInSingleWindow<T>>, ErrorContainer<T>, boolean, boolean, boolean) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
insertDataToTable(String, String, String, List<Map<String, Object>>) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
Inserts rows to a table using a BigQuery streaming write.
insertDeduplicate() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
 
insertDistributedSync() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
 
insertQuorum() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
 
InsertRetryPolicy - Class in org.apache.beam.sdk.io.gcp.bigquery
A retry policy for streaming BigQuery inserts.
InsertRetryPolicy() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy
 
InsertRetryPolicy.Context - Class in org.apache.beam.sdk.io.gcp.bigquery
Contains information about a failed insert.
insertRows(Schema, Row...) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery
 
insertTestData(Map<String, String>) - Static method in class org.apache.beam.sdk.io.hcatalog.test.HCatalogIOTestUtils
Inserts data into test datastore.
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.JdbcDriver
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamJavaTypeFactory
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelDataTypeSystem
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamAggregateProjectMergeRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamAggregationRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamBasicAggregationRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamCalcRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamCoGBKJoinRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamEnumerableConverterRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIntersectRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIOPushDownRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIOSinkRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamJoinAssociateRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamMinusRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSideInputJoinRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSideInputLookupJoinRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSortRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamUncollectRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamUnionRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamUnnestRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamValuesRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcRule
 
INSTANCE - Static variable in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow.Coder
 
INSTANCE - Static variable in class org.apache.beam.sdk.schemas.AutoValueSchema.AbstractGetterTypeSupplier
 
INSTANCE - Static variable in class org.apache.beam.sdk.schemas.JavaBeanSchema.GetterTypeSupplier
 
INSTANCE - Static variable in class org.apache.beam.sdk.schemas.JavaFieldSchema.JavaFieldTypeSupplier
 
INSTANCE - Static variable in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
 
INSTANCE - Static variable in class org.apache.beam.sdk.transforms.windowing.GlobalWindow
Singleton instance of GlobalWindow.
INSTANCE - Static variable in class org.apache.beam.sdk.transforms.windowing.PaneInfo.PaneInfoCoder
 
InstantCoder - Class in org.apache.beam.sdk.coders
A Coder for joda Instant that encodes it as a big endian Long shifted such that lexicographic ordering of the bytes corresponds to chronological order.
InstantDeserializer - Class in org.apache.beam.sdk.io.kafka.serialization
InstantDeserializer() - Constructor for class org.apache.beam.sdk.io.kafka.serialization.InstantDeserializer
 
instantiateCoder(String, RunnerApi.Components) - Static method in class org.apache.beam.runners.fnexecution.translation.PipelineTranslatorUtils
Creates a coder for a given PCollection id from the Proto definition.
instantiateRunnerWireCoder(PipelineNode.PCollectionNode, RunnerApi.Components) - Static method in class org.apache.beam.runners.fnexecution.wire.WireCoders
Instantiates a runner-side wire coder for the given PCollection.
instantiateRunnerWireCoder(PipelineNode.PCollectionNode, RunnerApi.Components, RunnerApi.ExecutableStagePayload.WireCoderSetting) - Static method in class org.apache.beam.runners.fnexecution.wire.WireCoders
Instantiates a runner-side wire coder for the given PCollection.
InstantSerializer - Class in org.apache.beam.sdk.io.kafka.serialization
Kafka Serializer for Instant.
InstantSerializer() - Constructor for class org.apache.beam.sdk.io.kafka.serialization.InstantSerializer
 
InstructionRequestHandler - Interface in org.apache.beam.runners.fnexecution.control
Interface for any function that can handle a Fn API BeamFnApi.InstructionRequest.
INT16 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
INT16 - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
The type of int16 fields.
INT32 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
INT32 - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
The type of int32 fields.
INT64 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
INT64 - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
The type of int64 fields.
INT8 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
INTEGER - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
integers() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for Integer.
integersGlobally() - Static method in class org.apache.beam.sdk.transforms.Max
Returns a PTransform that takes an input PCollection<Integer> and returns a PCollection<Integer> whose contents is the maximum of the input PCollection's elements, or Integer.MIN_VALUE if there are no elements.
integersGlobally() - Static method in class org.apache.beam.sdk.transforms.Min
Returns a PTransform that takes an input PCollection<Integer> and returns a PCollection<Integer> whose contents is a single value that is the minimum of the input PCollection's elements, or Integer.MAX_VALUE if there are no elements.
integersGlobally() - Static method in class org.apache.beam.sdk.transforms.Sum
Returns a PTransform that takes an input PCollection<Integer> and returns a PCollection<Integer> whose contents is the sum of the input PCollection's elements, or 0 if there are no elements.
integersPerKey() - Static method in class org.apache.beam.sdk.transforms.Max
Returns a PTransform that takes an input PCollection<KV<K, Integer>> and returns a PCollection<KV<K, Integer>> that contains an output element mapping each distinct key in the input PCollection to the maximum of the values associated with that key in the input PCollection.
integersPerKey() - Static method in class org.apache.beam.sdk.transforms.Min
Returns a PTransform that takes an input PCollection<KV<K, Integer>> and returns a PCollection<KV<K, Integer>> that contains an output element mapping each distinct key in the input PCollection to the minimum of the values associated with that key in the input PCollection.
integersPerKey() - Static method in class org.apache.beam.sdk.transforms.Sum
Returns a PTransform that takes an input PCollection<KV<K, Integer>> and returns a PCollection<KV<K, Integer>> that contains an output element mapping each distinct key in the input PCollection to the sum of the values associated with that key in the input PCollection.
interceptor() - Static method in class org.apache.beam.runners.fnexecution.GrpcContextHeaderAccessorProvider
 
interceptResponse(HttpResponse) - Method in class org.apache.beam.sdk.extensions.gcp.util.UploadIdResponseInterceptor
 
Internal - Annotation Type in org.apache.beam.sdk.annotations
Signifies that a publicly accessible API (public class, method or field) is intended for internal use only and not for public consumption.
interpolateKey(double) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
Returns a ByteKey key such that [startKey, key) represents approximately the specified fraction of the range [startKey, endKey).
intersects(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
Returns whether this window intersects the given window.
IntervalWindow - Class in org.apache.beam.sdk.transforms.windowing
An implementation of BoundedWindow that represents an interval from IntervalWindow.start (inclusive) to IntervalWindow.end (exclusive).
IntervalWindow(Instant, Instant) - Constructor for class org.apache.beam.sdk.transforms.windowing.IntervalWindow
Creates a new IntervalWindow that represents the half-open time interval [start, end).
IntervalWindow(Instant, ReadableDuration) - Constructor for class org.apache.beam.sdk.transforms.windowing.IntervalWindow
 
IntervalWindow.IntervalWindowCoder - Class in org.apache.beam.sdk.transforms.windowing
Encodes an IntervalWindow as a pair of its upper bound and duration.
IntervalWindowCoder() - Constructor for class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
 
into(TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.transforms.FlatMapElements
Returns a new FlatMapElements transform with the given type descriptor for the output type, but the mapping function yet to be specified using FlatMapElements.via(ProcessFunction).
into(TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.transforms.MapElements
Returns a new MapElements transform with the given type descriptor for the output type, but the mapping function yet to be specified using MapElements.via(ProcessFunction).
into(WindowFn<? super T, ?>) - Static method in class org.apache.beam.sdk.transforms.windowing.Window
Creates a Window PTransform that uses the given WindowFn to window the data.
InvalidTableException - Exception in org.apache.beam.sdk.extensions.sql.meta.provider
Exception thrown when the request for a table is invalid, such as invalid metadata.
InvalidTableException(String) - Constructor for exception org.apache.beam.sdk.extensions.sql.meta.provider.InvalidTableException
 
InvalidTableException(String, Throwable) - Constructor for exception org.apache.beam.sdk.extensions.sql.meta.provider.InvalidTableException
 
InvalidTableException(Throwable) - Constructor for exception org.apache.beam.sdk.extensions.sql.meta.provider.InvalidTableException
 
InvalidWindows<W extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
A WindowFn that represents an invalid pipeline state.
InvalidWindows(String, WindowFn<?, W>) - Constructor for class org.apache.beam.sdk.transforms.windowing.InvalidWindows
 
invokeAdvance(ReaderT) - Method in class org.apache.beam.runners.flink.metrics.ReaderInvocationUtil
 
invokeStart(ReaderT) - Method in class org.apache.beam.runners.flink.metrics.ReaderInvocationUtil
 
invokeWithExecutor(RunnerApi.Pipeline, Struct, String, ListeningExecutorService) - Method in class org.apache.beam.runners.flink.FlinkJobInvoker
 
invokeWithExecutor(RunnerApi.Pipeline, Struct, String, ListeningExecutorService) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobInvoker
Start running a job, abstracting its state as a JobInvocation instance.
invokeWithExecutor(RunnerApi.Pipeline, Struct, String, ListeningExecutorService) - Method in class org.apache.beam.runners.spark.SparkJobInvoker
 
inWindow(BoundedWindow) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
Creates a new PAssert.IterableAssert like this one, but with the assertion restricted to only run on the provided window.
inWindow(BoundedWindow) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
 
ioException() - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.StorageObjectOrIOException
Returns the IOException.
IS_MERGING_WINDOW_FN - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
IS_PAIR_LIKE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
IS_STREAM_LIKE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
IS_WRAPPER - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
isAbsolute() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
isAccessible() - Method in interface org.apache.beam.sdk.options.ValueProvider
Whether the contents of this ValueProvider is currently available via ValueProvider.get().
isAccessible() - Method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
 
isAccessible() - Method in class org.apache.beam.sdk.options.ValueProvider.RuntimeValueProvider
 
isAccessible() - Method in class org.apache.beam.sdk.options.ValueProvider.StaticValueProvider
 
isAliveOrThrow() - Method in class org.apache.beam.runners.fnexecution.environment.ProcessManager.RunningProcess
Checks if the underlying process is still running.
isAllowedLatenessSpecified() - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
isArray() - Method in class org.apache.beam.sdk.values.TypeDescriptor
Returns true if this type is known to be an array type.
isAtSplitPoint() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
Returns true if the reader is at a split point.
isAtSplitPoint() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
Returns true only for the first record; compressed sources cannot be split.
isAtSplitPoint() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
Returns whether the current record is at a split point (i.e., whether the current record would be the first record to be read by a source with a specified start offset of OffsetBasedSource.OffsetBasedReader.getCurrentOffset()).
isAutoBalanceWriteFilesShardingEnabled() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
isBlockOnRun() - Method in interface org.apache.beam.runners.direct.DirectOptions
 
isBounded() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSourceRel
 
isBounded() - Method in interface org.apache.beam.sdk.extensions.sql.impl.rel.BeamRelNode
Whether the collection of rows represented by this relational expression is bounded (known to be finite) or unbounded (may or may not be finite).
isBounded() - Method in class org.apache.beam.sdk.extensions.sql.impl.schema.BeamPCollectionTable
 
isBounded() - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTable
Whether this table is bounded (known to be finite) or unbounded (may or may not be finite).
isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.avro.AvroTable
 
isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
 
isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
 
isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable
 
isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.parquet.ParquetTable
 
isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestBoundedTable
 
isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestUnboundedTable
 
isBounded() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTable
 
isBounded() - Method in class org.apache.beam.sdk.values.PCollection
 
isBoundedCollection(Collection<PValue>) - Method in class org.apache.beam.runners.spark.SparkRunner.Evaluator
 
isCacheDisabled() - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
isCleanArtifactsPerJob() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver.ServerConfiguration
 
isClosed() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
isCollectionType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
isCompatible(Trigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
 
isCompatible(Trigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
 
isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
 
isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
 
isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
 
isCompatible(Trigger) - Method in class org.apache.beam.sdk.transforms.windowing.DefaultTrigger
 
isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
 
isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
 
isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
InvalidWindows objects with the same originalWindowFn are compatible.
isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
 
isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
 
isCompatible(Trigger) - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
For internal use only; no backwards-compatibility guarantees.
isCompatible(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
Deprecated.
please override verifyCompatibility to throw a useful error message; we will remove isCompatible at version 3.0.0
isCompositeType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
isCompound() - Method in class org.apache.beam.sdk.extensions.sql.impl.TableName
Whether it's a compound table name (with multiple path components).
isCompressed(String) - Static method in enum org.apache.beam.sdk.io.CompressedSource.CompressionMode
Deprecated.
Returns whether the file's extension matches of one of the known compression formats.
isCompressed(String) - Method in enum org.apache.beam.sdk.io.Compression
 
isCooperative() - Method in class org.apache.beam.runners.jet.processors.BoundedSourceP
 
isCooperative() - Method in class org.apache.beam.runners.jet.processors.UnboundedSourceP
 
isDateTimeType(Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
Returns true if the type is any of the various date time types.
isDateType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
isDdl(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
 
isDecimal(Schema.TypeName) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
Checks if type is decimal.
isDirectory() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
 
isDirectory() - Method in interface org.apache.beam.sdk.io.fs.ResourceId
Returns true if this ResourceId represents a directory, false otherwise.
isDisjoint(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
Returns whether this window is disjoint from the given window.
isDone() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataInboundObserver
 
isDone() - Method in class org.apache.beam.sdk.fn.data.CompletableFutureInboundDataClient
 
isDone() - Method in interface org.apache.beam.sdk.fn.data.InboundDataClient
Returns true if the client is done, either via completing successfully or by being cancelled.
isDone() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
isDone() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
 
isDone() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
 
isEmbeddedExecution() - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
 
isEmbeddedExecutionDebugMode() - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
 
isEmpty() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.SparkSideInputReader
 
isEmpty() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.utils.CachedSideInputReader
 
isEmpty() - Method in class org.apache.beam.runners.spark.util.CachedSideInputReader
 
isEmpty() - Method in class org.apache.beam.runners.spark.util.SparkSideInputReader
 
isEmpty() - Method in class org.apache.beam.sdk.io.range.ByteKey
Returns true if the byte[] backing this ByteKey is of length 0.
isEmpty() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
isEmpty() - Method in interface org.apache.beam.sdk.state.GroupingState
Returns a ReadableState whose ReadableState.read() method will return true if this state is empty at the point when that ReadableState.read() call returns.
isEmpty() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
 
isEmpty() - Method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
 
isEmpty() - Method in class org.apache.beam.sdk.transforms.Requirements
Whether this is an empty set of requirements.
isEnableStreamingEngine() - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
 
isEnforceEncodability() - Method in interface org.apache.beam.runners.direct.DirectOptions
 
isEnforceImmutability() - Method in interface org.apache.beam.runners.direct.DirectOptions
 
isEOF() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
isEOF() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
 
isEOF() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
 
isEOF() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
isEqualTo(T) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
Asserts that the value in question is equal to the provided value, according to Object.equals(java.lang.Object).
isEqWithEpsilon(RelOptCost) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
isExternalizedCheckpointsEnabled() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
Enables or disables externalized checkpoints.
isFinished() - Method in class org.apache.beam.runners.apex.ApexYarnLauncher.ProcessWatcher
 
isFirst() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
Return true if this is the first pane produced for the associated window.
IsFlinkNativeTransform() - Constructor for class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.IsFlinkNativeTransform
 
IsFlinkNativeTransform() - Constructor for class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.IsFlinkNativeTransform
 
isFnApi() - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
 
isForceStreaming() - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
 
isForceWatermarkSync() - Method in class org.apache.beam.runners.spark.io.CreateStream
 
isGetter(Method) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
 
isIn(Collection<T>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.isIn(Collection).
isIn(Coder<T>, Collection<T>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.isIn(Collection).
isIn(T[]) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.isIn(Object[]).
isIn(Coder<T>, T[]) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.isIn(Object[]).
isInboundEdgeOfVertex(Edge, String, String, String) - Method in interface org.apache.beam.runners.jet.DAGBuilder.WiringListener
 
isInboundEdgeOfVertex(Edge, String, String, String) - Method in class org.apache.beam.runners.jet.processors.FlattenP.Supplier
 
IsInf - Class in org.apache.beam.sdk.extensions.sql.impl.udf
IS_INF(X)
IsInf() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.udf.IsInf
 
isInf(Double) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.IsInf
 
isInf(Float) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.IsInf
 
isInfinite() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
isInputSortRelAndLimitOnly() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.AbstractBeamCalcRel
 
isInsertDefault() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
Returns whether or not this transformation applies a default value.
isIntegral(Schema.TypeName) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
Checks if type is integral.
isJoinLegal(Join) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
This method checks if a join is legal and can be converted into Beam SQL.
isKey(ImmutableBitSet) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
 
isLast() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
Return true if this is the last pane that will be produced in the associated window.
isLe(RelOptCost) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
isLimitOnly() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSortRel
 
isLocalSparkMaster(SparkPipelineOptions) - Static method in interface org.apache.beam.runners.spark.SparkPipelineOptions
Detects if the pipeline is run in spark local mode.
isLocalSparkMaster(SparkStructuredStreamingPipelineOptions) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
Detects if the pipeline is run in local spark mode.
isLogicalType(String) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
 
isLogicalType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
isLt(RelOptCost) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
isMapType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
isMetricsSupported() - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
Indicates whether metrics reporting is supported.
isModeSpecified() - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
isMutable() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
IsNan - Class in org.apache.beam.sdk.extensions.sql.impl.udf
IS_NAN(X)
IsNan() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.udf.IsNan
 
isNan(Float) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.IsNan
 
isNan(Double) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.IsNan
 
isNonMerging() - Method in class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator.GearpumpWindowFn
 
isNonMerging() - Method in class org.apache.beam.sdk.transforms.windowing.NonMergingWindowFn
 
isNonMerging() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
Returns true if this WindowFn never needs to merge any windows.
isNullable() - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
Returns whether the field is nullable.
isNumericType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
isOneOf(T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.isOneOf(T...).
isOneOf(Coder<T>, T...) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.isOneOf(T...).
isOutboundEdgeOfVertex(Edge, String, String, String) - Method in interface org.apache.beam.runners.jet.DAGBuilder.WiringListener
 
isOutboundEdgeOfVertex(Edge, String, String, String) - Method in class org.apache.beam.runners.jet.processors.FlattenP.Supplier
 
isParDoFusionEnabled() - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
 
isPrimitiveType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
isReadOnly() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
isReadSeekEfficient() - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata
 
isRegisterByteSizeObserverCheap(RandomAccessData) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
 
isRegisterByteSizeObserverCheap(BigDecimal) - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(Integer) - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(Long) - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(Short) - Method in class org.apache.beam.sdk.coders.BigEndianShortCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(BigInteger) - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(Boolean) - Method in class org.apache.beam.sdk.coders.BooleanCoder
 
isRegisterByteSizeObserverCheap(byte[]) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(Byte) - Method in class org.apache.beam.sdk.coders.ByteCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(T) - Method in class org.apache.beam.sdk.coders.Coder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(Double) - Method in class org.apache.beam.sdk.coders.DoubleCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(ReadableDuration) - Method in class org.apache.beam.sdk.coders.DurationCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(Float) - Method in class org.apache.beam.sdk.coders.FloatCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(Instant) - Method in class org.apache.beam.sdk.coders.InstantCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(IterableT) - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(KV<K, V>) - Method in class org.apache.beam.sdk.coders.KvCoder
Returns whether both keyCoder and valueCoder are considered not expensive.
isRegisterByteSizeObserverCheap(T) - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
LengthPrefixCoder is cheap if valueCoder is cheap.
isRegisterByteSizeObserverCheap(T) - Method in class org.apache.beam.sdk.coders.NullableCoder
NullableCoder is cheap if valueCoder is cheap.
isRegisterByteSizeObserverCheap(Integer) - Method in class org.apache.beam.sdk.coders.VarIntCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(Long) - Method in class org.apache.beam.sdk.coders.VarLongCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(Void) - Method in class org.apache.beam.sdk.coders.VoidCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(ByteString) - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
Returns whether Coder.registerByteSizeObserver(T, org.apache.beam.sdk.util.common.ElementByteSizeObserver) cheap enough to call for every element, that is, if this Coder can calculate the byte size of the element to be coded in roughly constant time (or lazily).
isRegisterByteSizeObserverCheap(HyperLogLogPlus) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.HyperLogLogPlusCoder
 
isRegisterByteSizeObserverCheap(EncodedBoundedWindow) - Method in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow.Coder
 
isRegisterByteSizeObserverCheap(KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
 
isRegisterByteSizeObserverCheap(ProducerRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
 
isRegisterByteSizeObserverCheap(OffsetRange) - Method in class org.apache.beam.sdk.io.range.OffsetRange.Coder
 
isRegisterByteSizeObserverCheap(RawUnionValue) - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
Since this coder uses elementCoders.get(index) and coders that are known to run in constant time, we defer the return value to that coder.
isRunnerDeterminedSharding() - Method in interface org.apache.beam.runners.direct.DirectTestOptions
 
isSetter(Method) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
 
isShutdownSourcesOnFinalWatermark() - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
Whether to shutdown sources when their watermark reaches +Inf.
isSideInputLookupJoin() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
 
isSimple() - Method in class org.apache.beam.sdk.extensions.sql.impl.TableName
Whether it's a simple name, with a single name component.
isSplittable() - Method in class org.apache.beam.sdk.io.CompressedSource
Determines whether a single file represented by this source is splittable.
isSplittable() - Method in class org.apache.beam.sdk.io.FileBasedSource
Determines whether a file represented by this source is can be split into bundles.
isStarted() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
Returns true if there has been a call to OffsetBasedSource.OffsetBasedReader.start().
isStarted() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
 
isStreaming() - Method in interface org.apache.beam.sdk.options.StreamingOptions
Set to true if running a streaming pipeline.
isStreamingEngine() - Method in interface org.apache.beam.runners.dataflow.TransformTranslator.TranslationContext
 
isStringType(Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
isStringType() - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
isSubtypeOf(Schema.TypeName) - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
isSubtypeOf(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.values.TypeDescriptor
Return true if this type is a subtype of the given type.
isSuccess() - Method in class org.apache.beam.sdk.io.tika.ParseResult
Returns whether this file was parsed successfully.
isSuccess() - Method in class org.apache.beam.sdk.testing.SuccessOrFailure
 
isSupertypeOf(Schema.TypeName) - Method in enum org.apache.beam.sdk.schemas.Schema.TypeName
Whether this is a super type of the another type.
isSupertypeOf(TypeDescriptor<?>) - Method in class org.apache.beam.sdk.values.TypeDescriptor
Returns true if this type is assignable from the given type.
isSupported() - Method in enum org.apache.beam.sdk.extensions.sql.meta.ProjectSupport
 
isTableEmpty(TableReference) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
Returns true if the table is empty.
isTableEmpty(TableReference) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
isTableResolved(Table) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
True if the table was resolved using the Calcite schema.
isTermainal() - Method in enum org.apache.beam.runners.local.ExecutionDriver.DriverState
 
isTerminal() - Method in enum org.apache.beam.sdk.PipelineResult.State
 
isTimer() - Method in class org.apache.beam.sdk.fn.data.LogicalEndpoint
 
isTimestampCombinerSpecified() - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
isTriggerSpecified() - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
isTrustSelfSignedCerts() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
 
isTupleTracingEnabled() - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
 
isUnknown() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
 
isUnknown() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
If any of the values for rowCount, rate or window is infinite, it returns true.
isUnknown() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
Return true if there is no timing information for the current PaneInfo.
isUpdate() - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
Whether to update the currently running pipeline with the same name as this one.
isValid(int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
isWholeStream - Variable in class org.apache.beam.sdk.coders.Coder.Context
Deprecated.
Whether the encoded or decoded value fills the remainder of the output or input (resp.) record/stream contents.
isWildcard(GcsPath) - Static method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
Returns true if the given spec contains wildcard.
isWrapperFor(Class<?>) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
isZero() - Method in class org.apache.beam.runners.spark.aggregators.NamedAggregatorsAccumulator
 
isZero() - Method in class org.apache.beam.runners.spark.metrics.MetricsContainerStepMapAccumulator
 
isZero() - Method in class org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregatorsAccumulator
 
isZero() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsContainerStepMapAccumulator
 
item(String, String) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
Create a display item for the specified key and string value.
item(String, ValueProvider<?>) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
Create a display item for the specified key and ValueProvider.
item(String, Integer) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
Create a display item for the specified key and integer value.
item(String, Long) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
Create a display item for the specified key and integer value.
item(String, Float) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
Create a display item for the specified key and floating point value.
item(String, Double) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
Create a display item for the specified key and floating point value.
item(String, Boolean) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
Create a display item for the specified key and boolean value.
item(String, Instant) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
Create a display item for the specified key and timestamp value.
item(String, Duration) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
Create a display item for the specified key and duration value.
item(String, Class<T>) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
Create a display item for the specified key and class value.
item(String, DisplayData.Type, T) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
Create a display item for the specified key, type, and value.
Item() - Constructor for class org.apache.beam.sdk.transforms.display.DisplayData.Item
 
items() - Method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Read
 
items() - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
 
items() - Method in class org.apache.beam.sdk.transforms.display.DisplayData
 
ItemSpec() - Constructor for class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
 
iterable(Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
 
iterable() - Static method in class org.apache.beam.sdk.transforms.Materializations
For internal use only; no backwards-compatibility guarantees.
ITERABLE_MATERIALIZATION_URN - Static variable in class org.apache.beam.sdk.transforms.Materializations
The URN for a Materialization where the primitive view type is an iterable of fully specified windowed values.
IterableCoder<T> - Class in org.apache.beam.sdk.coders
An IterableCoder encodes any Iterable in the format of IterableLikeCoder.
IterableCoder(Coder<T>) - Constructor for class org.apache.beam.sdk.coders.IterableCoder
 
IterableLikeCoder<T,IterableT extends java.lang.Iterable<T>> - Class in org.apache.beam.sdk.coders
An abstract base class with functionality for assembling a Coder for a class that implements Iterable.
IterableLikeCoder(Coder<T>, String) - Constructor for class org.apache.beam.sdk.coders.IterableLikeCoder
 
iterables() - Static method in class org.apache.beam.sdk.transforms.Flatten
Returns a PTransform that takes a PCollection<Iterable<T>> and returns a PCollection<T> containing all the elements from all the Iterables.
iterables() - Static method in class org.apache.beam.sdk.transforms.ToString
Transforms each item in the iterable of the input PCollection to a String using the Object.toString() method followed by a "," until the last element in the iterable.
iterables(String) - Static method in class org.apache.beam.sdk.transforms.ToString
Transforms each item in the iterable of the input PCollection to a String using the Object.toString() method followed by the specified delimiter until the last element in the iterable.
iterables(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for Iterable.
iterableView(PCollection<KV<Void, T>>, PCollectionViews.TypeDescriptorSupplier<T>, WindowingStrategy<?, W>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
Returns a PCollectionView<Iterable<T>> capable of processing elements windowed using the provided WindowingStrategy.
IterableViewFn(PCollectionViews.TypeDescriptorSupplier<T>) - Constructor for class org.apache.beam.sdk.values.PCollectionViews.IterableViewFn
 
iterableWithSize(int) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
iterableWithSize(SerializableMatcher<? super Integer>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
iterator() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
iterator() - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
 
iterator() - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices.FakeBigQueryServerStream
 

J

java8TimeToJodaTime(Instant) - Static method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils
 
JavaBeanSchema - Class in org.apache.beam.sdk.schemas
A SchemaProvider for Java Bean objects.
JavaBeanSchema() - Constructor for class org.apache.beam.sdk.schemas.JavaBeanSchema
 
JavaBeanSchema.GetterTypeSupplier - Class in org.apache.beam.sdk.schemas
FieldValueTypeSupplier that's based on getter methods.
JavaBeanSchema.SetterTypeSupplier - Class in org.apache.beam.sdk.schemas
FieldValueTypeSupplier that's based on setter methods.
JavaBeanUtils - Class in org.apache.beam.sdk.schemas.utils
A set of utilities to generate getter and setter classes for JavaBean objects.
JavaBeanUtils() - Constructor for class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
 
JavaFieldSchema - Class in org.apache.beam.sdk.schemas
A SchemaProvider for Java POJO objects.
JavaFieldSchema() - Constructor for class org.apache.beam.sdk.schemas.JavaFieldSchema
 
JavaFieldSchema.JavaFieldTypeSupplier - Class in org.apache.beam.sdk.schemas
FieldValueTypeSupplier that's based on public fields.
JavaFieldTypeSupplier() - Constructor for class org.apache.beam.sdk.schemas.JavaFieldSchema.JavaFieldTypeSupplier
 
JavaFilesystemLegacyArtifactStagingService - Class in org.apache.beam.runners.fnexecution.artifact
An LegacyArtifactStagingServiceGrpc.LegacyArtifactStagingServiceImplBase that loads artifacts into a Java FileSystem.
JavaFilesystemLegacyArtifactStagingService(FileSystem, String) - Constructor for class org.apache.beam.runners.fnexecution.artifact.JavaFilesystemLegacyArtifactStagingService
 
javaObjectToZetaSqlValue(Object, Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlUtils
 
javaTypeForFieldType(Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.FieldTypeDescriptors
JAXBCoder<T> - Class in org.apache.beam.sdk.io.xml
A coder for JAXB annotated objects.
JdbcConnection - Class in org.apache.beam.sdk.extensions.sql.impl
Beam JDBC Connection.
JdbcDriver - Class in org.apache.beam.sdk.extensions.sql.impl
Calcite JDBC driver with Beam defaults.
JdbcDriver() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.JdbcDriver
 
JdbcIO - Class in org.apache.beam.sdk.io.jdbc
IO to read and write data on JDBC.
JdbcIO.DataSourceConfiguration - Class in org.apache.beam.sdk.io.jdbc
A POJO describing a DataSource, either providing directly a DataSource or all properties allowing to create a DataSource.
JdbcIO.DataSourceProviderFromDataSourceConfiguration - Class in org.apache.beam.sdk.io.jdbc
Wraps a JdbcIO.DataSourceConfiguration to provide a DataSource.
JdbcIO.DefaultRetryStrategy - Class in org.apache.beam.sdk.io.jdbc
This is the default Predicate we use to detect DeadLock.
JdbcIO.PoolableDataSourceProvider - Class in org.apache.beam.sdk.io.jdbc
Wraps a JdbcIO.DataSourceConfiguration to provide a PoolingDataSource.
JdbcIO.PreparedStatementSetter<T> - Interface in org.apache.beam.sdk.io.jdbc
An interface used by the JdbcIO Write to set the parameters of the PreparedStatement used to setParameters into the database.
JdbcIO.Read<T> - Class in org.apache.beam.sdk.io.jdbc
Implementation of JdbcIO.read().
JdbcIO.ReadAll<ParameterT,OutputT> - Class in org.apache.beam.sdk.io.jdbc
Implementation of JdbcIO.readAll().
JdbcIO.ReadRows - Class in org.apache.beam.sdk.io.jdbc
Implementation of JdbcIO.readRows().
JdbcIO.RetryStrategy - Interface in org.apache.beam.sdk.io.jdbc
An interface used to control if we retry the statements when a SQLException occurs.
JdbcIO.RowMapper<T> - Interface in org.apache.beam.sdk.io.jdbc
An interface used by JdbcIO.Read for converting each row of the ResultSet into an element of the resulting PCollection.
JdbcIO.StatementPreparator - Interface in org.apache.beam.sdk.io.jdbc
An interface used by the JdbcIO Write to set the parameters of the PreparedStatement used to setParameters into the database.
JdbcIO.Write<T> - Class in org.apache.beam.sdk.io.jdbc
This class is used as the default return value of JdbcIO.write().
JdbcIO.WriteVoid<T> - Class in org.apache.beam.sdk.io.jdbc
A PTransform to write to a JDBC datasource.
jdbcUrl() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
 
JetMetricResults - Class in org.apache.beam.runners.jet.metrics
Jet specific MetricResults.
JetMetricResults(IMap<String, MetricUpdates>) - Constructor for class org.apache.beam.runners.jet.metrics.JetMetricResults
 
JetMetricsContainer - Class in org.apache.beam.runners.jet.metrics
Jet specific implementation of MetricsContainer.
JetMetricsContainer(String, String, Processor.Context) - Constructor for class org.apache.beam.runners.jet.metrics.JetMetricsContainer
 
JetPipelineOptions - Interface in org.apache.beam.runners.jet
Pipeline options specific to the Jet runner.
JetPipelineResult - Class in org.apache.beam.runners.jet
Jet specific implementation of PipelineResult.
JetRunner - Class in org.apache.beam.runners.jet
Jet specific implementation of Beam's PipelineRunner.
JetRunnerRegistrar - Class in org.apache.beam.runners.jet
Contains the PipelineRunnerRegistrar and PipelineOptionsRegistrar for the JetRunner.
JetRunnerRegistrar.Options - Class in org.apache.beam.runners.jet
Registers the JetPipelineOptions.
JetRunnerRegistrar.Runner - Class in org.apache.beam.runners.jet
Registers the JetRunner.
JmsIO - Class in org.apache.beam.sdk.io.jms
An unbounded source for JMS destinations (queues or topics).
JmsIO.MessageMapper<T> - Interface in org.apache.beam.sdk.io.jms
An interface used by JmsIO.Read for converting each jms Message into an element of the resulting PCollection.
JmsIO.Read<T> - Class in org.apache.beam.sdk.io.jms
A PTransform to read from a JMS destination.
JmsIO.Write - Class in org.apache.beam.sdk.io.jms
A PTransform to write to a JMS queue.
JmsRecord - Class in org.apache.beam.sdk.io.jms
JmsRecord contains message payload of the record as well as metadata (JMS headers and properties).
JmsRecord(String, long, String, Destination, Destination, int, boolean, String, long, int, Map<String, Object>, String) - Constructor for class org.apache.beam.sdk.io.jms.JmsRecord
 
JOB_ID - Static variable in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
MRJobConfig.ID.
JobBundleFactory - Interface in org.apache.beam.runners.fnexecution.control
A factory that has all job-scoped information, and can be combined with stage-scoped information to create a StageBundleFactory.
jobId - Variable in class org.apache.beam.runners.dataflow.DataflowPipelineJob
The id for the job.
jobId() - Method in class org.apache.beam.runners.fnexecution.provisioning.JobInfo
 
JobInfo - Class in org.apache.beam.runners.fnexecution.provisioning
A subset of ProvisionApi.ProvisionInfo that specifies a unique job, while omitting fields that are not known to the runner operator.
JobInfo() - Constructor for class org.apache.beam.runners.fnexecution.provisioning.JobInfo
 
JobInvocation - Class in org.apache.beam.runners.fnexecution.jobsubmission
Internal representation of a Job which has been invoked (prepared and run) by a client.
JobInvocation(JobInfo, ListeningExecutorService, RunnerApi.Pipeline, PortablePipelineRunner) - Constructor for class org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation
 
JobInvoker - Class in org.apache.beam.runners.fnexecution.jobsubmission
Factory to create JobInvocation instances.
JobInvoker(String) - Constructor for class org.apache.beam.runners.fnexecution.jobsubmission.JobInvoker
 
jobName() - Method in class org.apache.beam.runners.fnexecution.provisioning.JobInfo
 
JobNameFactory() - Constructor for class org.apache.beam.sdk.options.PipelineOptions.JobNameFactory
 
JobPreparation - Class in org.apache.beam.runners.fnexecution.jobsubmission
A job that has been prepared, but not invoked.
JobPreparation() - Constructor for class org.apache.beam.runners.fnexecution.jobsubmission.JobPreparation
 
JobServerDriver - Class in org.apache.beam.runners.fnexecution.jobsubmission
Shared code for starting and serving an InMemoryJobService.
JobServerDriver(JobServerDriver.ServerConfiguration, ServerFactory, ServerFactory, JobServerDriver.JobInvokerFactory) - Constructor for class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver
 
JobServerDriver.JobInvokerFactory - Interface in org.apache.beam.runners.fnexecution.jobsubmission
 
JobServerDriver.ServerConfiguration - Class in org.apache.beam.runners.fnexecution.jobsubmission
Configuration for the jobServer.
JobSpecification(Job, RunnerApi.Pipeline, Map<AppliedPTransform<?, ?, ?>, String>) - Constructor for class org.apache.beam.runners.dataflow.DataflowPipelineTranslator.JobSpecification
 
jobToString(Job) - Static method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator
Renders a Job as a string.
JodaTimestampConversion() - Constructor for class org.apache.beam.sdk.coders.AvroCoder.JodaTimestampConversion
 
jodaTimeToJava8Time(Instant) - Static method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils
 
Join - Class in org.apache.beam.sdk.extensions.joinlibrary
Utility class with different versions of joins.
Join() - Constructor for class org.apache.beam.sdk.extensions.joinlibrary.Join
 
join(String, CoGroup.By) - Method in class org.apache.beam.sdk.schemas.transforms.CoGroup.ExpandCrossProduct
Select the following fields for the specified PCollection with the specified join args.
join(String, CoGroup.By) - Method in class org.apache.beam.sdk.schemas.transforms.CoGroup.Impl
Select the following fields for the specified PCollection with the specified join args.
join(CoGroup.By) - Static method in class org.apache.beam.sdk.schemas.transforms.CoGroup
Join all input PCollections using the same args.
join(String, CoGroup.By) - Static method in class org.apache.beam.sdk.schemas.transforms.CoGroup
Specify the following join arguments (including fields to join by_ for the specified PCollection.
Join - Class in org.apache.beam.sdk.schemas.transforms
A transform that performs equijoins across two schema PCollections.
Join() - Constructor for class org.apache.beam.sdk.schemas.transforms.Join
 
Join.FieldsEqual - Class in org.apache.beam.sdk.schemas.transforms
Predicate object to specify fields to compare when doing an equi-join.
Join.FieldsEqual.Impl - Class in org.apache.beam.sdk.schemas.transforms
Implementation class for FieldsEqual.
Join.FullOuterJoin<K,V1,V2> - Class in org.apache.beam.sdk.extensions.joinlibrary
PTransform representing a full outer join of two collections of KV elements.
Join.Impl<LhsT,RhsT> - Class in org.apache.beam.sdk.schemas.transforms
Implementation class .
Join.InnerJoin<K,V1,V2> - Class in org.apache.beam.sdk.extensions.joinlibrary
PTransform representing an inner join of two collections of KV elements.
Join.LeftOuterJoin<K,V1,V2> - Class in org.apache.beam.sdk.extensions.joinlibrary
PTransform representing a left outer join of two collections of KV elements.
Join.RightOuterJoin<K,V1,V2> - Class in org.apache.beam.sdk.extensions.joinlibrary
PTransform representing a right outer join of two collections of KV elements.
JoinAsLookup(RexNode, BeamSqlSeekableTable, Schema, Schema, int, int) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamJoinTransforms.JoinAsLookup
 
JoinRelOptRuleCall - Class in org.apache.beam.sdk.extensions.sql.impl.rule
This is a class to catch the built join and check if it is a legal join before passing it to the actual RelOptRuleCall.
JoinRelOptRuleCall.JoinChecker - Interface in org.apache.beam.sdk.extensions.sql.impl.rule
This is a function gets the output relation and checks if it is a legal relational node.
joinStreams(JavaStreamingContext, List<JavaDStream<WindowedValue<T>>>) - Static method in class org.apache.beam.runners.spark.util.SparkCompat
Union of dStreams in the given StreamingContext.
jsonBytesLike(String) - Static method in class org.apache.beam.sdk.testing.JsonMatcher
 
jsonBytesLike(Map<String, Object>) - Static method in class org.apache.beam.sdk.testing.JsonMatcher
 
JsonMatcher<T> - Class in org.apache.beam.sdk.testing
Matcher to compare a string or byte[] representing a JSON Object, independent of field order.
JsonMatcher(Map<String, Object>) - Constructor for class org.apache.beam.sdk.testing.JsonMatcher
 
jsonStringLike(String) - Static method in class org.apache.beam.sdk.testing.JsonMatcher
 
jsonStringLike(Map<String, Object>) - Static method in class org.apache.beam.sdk.testing.JsonMatcher
 
JsonToRow - Class in org.apache.beam.sdk.transforms
Experimental
JsonToRow() - Constructor for class org.apache.beam.sdk.transforms.JsonToRow
 
JvmInitializer - Interface in org.apache.beam.sdk.harness
A service interface for defining one-time initialization of the JVM during pipeline execution.
JvmInitializers - Class in org.apache.beam.sdk.fn
Helpers for executing JvmInitializer implementations.
JvmInitializers() - Constructor for class org.apache.beam.sdk.fn.JvmInitializers
 

K

KafkaCheckpointMark - Class in org.apache.beam.sdk.io.kafka
Checkpoint for a KafkaUnboundedReader.
KafkaCheckpointMark(List<KafkaCheckpointMark.PartitionMark>, Optional<KafkaUnboundedReader<?, ?>>) - Constructor for class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark
 
KafkaCheckpointMark.PartitionMark - Class in org.apache.beam.sdk.io.kafka
A tuple to hold topic, partition, and offset that comprise the checkpoint for a single partition.
KafkaIO - Class in org.apache.beam.sdk.io.kafka
An unbounded source and a sink for Kafka topics.
KafkaIO.Read<K,V> - Class in org.apache.beam.sdk.io.kafka
A PTransform to read from Kafka topics.
KafkaIO.Read.External - Class in org.apache.beam.sdk.io.kafka
Exposes KafkaIO.TypedWithoutMetadata as an external transform for cross-language usage.
KafkaIO.Read.External.Configuration - Class in org.apache.beam.sdk.io.kafka
Parameters class to expose the Read transform to an external SDK.
KafkaIO.TypedWithoutMetadata<K,V> - Class in org.apache.beam.sdk.io.kafka
A PTransform to read from Kafka topics.
KafkaIO.Write<K,V> - Class in org.apache.beam.sdk.io.kafka
A PTransform to write to a Kafka topic with KVs .
KafkaIO.Write.External - Class in org.apache.beam.sdk.io.kafka
Exposes KafkaIO.Write as an external transform for cross-language usage.
KafkaIO.Write.External.Configuration - Class in org.apache.beam.sdk.io.kafka
Parameters class to expose the Write transform to an external SDK.
KafkaIO.WriteRecords<K,V> - Class in org.apache.beam.sdk.io.kafka
A PTransform to write to a Kafka topic with ProducerRecord's.
KafkaPublishTimestampFunction<T> - Interface in org.apache.beam.sdk.io.kafka
An interface for providing custom timestamp for elements written to Kafka.
KafkaRecord<K,V> - Class in org.apache.beam.sdk.io.kafka
KafkaRecord contains key and value of the record as well as metadata for the record (topic name, partition id, and offset).
KafkaRecord(String, int, long, long, KafkaTimestampType, Headers, K, V) - Constructor for class org.apache.beam.sdk.io.kafka.KafkaRecord
 
KafkaRecord(String, int, long, long, KafkaTimestampType, Headers, KV<K, V>) - Constructor for class org.apache.beam.sdk.io.kafka.KafkaRecord
 
KafkaRecordCoder<K,V> - Class in org.apache.beam.sdk.io.kafka
KafkaRecordCoder(Coder<K>, Coder<V>) - Constructor for class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
 
KafkaTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.kafka
Kafka table provider.
KafkaTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.KafkaTableProvider
 
KafkaTimestampType - Enum in org.apache.beam.sdk.io.kafka
This is a copy of Kafka's TimestampType.
KEEP_NESTED_NAME - Static variable in class org.apache.beam.sdk.schemas.utils.SelectHelpers
This policy keeps the raw nested field name.
keepMostNestedFieldName() - Method in class org.apache.beam.sdk.schemas.transforms.Select.Flattened
For nested fields, keep just the most-nested field name.
keyCoder() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.BagUserStateSpec
 
keyCoder() - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.MultimapSideInputHandler
Returns the Coder to use for the elements of the resulting keys iterable.
KeyedByTimestamp(WindowFn<KV<K, V>, BoundedWindow>, TimestampCombiner) - Constructor for class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator.KeyedByTimestamp
 
KeyedPCollectionTuple<K> - Class in org.apache.beam.sdk.transforms.join
An immutable tuple of keyed PCollections with key type K.
KeyedPCollectionTuple.TaggedKeyedPCollection<K,V> - Class in org.apache.beam.sdk.transforms.join
A utility class to help ensure coherence of tag and input PCollection types.
keyedValues() - Static method in class org.apache.beam.sdk.transforms.Deduplicate
Returns a deduplication transform that deduplicates keyed values using the key for up to 10 mins within the processing time domain.
keys() - Method in interface org.apache.beam.sdk.state.MapState
Returns an Iterable over the keys contained in this map.
Keys<K> - Class in org.apache.beam.sdk.transforms
Keys<K> takes a PCollection of KV<K, V>s and returns a PCollection<K> of the keys.
keySet() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
KinesisIO - Class in org.apache.beam.sdk.io.kinesis
PTransforms for reading from and writing to Kinesis streams.
KinesisIO() - Constructor for class org.apache.beam.sdk.io.kinesis.KinesisIO
 
KinesisIO.Read - Class in org.apache.beam.sdk.io.kinesis
Implementation of KinesisIO.read().
KinesisIO.Write - Class in org.apache.beam.sdk.io.kinesis
Implementation of KinesisIO.write().
KinesisPartitioner - Interface in org.apache.beam.sdk.io.kinesis
Kinesis interface for custom partitioner.
KinesisRecord - Class in org.apache.beam.sdk.io.kinesis
UserRecord enhanced with utility methods.
KinesisRecord(UserRecord, String, String) - Constructor for class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
KinesisRecord(ByteBuffer, String, long, String, Instant, Instant, String, String) - Constructor for class org.apache.beam.sdk.io.kinesis.KinesisRecord
 
knownBuilders() - Method in interface org.apache.beam.sdk.expansion.ExternalTransformRegistrar
A mapping from URN to an ExternalTransformBuilder class.
knownBuilders() - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead
 
knownBuilders() - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite
 
knownBuilders() - Method in class org.apache.beam.sdk.io.GenerateSequence.External
 
knownBuilders() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External
 
knownBuilders() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write.External
 
knownTransforms() - Method in interface org.apache.beam.sdk.expansion.service.ExpansionService.ExpansionServiceRegistrar
 
knownTransforms() - Method in class org.apache.beam.sdk.expansion.service.ExpansionService.ExternalTransformRegistrarLoader
 
knownUrns() - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator
 
knownUrns() - Method in interface org.apache.beam.runners.flink.FlinkPortablePipelineTranslator
 
knownUrns() - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator
 
kv(SerializableMatcher<? super K>, SerializableMatcher<? super V>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher that matches any KV with matching key and value.
KV<K,V> - Class in org.apache.beam.sdk.values
An immutable key/value pair.
KV.OrderByKey<K extends java.lang.Comparable<? super K>,V> - Class in org.apache.beam.sdk.values
A Comparator that orders KVs by the natural ordering of their keys.
KV.OrderByValue<K,V extends java.lang.Comparable<? super V>> - Class in org.apache.beam.sdk.values
A Comparator that orders KVs by the natural ordering of their values.
KvCoder<K,V> - Class in org.apache.beam.sdk.coders
A KvCoder encodes KVs.
KVHelpers - Class in org.apache.beam.runners.spark.structuredstreaming.translation.helpers
Helper functions for working with KV.
KVHelpers() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.KVHelpers
 
kvs() - Static method in class org.apache.beam.sdk.transforms.ToString
Transforms each element of the input PCollection to a String by using the Object.toString() on the key followed by a "," followed by the Object.toString() of the value.
kvs(String) - Static method in class org.apache.beam.sdk.transforms.ToString
Transforms each element of the input PCollection to a String by using the Object.toString() on the key followed by the specified delimiter followed by the Object.toString() of the value.
kvs(TypeDescriptor<K>, TypeDescriptor<V>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
KvSwap<K,V> - Class in org.apache.beam.sdk.transforms
KvSwap<K, V> takes a PCollection<KV<K, V>> and returns a PCollection<KV<V, K>>, where all the keys and values have been swapped.
kvWithKey(K) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher that matches any KV with the specified key.
kvWithKey(Coder<K>, K) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher that matches any KV with the specified key.
kvWithKey(SerializableMatcher<? super K>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher that matches any KV with matching key.
kvWithValue(V) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher that matches any KV with the specified value.
kvWithValue(Coder<V>, V) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher that matches any KV with the specified value.
kvWithValue(SerializableMatcher<? super V>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher that matches any KV with matching value.

L

LargeKeys - Interface in org.apache.beam.sdk.testing
Category tags for tests which validate that a Beam runner can handle keys up to a given size.
LargeKeys.Above100KB - Interface in org.apache.beam.sdk.testing
Tests if a runner supports 100KB keys.
LargeKeys.Above100MB - Interface in org.apache.beam.sdk.testing
Tests if a runner supports 100MB keys.
LargeKeys.Above10KB - Interface in org.apache.beam.sdk.testing
Tests if a runner supports 10KB keys.
LargeKeys.Above10MB - Interface in org.apache.beam.sdk.testing
Tests if a runner supports 10MB keys.
LargeKeys.Above1MB - Interface in org.apache.beam.sdk.testing
Tests if a runner supports 1MB keys.
largest(int) - Static method in class org.apache.beam.sdk.transforms.Top
Returns a PTransform that takes an input PCollection<T> and returns a PCollection<List<T>> with a single element containing the largest count elements of the input PCollection<T>, in decreasing order, sorted according to their natural order.
Largest() - Constructor for class org.apache.beam.sdk.transforms.Top.Largest
Deprecated.
 
largestDoublesFn(int) - Static method in class org.apache.beam.sdk.transforms.Top
Returns a Top.TopCombineFn that aggregates the largest count double values.
largestFn(int) - Static method in class org.apache.beam.sdk.transforms.Top
Returns a Top.TopCombineFn that aggregates the largest count values.
largestIntsFn(int) - Static method in class org.apache.beam.sdk.transforms.Top
Returns a Top.TopCombineFn that aggregates the largest count int values.
largestLongsFn(int) - Static method in class org.apache.beam.sdk.transforms.Top
Returns a Top.TopCombineFn that aggregates the largest count long values.
largestPerKey(int) - Static method in class org.apache.beam.sdk.transforms.Top
Returns a PTransform that takes an input PCollection<KV<K, V>> and returns a PCollection<KV<K, List<V>>> that contains an output element mapping each distinct key in the input PCollection to the largest count values associated with that key in the input PCollection<KV<K, V>>, in decreasing order, sorted according to their natural order.
LargestUnique(long) - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn.LargestUnique
Creates a heap to track the largest sampleSize elements.
lastModifiedMillis() - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata
Last modification timestamp in milliseconds since Unix epoch.
Latest - Class in org.apache.beam.sdk.transforms
PTransform and Combine.CombineFn for computing the latest element in a PCollection.
launchApp(StreamingApplication, Properties) - Method in class org.apache.beam.runners.apex.ApexYarnLauncher
 
launchApp(ApexYarnLauncher.LaunchParams) - Method in class org.apache.beam.runners.apex.ApexYarnLauncher
 
LaunchParams(DAG, Attribute.AttributeMap, Properties) - Constructor for class org.apache.beam.runners.apex.ApexYarnLauncher.LaunchParams
 
leaseWorkItem(String, LeaseWorkItemRequest) - Method in class org.apache.beam.runners.dataflow.DataflowClient
Leases the work item for jobId.
leaveCompositeTransform(TransformHierarchy.Node) - Method in class org.apache.beam.runners.gearpump.translators.GearpumpPipelineTranslator
 
leaveCompositeTransform(TransformHierarchy.Node) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
 
leaveCompositeTransform(TransformHierarchy.Node) - Method in class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
 
leaveCompositeTransform(TransformHierarchy.Node) - Method in interface org.apache.beam.sdk.Pipeline.PipelineVisitor
Called for each composite transform after all of its component transforms and their outputs have been visited.
leavePipeline(Pipeline) - Method in class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
 
leavePipeline(Pipeline) - Method in interface org.apache.beam.sdk.Pipeline.PipelineVisitor
Called when all values and transforms in a Pipeline have been visited.
LEFT - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamJoinPushThroughJoinRule
Instance of the rule that works on logical joins only, and pushes to the left.
left(String...) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual.Impl
 
left(Integer...) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual.Impl
 
left(FieldAccessDescriptor) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual.Impl
 
left(String...) - Static method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual
 
left(Integer...) - Static method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual
 
left(FieldAccessDescriptor) - Static method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual
 
leftOuterBroadcastJoin(PCollection<RhsT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Join
Perform a left outer join, broadcasting the right side.
leftOuterJoin(String, PCollection<KV<K, V1>>, PCollection<KV<K, V2>>, V2) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
Left Outer Join of two collections of KV elements.
leftOuterJoin(PCollection<KV<K, V1>>, PCollection<KV<K, V2>>, V2) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
 
leftOuterJoin(PCollection<RhsT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Join
Perform a left outer join.
LegacyArtifactRetrievalService - Interface in org.apache.beam.runners.fnexecution.artifact
An implementation of the Beam Artifact Retrieval Service.
lengthBytes(byte[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
LengthPrefixCoder<T> - Class in org.apache.beam.sdk.coders
A Coder which is able to take any existing coder and wrap it such that it is only invoked in the outer context.
LengthPrefixUnknownCoders - Class in org.apache.beam.runners.fnexecution.wire
Utilities for replacing or wrapping unknown coders with LengthPrefixCoder.
LengthPrefixUnknownCoders() - Constructor for class org.apache.beam.runners.fnexecution.wire.LengthPrefixUnknownCoders
 
lengthString(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
lessThan(T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
lessThan(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
lessThan(T) - Static method in class org.apache.beam.sdk.transforms.Filter
Returns a PTransform that takes an input PCollection and returns a PCollection with elements that are less than a given value, based on the elements' natural ordering.
lessThanEq(T) - Static method in class org.apache.beam.sdk.transforms.Filter
Returns a PTransform that takes an input PCollection<T> and returns a PCollection<T> with elements that are less than or equal to a given value, based on the elements' natural ordering.
lessThanOrEqualTo(T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
lessThanOrEqualTo(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
LHS_TAG - Static variable in class org.apache.beam.sdk.schemas.transforms.Join
 
LIKE - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
like(String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
LIKE_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamBuiltinMethods
 
LimitNumberOfFiles(int) - Constructor for class org.apache.beam.sdk.io.TextRowCountEstimator.LimitNumberOfFiles
 
LimitNumberOfTotalBytes(long) - Constructor for class org.apache.beam.sdk.io.TextRowCountEstimator.LimitNumberOfTotalBytes
 
LinesReadConverter() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTableProvider.LinesReadConverter
 
LinesWriteConverter() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTableProvider.LinesWriteConverter
 
ListCoder<T> - Class in org.apache.beam.sdk.coders
A Coder for List, using the format of IterableLikeCoder.
ListCoder(Coder<T>) - Constructor for class org.apache.beam.sdk.coders.ListCoder
 
listJobMessages(String, String) - Method in class org.apache.beam.runners.dataflow.DataflowClient
Lists job messages with the given jobId.
listJobs(String) - Method in class org.apache.beam.runners.dataflow.DataflowClient
Lists Dataflow Jobs in the project associated with the DataflowPipelineOptions.
listObjects(String, String, String) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
Lists Objects given the bucket, prefix, pageToken.
lists(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for List.
listSubscriptions(PubsubClient.ProjectPath, PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
Return a list of subscriptions for topic in project.
listSubscriptions(PubsubClient.ProjectPath, PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
 
listSubscriptions(PubsubClient.ProjectPath, PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
 
listSubscriptions(PubsubClient.ProjectPath, PubsubClient.TopicPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
listTopics(PubsubClient.ProjectPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
Return a list of topics for project.
listTopics(PubsubClient.ProjectPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
 
listTopics(PubsubClient.ProjectPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
 
listTopics(PubsubClient.ProjectPath) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
listView(PCollection<KV<Void, T>>, PCollectionViews.TypeDescriptorSupplier<T>, WindowingStrategy<?, W>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
Returns a PCollectionView<List<T>> capable of processing elements windowed using the provided WindowingStrategy.
ListViewFn(PCollectionViews.TypeDescriptorSupplier<T>) - Constructor for class org.apache.beam.sdk.values.PCollectionViews.ListViewFn
 
LocalFileSystemRegistrar - Class in org.apache.beam.sdk.io
AutoService registrar for the LocalFileSystem.
LocalFileSystemRegistrar() - Constructor for class org.apache.beam.sdk.io.LocalFileSystemRegistrar
 
LocalResources - Class in org.apache.beam.sdk.io
Helper functions for producing a ResourceId that references a local file or directory.
location(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.Table.Builder
 
log(BeamFnApi.LogEntry) - Method in interface org.apache.beam.runners.fnexecution.logging.LogWriter
Write the contents of the Log Entry to some logging backend.
log(BeamFnApi.LogEntry) - Method in class org.apache.beam.runners.fnexecution.logging.Slf4jLogWriter
 
LogAppendTimePolicy(Optional<Instant>) - Constructor for class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.LogAppendTimePolicy
 
logging(StreamObserver<BeamFnApi.LogControl>) - Method in class org.apache.beam.runners.fnexecution.logging.GrpcLoggingService
 
LoggingHandler() - Constructor for class org.apache.beam.runners.dataflow.util.MonitoringUtil.LoggingHandler
 
LogicalEndpoint - Class in org.apache.beam.sdk.fn.data
A logical endpoint is a pair of an instruction ID corresponding to the BeamFnApi.ProcessBundleRequest and the transform within the processing graph.
LogicalEndpoint() - Constructor for class org.apache.beam.sdk.fn.data.LogicalEndpoint
 
logicalType(Schema.LogicalType<InputT, BaseT>) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
Creates a logical type based on a primitive field type.
LogWriter - Interface in org.apache.beam.runners.fnexecution.logging
A consumer of Beam Log Entries.
longs() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for Long.
longsGlobally() - Static method in class org.apache.beam.sdk.transforms.Max
Returns a PTransform that takes an input PCollection<Long> and returns a PCollection<Long> whose contents is the maximum of the input PCollection's elements, or Long.MIN_VALUE if there are no elements.
longsGlobally() - Static method in class org.apache.beam.sdk.transforms.Min
Returns a PTransform that takes an input PCollection<Long> and returns a PCollection<Long> whose contents is the minimum of the input PCollection's elements, or Long.MAX_VALUE if there are no elements.
longsGlobally() - Static method in class org.apache.beam.sdk.transforms.Sum
Returns a PTransform that takes an input PCollection<Long> and returns a PCollection<Long> whose contents is the sum of the input PCollection's elements, or 0 if there are no elements.
longsPerKey() - Static method in class org.apache.beam.sdk.transforms.Max
Returns a PTransform that takes an input PCollection<KV<K, Long>> and returns a PCollection<KV<K, Long>> that contains an output element mapping each distinct key in the input PCollection to the maximum of the values associated with that key in the input PCollection.
longsPerKey() - Static method in class org.apache.beam.sdk.transforms.Min
Returns a PTransform that takes an input PCollection<KV<K, Long>> and returns a PCollection<KV<K, Long>> that contains an output element mapping each distinct key in the input PCollection to the minimum of the values associated with that key in the input PCollection.
longsPerKey() - Static method in class org.apache.beam.sdk.transforms.Sum
Returns a PTransform that takes an input PCollection<KV<K, Long>> and returns a PCollection<KV<K, Long>> that contains an output element mapping each distinct key in the input PCollection to the sum of the values associated with that key in the input PCollection.
lpad(String, Long) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
lpad(String, Long, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
lpad(byte[], Long) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
lpad(byte[], Long, byte[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
LTRIM - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
ltrim(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
ltrim(String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
LTRIM_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamBuiltinMethods
 

M

main(String[]) - Static method in class org.apache.beam.runners.apex.ApexYarnLauncher
The main method expects the serialized DAG and will launch the YARN application.
main(String[]) - Static method in class org.apache.beam.runners.flink.FlinkJobServerDriver
 
main(String[]) - Static method in class org.apache.beam.runners.flink.FlinkMiniClusterEntryPoint
 
main(String[]) - Static method in class org.apache.beam.runners.flink.FlinkPipelineRunner
Main method to be called only as the entry point to an executable jar with structure as defined in PortablePipelineJarUtils.
main(String[]) - Static method in class org.apache.beam.runners.flink.FlinkPortableClientEntryPoint
Main method to be called standalone or by Flink (CLI or REST API).
main(String[]) - Static method in class org.apache.beam.runners.spark.SparkJobServerDriver
 
main(String[]) - Static method in class org.apache.beam.runners.spark.SparkPipelineRunner
Main method to be called only as the entry point to an executable jar with structure as defined in PortablePipelineJarUtils.
main(String[]) - Static method in class org.apache.beam.runners.spark.structuredstreaming.examples.WordCount
 
main(String[]) - Static method in class org.apache.beam.sdk.expansion.service.ExpansionService
 
main(String[]) - Static method in class org.apache.beam.sdk.extensions.sql.example.BeamSqlDataCatalogExample
 
makeCost(double, double, double) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel.Factory
 
makeCost(double, double) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel.Factory
 
makeHugeCost() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel.Factory
 
makeInfiniteCost() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel.Factory
 
makeTinyCost() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel.Factory
 
makeZeroCost() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel.Factory
 
ManagedChannelFactory - Class in org.apache.beam.sdk.fn.channel
A Factory which creates an underlying ManagedChannel implementation.
ManagedChannelFactory() - Constructor for class org.apache.beam.sdk.fn.channel.ManagedChannelFactory
 
MANIFEST - Static variable in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService
 
MANIFEST - Static variable in class org.apache.beam.runners.fnexecution.artifact.JavaFilesystemLegacyArtifactStagingService
 
Manual(Instant) - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.Manual
 
ManualDockerEnvironmentOptions - Interface in org.apache.beam.sdk.options
Pipeline options to tune DockerEnvironment.
ManualDockerEnvironmentOptions.Options - Class in org.apache.beam.sdk.options
ManualWatermarkEstimator<WatermarkEstimatorStateT> - Interface in org.apache.beam.sdk.transforms.splittabledofn
A WatermarkEstimator which is controlled manually from within a DoFn.
map(WindowedValue<KV<K, V>>) - Method in class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator.KeyedByTimestamp
 
map(TranslatorUtils.RawUnionValue) - Method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils.FromRawUnionValue
 
map(ResultSet) - Method in interface org.apache.beam.sdk.io.cassandra.Mapper
This method is called when reading data from Cassandra.
map(Schema.FieldType, Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
Create a map type for the given key and value types.
map(Schema.FieldType, Schema.FieldType, boolean) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
Deprecated.
Set the nullability on the valueType instead
map() - Static method in class org.apache.beam.sdk.state.StateSpecs
Create a StateSpec for a SetState, optimized for key lookups and writes.
map(Coder<K>, Coder<V>) - Static method in class org.apache.beam.sdk.state.StateSpecs
Identical to StateSpecs.map(), but with key and value coders explicitly supplied.
MAP_TYPES - Static variable in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
MapCoder<K,V> - Class in org.apache.beam.sdk.coders
A Coder for Maps that encodes them according to provided coders for keys and values.
MapControlClientPool - Class in org.apache.beam.runners.fnexecution.control
A ControlClientPool backed by a client map.
MapElements<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
PTransforms for mapping a simple function over the elements of a PCollection.
MapElements.MapWithFailures<InputT,OutputT,FailureT> - Class in org.apache.beam.sdk.transforms
A PTransform that adds exception handling to MapElements.
mapMessage(Message) - Method in interface org.apache.beam.sdk.io.jms.JmsIO.MessageMapper
 
mapOutputs(Map<TupleTag<?>, PValue>, PCollectionTuple) - Method in class org.apache.beam.runners.direct.ParDoMultiOverrideFactory
 
Mapper<T> - Interface in org.apache.beam.sdk.io.cassandra
This interface allows you to implement a custom mapper to read and persist elements from/to Cassandra.
mapQualifier() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
mapQualifier() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierListContext
 
MapQualifierContext(ParserRuleContext, int) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierContext
 
MapQualifierListContext(FieldSpecifierNotationParser.QualifierListContext) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierListContext
 
mapRow(ResultSet) - Method in interface org.apache.beam.sdk.io.jdbc.JdbcIO.RowMapper
 
maps(TypeDescriptor<K>, TypeDescriptor<V>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for Map.
mapSourceFunction(SerializablePipelineOptions, String) - Static method in class org.apache.beam.runners.spark.stateful.StateSpecFunctions
A StateSpec function to support reading from an UnboundedSource.
MapState<K,V> - Interface in org.apache.beam.sdk.state
A ReadableState cell mapping keys to values.
mapView(PCollection<KV<Void, KV<K, V>>>, PCollectionViews.TypeDescriptorSupplier<K>, PCollectionViews.TypeDescriptorSupplier<V>, WindowingStrategy<?, W>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
Returns a PCollectionView<Map<K, V>> capable of processing elements windowed using the provided WindowingStrategy.
MapViewFn(PCollectionViews.TypeDescriptorSupplier<K>, PCollectionViews.TypeDescriptorSupplier<V>) - Constructor for class org.apache.beam.sdk.values.PCollectionViews.MapViewFn
 
markDone() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
Marks this range tracker as being done.
markDone() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
Marks this range tracker as being done.
match() - Static method in class org.apache.beam.sdk.io.FileIO
Matches a filepattern using FileSystems.match(java.util.List<java.lang.String>) and produces a collection of matched resources (both files and directories) as MatchResult.Metadata.
Match() - Constructor for class org.apache.beam.sdk.io.FileIO.Match
 
match(List<String>) - Method in class org.apache.beam.sdk.io.FileSystem
This is the entry point to convert user-provided specs to ResourceIds.
match(List<String>) - Static method in class org.apache.beam.sdk.io.FileSystems
This is the entry point to convert user-provided specs to ResourceIds.
match(List<String>, EmptyMatchTreatment) - Static method in class org.apache.beam.sdk.io.FileSystems
Like FileSystems.match(List), but with a configurable EmptyMatchTreatment.
match(String) - Static method in class org.apache.beam.sdk.io.FileSystems
Like FileSystems.match(List), but for a single resource specification.
match(String, EmptyMatchTreatment) - Static method in class org.apache.beam.sdk.io.FileSystems
Like FileSystems.match(String), but with a configurable EmptyMatchTreatment.
match(StateSpec.Cases<ResultT>) - Method in interface org.apache.beam.sdk.state.StateSpec
For internal use only; no backwards-compatibility guarantees.
matchAll() - Static method in class org.apache.beam.sdk.io.FileIO
Like FileIO.match(), but matches each filepattern in a collection of filepatterns.
MatchAll() - Constructor for class org.apache.beam.sdk.io.FileIO.MatchAll
 
MatchConfiguration() - Constructor for class org.apache.beam.sdk.io.FileIO.MatchConfiguration
 
MatcherAndError() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.CustomHttpErrors.MatcherAndError
 
MatcherCheckerFn(SerializableMatcher<T>) - Constructor for class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert.MatcherCheckerFn
 
matches(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamCalcRule
 
matches(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamCoGBKJoinRule
 
matches(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSideInputJoinRule
 
matches(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSideInputLookupJoinRule
 
matches(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.BeamZetaSqlCalcRule
 
matches(String) - Method in enum org.apache.beam.sdk.io.CompressedSource.CompressionMode
Deprecated.
Returns true if the given file name implies that the contents are compressed according to the compression embodied by this factory.
matches(String) - Method in enum org.apache.beam.sdk.io.Compression
 
matches(String) - Method in enum org.apache.beam.sdk.io.TextIO.CompressionType
Deprecated.
 
matches(String) - Method in enum org.apache.beam.sdk.io.TFRecordIO.CompressionType
Deprecated.
 
matches(String) - Method in enum org.apache.beam.sdk.io.xml.XmlIO.Read.CompressionType
Deprecated.
 
matches(MetricsFilter, MetricKey) - Static method in class org.apache.beam.sdk.metrics.MetricFiltering
Matching logic is implemented here rather than in MetricsFilter because we would like MetricsFilter to act as a "dumb" value-object, with the possibility of replacing it with a Proto/JSON/etc.
matches(Object) - Method in class org.apache.beam.sdk.testing.RegexMatcher
 
matches(String) - Static method in class org.apache.beam.sdk.testing.RegexMatcher
 
matches(Object) - Method in class org.apache.beam.sdk.testing.TestPipelineOptions.AlwaysPassMatcher
 
matches(String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.Matches PTransform that checks if the entire line matches the Regex.
matches(Pattern) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.Matches PTransform that checks if the entire line matches the Regex.
matches(String, int) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.Matches PTransform that checks if the entire line matches the Regex.
matches(Pattern, int) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.Matches PTransform that checks if the entire line matches the Regex.
matches(String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.MatchesName PTransform that checks if the entire line matches the Regex.
matches(Pattern, String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.MatchesName PTransform that checks if the entire line matches the Regex.
Matches(Pattern, int) - Constructor for class org.apache.beam.sdk.transforms.Regex.Matches
 
matchesKV(String, int, int) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.MatchesKV PTransform that checks if the entire line matches the Regex.
matchesKV(Pattern, int, int) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.MatchesKV PTransform that checks if the entire line matches the Regex.
matchesKV(String, String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.MatchesNameKV PTransform that checks if the entire line matches the Regex.
matchesKV(Pattern, String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.MatchesNameKV PTransform that checks if the entire line matches the Regex.
MatchesKV(Pattern, int, int) - Constructor for class org.apache.beam.sdk.transforms.Regex.MatchesKV
 
MatchesName(Pattern, String) - Constructor for class org.apache.beam.sdk.transforms.Regex.MatchesName
 
MatchesNameKV(Pattern, String, String) - Constructor for class org.apache.beam.sdk.transforms.Regex.MatchesNameKV
 
matchesSafely(BigqueryMatcher.TableAndQuery) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher
 
matchesSafely(ShardedFile) - Method in class org.apache.beam.sdk.testing.FileChecksumMatcher
 
matchesSafely(T) - Method in class org.apache.beam.sdk.testing.JsonMatcher
 
matchesScope(String, Set<String>) - Static method in class org.apache.beam.sdk.metrics.MetricFiltering
matchesScope(actualScope, scopes) returns true if the scope of a metric is matched by any of the filters in scopes.
matchNewResource(String, boolean) - Method in class org.apache.beam.sdk.io.FileSystem
Returns a new ResourceId for this filesystem that represents the named resource.
matchNewResource(String, boolean) - Static method in class org.apache.beam.sdk.io.FileSystems
Returns a new ResourceId that represents the named resource of a type corresponding to the resource type.
matchResources(List<ResourceId>) - Static method in class org.apache.beam.sdk.io.FileSystems
Returns MatchResults for the given resourceIds.
MatchResult - Class in org.apache.beam.sdk.io.fs
MatchResult.Metadata - Class in org.apache.beam.sdk.io.fs
MatchResult.Metadata of a matched file.
MatchResult.Metadata.Builder - Class in org.apache.beam.sdk.io.fs
Builder class for MatchResult.Metadata.
MatchResult.Status - Enum in org.apache.beam.sdk.io.fs
Status of a MatchResult.
matchSingleFileSpec(String) - Static method in class org.apache.beam.sdk.io.FileSystems
Returns the MatchResult.Metadata for a single file resource.
Materialization<T> - Interface in org.apache.beam.sdk.transforms
For internal use only; no backwards-compatibility guarantees.
Materializations - Class in org.apache.beam.sdk.transforms
For internal use only; no backwards-compatibility guarantees.
Materializations() - Constructor for class org.apache.beam.sdk.transforms.Materializations
 
Materializations.IterableView<V> - Interface in org.apache.beam.sdk.transforms
Represents the PrimitiveViewT supplied to the ViewFn when it declares to use the iterable materialization.
Materializations.MultimapView<K,V> - Interface in org.apache.beam.sdk.transforms
Represents the PrimitiveViewT supplied to the ViewFn when it declares to use the multimap materialization.
materializedOrAlias() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.Column
 
Max - Class in org.apache.beam.sdk.transforms
PTransforms for computing the maximum of the elements in a PCollection, or the maximum of the values associated with each key in a PCollection of KVs.
MAX_UNIX_MILLIS - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
maxCumulativeBackoff() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
 
MAXIMUM_PRECISION - Static variable in class org.apache.beam.sdk.extensions.zetasketch.HllCount
The maximum precision value you can set in HllCount.Init.Builder.withPrecision(int) is 24.
maximumLookback() - Method in class org.apache.beam.sdk.transforms.windowing.WindowMappingFn
The maximum distance between the end of any main input window mainWindow and the end of the side input window returned by WindowMappingFn.getSideInputWindow(BoundedWindow)
maxInsertBlockSize() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
 
maxRetries() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
 
maxTimestamp() - Method in class org.apache.beam.sdk.fn.windowing.EncodedBoundedWindow
 
maxTimestamp() - Method in class org.apache.beam.sdk.transforms.windowing.BoundedWindow
Returns the inclusive upper bound of timestamps for values in this window.
maxTimestamp() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow
 
maxTimestamp() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
Returns the largest timestamp that can be included in this window.
mayFinish() - Method in class org.apache.beam.sdk.transforms.windowing.AfterEach
 
mayFinish() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
 
mayFinish() - Method in class org.apache.beam.sdk.transforms.windowing.DefaultTrigger
 
mayFinish() - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
 
mayFinish() - Method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
 
mayFinish() - Method in class org.apache.beam.sdk.transforms.windowing.ReshuffleTrigger
Deprecated.
 
mayFinish() - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
For internal use only; no backwards-compatibility guarantees.
mayFinish() - Method in class org.apache.beam.sdk.transforms.windowing.Trigger.OnceTrigger
 
Mean - Class in org.apache.beam.sdk.transforms
PTransforms for computing the arithmetic mean (a.k.a.
merge(Accumulator<MetricsContainerStepMap, MetricsContainerStepMap>) - Method in class org.apache.beam.runners.flink.metrics.MetricsAccumulator
 
merge(NamedAggregators) - Method in class org.apache.beam.runners.spark.aggregators.NamedAggregators
Merges another NamedAggregators instance with this instance.
merge(NamedAggregators.State<InputT, InterT, OutputT>) - Method in interface org.apache.beam.runners.spark.aggregators.NamedAggregators.State
 
merge(AccumulatorV2<NamedAggregators, NamedAggregators>) - Method in class org.apache.beam.runners.spark.aggregators.NamedAggregatorsAccumulator
 
merge(AccumulatorV2<MetricsContainerStepMap, MetricsContainerStepMap>) - Method in class org.apache.beam.runners.spark.metrics.MetricsContainerStepMapAccumulator
 
merge(NamedAggregators) - Method in class org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregators
Merges another NamedAggregators instance with this instance.
merge(NamedAggregators.State<InputT, InterT, OutputT>) - Method in interface org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregators.State
 
merge(AccumulatorV2<NamedAggregators, NamedAggregators>) - Method in class org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregatorsAccumulator
 
merge(AccumulatorV2<MetricsContainerStepMap, MetricsContainerStepMap>) - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsContainerStepMapAccumulator
 
merge(BoundedWindow, Iterable<? extends Instant>) - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
Merges the given timestamps, which may have originated in separate windows, into the context of the result window.
merge(BoundedWindow, Instant...) - Method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
merge(Collection<W>, W) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn.MergeContext
Signals to the framework that the windows in toBeMerged should be merged together to form mergeResult.
mergeAccumulator(AccumT) - Method in interface org.apache.beam.sdk.transforms.Combine.AccumulatingCombineFn.Accumulator
Adds the input values represented by the given accumulator into this accumulator.
mergeAccumulators(Iterable<HyperLogLogPlus>) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.ApproximateDistinctFn
 
mergeAccumulators(Iterable<SketchFrequencies.Sketch<InputT>>) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.CountMinSketchFn
 
mergeAccumulators(Iterable<MergingDigest>) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.TDigestQuantilesFn
 
mergeAccumulators(Iterable<CovarianceAccumulator>) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CovarianceFn
 
mergeAccumulators(Iterable<VarianceAccumulator>) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.VarianceFn
 
mergeAccumulators(Iterable<AccumT>) - Method in interface org.apache.beam.sdk.state.CombiningState
Merge the given accumulators according to the underlying Combine.CombineFn.
mergeAccumulators(Iterable<ApproximateUnique.ApproximateUniqueCombineFn.LargestUnique>) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.ApproximateUniqueCombineFn
 
mergeAccumulators(Iterable<AccumT>) - Method in class org.apache.beam.sdk.transforms.Combine.AccumulatingCombineFn
 
mergeAccumulators(Iterable<double[]>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineDoubleFn
 
mergeAccumulators(Iterable<Combine.Holder<V>>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
 
mergeAccumulators(Iterable<int[]>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineIntegerFn
 
mergeAccumulators(Iterable<long[]>) - Method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineLongFn
 
mergeAccumulators(Iterable<AccumT>) - Method in class org.apache.beam.sdk.transforms.Combine.CombineFn
Returns an accumulator representing the accumulation of all the input values accumulated in the merging accumulators.
mergeAccumulators(Iterable<List<V>>) - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
 
mergeAccumulators(Iterable<Object[]>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
 
mergeAccumulators(Iterable<Object[]>, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
 
mergeAccumulators(Iterable<AccumT>, CombineWithContext.Context) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.CombineFnWithContext
Returns an accumulator representing the accumulation of all the input values accumulated in the merging accumulators.
mergeAccumulators(Iterable<Top.BoundedHeap<KV<Integer, T>, SerializableComparator<KV<Integer, T>>>>) - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
 
MergeContext() - Constructor for class org.apache.beam.sdk.transforms.windowing.WindowFn.MergeContext
 
MergeOverlappingIntervalWindows - Class in org.apache.beam.sdk.transforms.windowing
For internal use only; no backwards compatibility guarantees.
MergeOverlappingIntervalWindows() - Constructor for class org.apache.beam.sdk.transforms.windowing.MergeOverlappingIntervalWindows
 
mergeWideningNullable(Schema, Schema) - Static method in class org.apache.beam.sdk.schemas.SchemaUtils
Given two schema that have matching types, return a nullable-widened schema.
mergeWindows(WindowFn<Object, W>.MergeContext) - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
 
mergeWindows(WindowFn<?, IntervalWindow>.MergeContext) - Static method in class org.apache.beam.sdk.transforms.windowing.MergeOverlappingIntervalWindows
Merge overlapping IntervalWindows.
mergeWindows(WindowFn<T, W>.MergeContext) - Method in class org.apache.beam.sdk.transforms.windowing.NonMergingWindowFn
 
mergeWindows(WindowFn<Object, IntervalWindow>.MergeContext) - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
 
mergeWindows(WindowFn<T, W>.MergeContext) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
Does whatever merging of windows is necessary.
message() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.OutgoingMessage
Underlying Message.
message() - Method in class org.apache.beam.sdk.schemas.transforms.Cast.CompatibilityError
 
MessageCoder - Class in org.apache.beam.sdk.io.aws2.sqs
Custom Coder for handling SendMessageRequest for using in Write.
MessageCoderRegistrar - Class in org.apache.beam.sdk.io.aws2.sqs
A CoderProviderRegistrar for standard types used with SqsIO.
MessageCoderRegistrar() - Constructor for class org.apache.beam.sdk.io.aws2.sqs.MessageCoderRegistrar
 
messageSchema() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubMessageToRow
Schema of the Pubsub message.
Metadata(long, Instant, Instant, long, MetricsContainerStepMap) - Constructor for class org.apache.beam.runners.spark.io.SparkUnboundedSource.Metadata
 
metadata() - Method in class org.apache.beam.sdk.io.fs.MatchResult
MatchResult.Metadata of matched files.
Metadata() - Constructor for class org.apache.beam.sdk.io.fs.MatchResult.Metadata
 
MetadataCoder - Class in org.apache.beam.sdk.io.fs
MetadataCoderV2 - Class in org.apache.beam.sdk.io.fs
metaStore(MetaStore) - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlCli
 
metaStore(MetaStore, boolean, PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.BeamSqlCli
 
MetaStore - Interface in org.apache.beam.sdk.extensions.sql.meta.store
The interface to handle CRUD of BeamSql table metadata.
METHOD - Static variable in interface org.apache.beam.sdk.extensions.sql.impl.planner.NodeStatsMetadata
 
method - Variable in class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase
Method that implements the function.
Metric - Interface in org.apache.beam.sdk.metrics
Marker interface for all user-facing metrics.
MetricFiltering - Class in org.apache.beam.sdk.metrics
Implements matching for metrics filters.
MetricKey - Class in org.apache.beam.sdk.metrics
Metrics are keyed by the step name they are associated with and the name of the metric.
MetricKey() - Constructor for class org.apache.beam.sdk.metrics.MetricKey
 
metricName() - Method in class org.apache.beam.sdk.metrics.MetricKey
The name of the metric.
MetricName - Class in org.apache.beam.sdk.metrics
The name of a metric consists of a MetricName.getNamespace() and a MetricName.getName().
MetricName() - Constructor for class org.apache.beam.sdk.metrics.MetricName
 
MetricNameFilter - Class in org.apache.beam.sdk.metrics
The name of a metric.
MetricNameFilter() - Constructor for class org.apache.beam.sdk.metrics.MetricNameFilter
 
MetricQueryResults - Class in org.apache.beam.sdk.metrics
The results of a query for metrics.
MetricQueryResults() - Constructor for class org.apache.beam.sdk.metrics.MetricQueryResults
 
metricRegistry() - Method in class org.apache.beam.runners.spark.metrics.AggregatorMetricSource
 
metricRegistry() - Method in class org.apache.beam.runners.spark.metrics.CompositeSource
 
metricRegistry() - Method in class org.apache.beam.runners.spark.metrics.SparkBeamMetricSource
 
metricRegistry() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.AggregatorMetricSource
 
metricRegistry() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.CompositeSource
 
metricRegistry() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.SparkBeamMetricSource
 
MetricResult<T> - Class in org.apache.beam.sdk.metrics
The results of a single current metric.
MetricResult() - Constructor for class org.apache.beam.sdk.metrics.MetricResult
 
MetricResults - Class in org.apache.beam.sdk.metrics
Methods for interacting with the metrics of a pipeline that has been executed.
MetricResults() - Constructor for class org.apache.beam.sdk.metrics.MetricResults
 
metrics() - Method in class org.apache.beam.runners.apex.ApexRunnerResult
 
metrics() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
 
metrics() - Method in class org.apache.beam.runners.direct.DirectRunner.DirectPipelineResult
 
metrics() - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
 
metrics() - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
 
Metrics - Class in org.apache.beam.runners.flink.metrics
Helper for pretty-printing Flink metrics.
Metrics() - Constructor for class org.apache.beam.runners.flink.metrics.Metrics
 
metrics() - Method in class org.apache.beam.runners.gearpump.GearpumpPipelineResult
 
metrics() - Method in class org.apache.beam.runners.jet.FailedRunningPipelineResults
 
metrics() - Method in class org.apache.beam.runners.jet.JetPipelineResult
 
metrics() - Method in class org.apache.beam.runners.spark.SparkPipelineResult
 
metrics() - Method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineResult
 
Metrics - Class in org.apache.beam.sdk.metrics
The Metrics is a utility class for producing various kinds of metrics for reporting properties of an executing pipeline.
metrics() - Method in interface org.apache.beam.sdk.PipelineResult
Returns the object to access metrics from the pipeline.
MetricsAccumulator - Class in org.apache.beam.runners.flink.metrics
Accumulator of MetricsContainerStepMap.
MetricsAccumulator() - Constructor for class org.apache.beam.runners.flink.metrics.MetricsAccumulator
 
MetricsAccumulator - Class in org.apache.beam.runners.spark.metrics
For resilience, Accumulators are required to be wrapped in a Singleton.
MetricsAccumulator() - Constructor for class org.apache.beam.runners.spark.metrics.MetricsAccumulator
 
MetricsAccumulator - Class in org.apache.beam.runners.spark.structuredstreaming.metrics
For resilience, Accumulators are required to be wrapped in a Singleton.
MetricsAccumulator() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsAccumulator
 
MetricsAccumulator.AccumulatorCheckpointingSparkListener - Class in org.apache.beam.runners.spark.metrics
Spark Listener which checkpoints MetricsContainerStepMap values for fault-tolerance.
MetricsContainer - Interface in org.apache.beam.sdk.metrics
Holds the metrics for a single step.
MetricsContainerStepMapAccumulator - Class in org.apache.beam.runners.spark.metrics
AccumulatorV2 implementation for MetricsContainerStepMap.
MetricsContainerStepMapAccumulator(MetricsContainerStepMap) - Constructor for class org.apache.beam.runners.spark.metrics.MetricsContainerStepMapAccumulator
 
MetricsContainerStepMapAccumulator - Class in org.apache.beam.runners.spark.structuredstreaming.metrics
AccumulatorV2 implementation for MetricsContainerStepMap.
MetricsContainerStepMapAccumulator(MetricsContainerStepMap) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsContainerStepMapAccumulator
 
MetricsEnvironment - Class in org.apache.beam.sdk.metrics
Manages and provides the metrics container associated with each thread.
MetricsEnvironment() - Constructor for class org.apache.beam.sdk.metrics.MetricsEnvironment
 
MetricsFilter - Class in org.apache.beam.sdk.metrics
Simple POJO representing a filter for querying metrics.
MetricsFilter() - Constructor for class org.apache.beam.sdk.metrics.MetricsFilter
 
MetricsFilter.Builder - Class in org.apache.beam.sdk.metrics
Builder for creating a MetricsFilter.
MetricsOptions - Interface in org.apache.beam.sdk.metrics
Extension of PipelineOptions that defines MetricsSink specific options.
MetricsOptions.NoOpMetricsSink - Class in org.apache.beam.sdk.metrics
A DefaultValueFactory that obtains the class of the NoOpMetricsSink if it exists on the classpath, and throws an exception otherwise.
MetricsSink - Interface in org.apache.beam.sdk.metrics
Interface for all metric sinks.
MicrobatchSource<T,CheckpointMarkT extends UnboundedSource.CheckpointMark> - Class in org.apache.beam.runners.spark.io
A Source that accommodates Spark's micro-batch oriented nature and wraps an UnboundedSource.
MicrobatchSource.Reader - Class in org.apache.beam.runners.spark.io
Mostly based on BoundedReadFromUnboundedSource's UnboundedToBoundedSourceAdapter, with some adjustments for Spark specifics.
MILLIS_PER_DAY - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
mimeType() - Method in class org.apache.beam.sdk.io.fs.CreateOptions
The file-like resource mime type.
Min - Class in org.apache.beam.sdk.transforms
PTransforms for computing the minimum of the elements in a PCollection, or the minimum of the values associated with each key in a PCollection of KVs.
MIN_UNIX_MILLIS - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
MINIMUM_PRECISION - Static variable in class org.apache.beam.sdk.extensions.zetasketch.HllCount
The minimum precision value you can set in HllCount.Init.Builder.withPrecision(int) is 10.
MINIMUM_UPLOAD_BUFFER_SIZE_BYTES - Static variable in class org.apache.beam.sdk.io.aws.options.S3Options.S3UploadBufferSizeBytesFactory
 
minus(RelOptCost) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
minus(NodeStats) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
 
modeNames - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
modifyAckDeadline(PubsubClient.SubscriptionPath, List<String>, int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
Modify the ack deadline for messages from subscription with ackIds to be deadlineSeconds from now.
modifyAckDeadline(PubsubClient.SubscriptionPath, List<String>, int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
 
modifyAckDeadline(PubsubClient.SubscriptionPath, List<String>, int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
 
modifyAckDeadline(PubsubClient.SubscriptionPath, List<String>, int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
modifyEnvironmentBeforeSubmission(Environment) - Method in class org.apache.beam.runners.dataflow.DataflowRunnerHooks
Allows the user to modify the environment of their job before their job is submitted to the service for execution.
MongoDbGridFSIO - Class in org.apache.beam.sdk.io.mongodb
IO to read and write data on MongoDB GridFS.
MongoDbGridFSIO() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO
 
MongoDbGridFSIO.ConnectionConfiguration - Class in org.apache.beam.sdk.io.mongodb
Encapsulate the MongoDB GridFS connection logic.
MongoDbGridFSIO.Parser<T> - Interface in org.apache.beam.sdk.io.mongodb
Interface for the parser that is used to parse the GridFSDBFile into the appropriate types.
MongoDbGridFSIO.ParserCallback<T> - Interface in org.apache.beam.sdk.io.mongodb
Callback for the parser to use to submit data.
MongoDbGridFSIO.Read<T> - Class in org.apache.beam.sdk.io.mongodb
A PTransform to read data from MongoDB GridFS.
MongoDbGridFSIO.Read.BoundedGridFSSource - Class in org.apache.beam.sdk.io.mongodb
A BoundedSource for MongoDB GridFS.
MongoDbGridFSIO.Write<T> - Class in org.apache.beam.sdk.io.mongodb
A PTransform to write data to MongoDB GridFS.
MongoDbGridFSIO.WriteFn<T> - Interface in org.apache.beam.sdk.io.mongodb
Function that is called to write the data to the give GridFS OutputStream.
MongoDbIO - Class in org.apache.beam.sdk.io.mongodb
IO to read and write data on MongoDB.
MongoDbIO.Read - Class in org.apache.beam.sdk.io.mongodb
A PTransform to read data from MongoDB.
MongoDbIO.Write - Class in org.apache.beam.sdk.io.mongodb
A PTransform to write to a MongoDB database.
MongoDbTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.mongodb
 
MongoDbTable.DocumentToRow - Class in org.apache.beam.sdk.extensions.sql.meta.provider.mongodb
 
MongoDbTable.RowToDocument - Class in org.apache.beam.sdk.extensions.sql.meta.provider.mongodb
 
MongoDbTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.mongodb
MongoDbTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTableProvider
 
MonitoringUtil - Class in org.apache.beam.runners.dataflow.util
A helper class for monitoring jobs submitted to the service.
MonitoringUtil(DataflowClient) - Constructor for class org.apache.beam.runners.dataflow.util.MonitoringUtil
Construct a helper for monitoring.
MonitoringUtil.JobMessagesHandler - Interface in org.apache.beam.runners.dataflow.util
An interface that can be used for defining callbacks to receive a list of JobMessages containing monitoring information.
MonitoringUtil.LoggingHandler - Class in org.apache.beam.runners.dataflow.util
A handler that logs monitoring messages.
MonitoringUtil.TimeStampComparator - Class in org.apache.beam.runners.dataflow.util
Comparator for sorting rows in increasing order based on timestamp.
MonotonicallyIncreasing(Instant) - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.MonotonicallyIncreasing
 
months(int) - Static method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows
Returns a WindowFn that windows elements into periods measured by months.
MoveOptions - Interface in org.apache.beam.sdk.io.fs
MoveOptions.StandardMoveOptions - Enum in org.apache.beam.sdk.io.fs
Defines the standard MoveOptions.
MqttIO - Class in org.apache.beam.sdk.io.mqtt
An unbounded source for MQTT broker.
MqttIO.ConnectionConfiguration - Class in org.apache.beam.sdk.io.mqtt
A POJO describing a MQTT connection.
MqttIO.Read - Class in org.apache.beam.sdk.io.mqtt
A PTransform to read from a MQTT broker.
MqttIO.Write - Class in org.apache.beam.sdk.io.mqtt
A PTransform to write and send a message to a MQTT server.
multimap() - Static method in class org.apache.beam.sdk.transforms.Materializations
For internal use only; no backwards-compatibility guarantees.
MULTIMAP_MATERIALIZATION_URN - Static variable in class org.apache.beam.sdk.transforms.Materializations
The URN for a Materialization where the primitive view type is a multimap of fully specified windowed values.
multimapView(PCollection<KV<Void, KV<K, V>>>, PCollectionViews.TypeDescriptorSupplier<K>, PCollectionViews.TypeDescriptorSupplier<V>, WindowingStrategy<?, W>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
Returns a PCollectionView<Map<K, Iterable<V>>> capable of processing elements windowed using the provided WindowingStrategy.
MultimapViewFn(PCollectionViews.TypeDescriptorSupplier<K>, PCollectionViews.TypeDescriptorSupplier<V>) - Constructor for class org.apache.beam.sdk.values.PCollectionViews.MultimapViewFn
 
MultiOuputCoder<T> - Class in org.apache.beam.runners.spark.structuredstreaming.translation.helpers
Coder to serialize and deserialize Tuple2, WindowedValue{/@code} to be used in spark encoders while applying DoFn.
multiOutputOverrideFactory(DataflowPipelineOptions) - Static method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides
Returns a PTransformOverrideFactory that replaces a multi-output ParDo with a composite transform specialized for the DataflowRunner.
multiply(double) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
 
multiplyBy(double) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
MutationGroup - Class in org.apache.beam.sdk.io.gcp.spanner
A bundle of mutations that must be submitted atomically.

N

name(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.Table.Builder
 
name() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.Column
 
name - Variable in enum org.apache.beam.sdk.io.kafka.KafkaTimestampType
 
name() - Method in class org.apache.beam.sdk.metrics.MetricName
Deprecated.
to be removed once Dataflow no longer requires this method.
name() - Method in interface org.apache.beam.sdk.schemas.FieldValueGetter
 
name() - Method in interface org.apache.beam.sdk.schemas.FieldValueSetter
Returns the name of the field.
name(TypeDescription) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.InjectPackageStrategy
 
name - Variable in class org.apache.beam.sdk.transforms.PTransform
The base name of this PTransform, e.g., from defaults, or null if not yet assigned.
named() - Method in class org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters
 
named(String, String) - Static method in class org.apache.beam.sdk.metrics.MetricName
 
named(Class<?>, String) - Static method in class org.apache.beam.sdk.metrics.MetricName
 
named(String, String) - Static method in class org.apache.beam.sdk.metrics.MetricNameFilter
 
named(Class<?>, String) - Static method in class org.apache.beam.sdk.metrics.MetricNameFilter
 
NamedAggregators - Class in org.apache.beam.runners.spark.aggregators
This class wraps a map of named aggregators.
NamedAggregators() - Constructor for class org.apache.beam.runners.spark.aggregators.NamedAggregators
Constructs a new NamedAggregators instance.
NamedAggregators - Class in org.apache.beam.runners.spark.structuredstreaming.aggregators
This class wraps a map of named aggregators.
NamedAggregators() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregators
Constructs a new NamedAggregators instance.
NamedAggregators.State<InputT,InterT,OutputT> - Interface in org.apache.beam.runners.spark.aggregators
 
NamedAggregators.State<InputT,InterT,OutputT> - Interface in org.apache.beam.runners.spark.structuredstreaming.aggregators
 
NamedAggregatorsAccumulator - Class in org.apache.beam.runners.spark.aggregators
AccumulatorV2 implementation for NamedAggregators.
NamedAggregatorsAccumulator(NamedAggregators) - Constructor for class org.apache.beam.runners.spark.aggregators.NamedAggregatorsAccumulator
 
NamedAggregatorsAccumulator - Class in org.apache.beam.runners.spark.structuredstreaming.aggregators
AccumulatorV2 implementation for NamedAggregators.
NamedAggregatorsAccumulator(NamedAggregators) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregatorsAccumulator
 
nameOf(int) - Method in class org.apache.beam.sdk.schemas.Schema
Return the name of field by index.
names() - Method in class org.apache.beam.sdk.metrics.MetricsFilter
 
namespace() - Method in class org.apache.beam.sdk.metrics.MetricName
Deprecated.
to be removed once Dataflow no longer requires this method.
NanosDuration - Class in org.apache.beam.sdk.schemas.logicaltypes
A duration represented in nanoseconds.
NanosDuration() - Constructor for class org.apache.beam.sdk.schemas.logicaltypes.NanosDuration
 
NanosInstant - Class in org.apache.beam.sdk.schemas.logicaltypes
A timestamp represented as nanoseconds since the epoch.
NanosInstant() - Constructor for class org.apache.beam.sdk.schemas.logicaltypes.NanosInstant
 
narrowing(Schema) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
 
Narrowing() - Constructor for class org.apache.beam.sdk.schemas.transforms.Cast.Narrowing
 
nativeSQL(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
Natural() - Constructor for class org.apache.beam.sdk.transforms.Top.Natural
 
naturalOrder(T) - Static method in class org.apache.beam.sdk.transforms.Max
 
naturalOrder() - Static method in class org.apache.beam.sdk.transforms.Max
 
naturalOrder(T) - Static method in class org.apache.beam.sdk.transforms.Min
 
naturalOrder() - Static method in class org.apache.beam.sdk.transforms.Min
 
NeedsDocker - Interface in org.apache.beam.runners.fnexecution.environment.testing
Category for integration tests that require Docker.
NeedsRunner - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilize TestPipeline for execution and expect to be executed by a PipelineRunner.
NESTED - Static variable in class org.apache.beam.sdk.coders.Coder.Context
Deprecated.
The nested context: the value being encoded or decoded is (potentially) a part of a larger record/stream contents, and may have other parts encoded or decoded after it.
nested() - Method in class org.apache.beam.sdk.coders.Coder.Context
Deprecated.
 
nestedFieldsById() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return the nested fields keyed by field ids.
nestedFieldsByName() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return the nested fields keyed by field name.
never() - Static method in class org.apache.beam.sdk.transforms.Watch.Growth
Returns a Watch.Growth.TerminationCondition that never holds (i.e., poll each input until its output is complete).
Never - Class in org.apache.beam.sdk.transforms.windowing
A Trigger which never fires.
Never() - Constructor for class org.apache.beam.sdk.transforms.windowing.Never
 
Never.NeverTrigger - Class in org.apache.beam.sdk.transforms.windowing
The actual trigger class for Never triggers.
neverRetry() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy
Never retry any failures.
newBundle(Map<String, RemoteOutputReceiver<?>>, BundleProgressHandler) - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor
Start a new bundle for the given BeamFnApi.ProcessBundleDescriptor identifier.
newBundle(Map<String, RemoteOutputReceiver<?>>, StateRequestHandler, BundleProgressHandler) - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor
Start a new bundle for the given BeamFnApi.ProcessBundleDescriptor identifier.
newBundle(Map<String, RemoteOutputReceiver<?>>, Map<KV<String, String>, RemoteOutputReceiver<Timer<?>>>, StateRequestHandler, BundleProgressHandler) - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor
Start a new bundle for the given BeamFnApi.ProcessBundleDescriptor identifier.
newBundle(Map<String, RemoteOutputReceiver<?>>, Map<KV<String, String>, RemoteOutputReceiver<Timer<?>>>, StateRequestHandler, BundleProgressHandler, BundleSplitHandler, BundleCheckpointHandler, BundleFinalizationHandler) - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor
Start a new bundle for the given BeamFnApi.ProcessBundleDescriptor identifier.
newClient(String, String, PubsubOptions) - Method in interface org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.PubsubClientFactory
Construct a new Pubsub client.
newClouddebuggerClient(DataflowPipelineOptions) - Static method in class org.apache.beam.runners.dataflow.util.DataflowTransport
 
newConfiguration(SerializableConfiguration) - Static method in class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
Returns new populated Configuration object.
newConnection(UnregisteredDriver, AvaticaFactory, String, Properties, CalciteSchema, JavaTypeFactory) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteFactoryWrapper
 
newDatabaseMetaData(AvaticaConnection) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteFactoryWrapper
 
newDataflowClient(DataflowPipelineOptions) - Static method in class org.apache.beam.runners.dataflow.util.DataflowTransport
Returns a Google Cloud Dataflow client builder.
newJob(SerializableConfiguration) - Static method in class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
Returns new configured Job object.
newPopulation(Schema.TypeName) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CovarianceFn
 
newPopulation(SerializableFunction<BigDecimal, V>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CovarianceFn
 
newPopulation(Schema.TypeName) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.VarianceFn
 
newPopulation(SerializableFunction<BigDecimal, V>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.VarianceFn
 
newPreparedStatement(AvaticaConnection, Meta.StatementHandle, Meta.Signature, int, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteFactoryWrapper
 
newProvider(T) - Method in class org.apache.beam.sdk.testing.TestPipeline
Returns a new ValueProvider that is inaccessible before TestPipeline.run(), but will be accessible while the pipeline runs.
newResultSet(AvaticaStatement, QueryState, Meta.Signature, TimeZone, Meta.Frame) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteFactoryWrapper
 
newResultSetMetaData(AvaticaStatement, Meta.Signature) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteFactoryWrapper
 
newSample(Schema.TypeName) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CovarianceFn
 
newSample(SerializableFunction<BigDecimal, V>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.CovarianceFn
 
newSample(Schema.TypeName) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.VarianceFn
 
newSample(SerializableFunction<BigDecimal, V>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.transform.agg.VarianceFn
 
newStatement(AvaticaConnection, Meta.StatementHandle, int, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteFactoryWrapper
 
newStorageClient(GcsOptions) - Static method in class org.apache.beam.sdk.extensions.gcp.util.Transport
Returns a Cloud Storage client builder using the specified GcsOptions.
newTracker() - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
 
newTracker() - Method in class org.apache.beam.sdk.io.range.OffsetRange
 
newTracker() - Method in interface org.apache.beam.sdk.transforms.splittabledofn.HasDefaultTracker
Creates a new tracker for this.
newWatermarkEstimator() - Method in interface org.apache.beam.sdk.transforms.splittabledofn.HasDefaultWatermarkEstimator
Creates a new watermark estimator for this.
next() - Method in class org.apache.beam.runners.jet.processors.BoundedSourceP
 
next() - Method in class org.apache.beam.sdk.fn.stream.DataStreams.BlockingQueueIterator
 
next() - Method in class org.apache.beam.sdk.fn.stream.DataStreams.DataStreamDecoder
 
nextBatch(TimestampedValue<T>...) - Method in class org.apache.beam.runners.spark.io.CreateStream
Enqueue next micro-batch elements.
nextBatch(T...) - Method in class org.apache.beam.runners.spark.io.CreateStream
For non-timestamped elements.
nextFieldId() - Method in class org.apache.beam.sdk.values.Row.Builder
 
NO_ARTIFACTS_STAGED_TOKEN - Static variable in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactStagingService
 
NO_FIRING - Static variable in class org.apache.beam.sdk.transforms.windowing.PaneInfo
PaneInfo to use for elements on (and before) initial window assignment (including elements read from sources) before they have passed through a GroupByKey and are associated with a particular trigger firing.
NodeStats - Class in org.apache.beam.sdk.extensions.sql.impl.planner
This is a utility class to represent rowCount, rate and window.
NodeStats() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
 
NodeStatsMetadata - Interface in org.apache.beam.sdk.extensions.sql.impl.planner
This is a metadata used for row count and rate estimation.
NodeStatsMetadata.Handler - Interface in org.apache.beam.sdk.extensions.sql.impl.planner
Handler API.
NON_PARALLEL_INPUTS - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
NonCumulativeCostImpl() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner.NonCumulativeCostImpl
 
NonDeterministicException(Coder<?>, String, Coder.NonDeterministicException) - Constructor for exception org.apache.beam.sdk.coders.Coder.NonDeterministicException
 
NonDeterministicException(Coder<?>, String) - Constructor for exception org.apache.beam.sdk.coders.Coder.NonDeterministicException
 
NonDeterministicException(Coder<?>, List<String>) - Constructor for exception org.apache.beam.sdk.coders.Coder.NonDeterministicException
 
NonDeterministicException(Coder<?>, List<String>, Coder.NonDeterministicException) - Constructor for exception org.apache.beam.sdk.coders.Coder.NonDeterministicException
 
none() - Static method in class org.apache.beam.sdk.schemas.Schema.Options
 
none() - Static method in class org.apache.beam.sdk.transforms.display.DisplayData
Default empty DisplayData instance.
NonMergingWindowFn<T,W extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
Abstract base class for WindowFns that do not merge windows.
NonMergingWindowFn() - Constructor for class org.apache.beam.sdk.transforms.windowing.NonMergingWindowFn
 
nonSeekableInputIndex() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
 
NOOP_CHECKPOINT_MARK - Static variable in interface org.apache.beam.sdk.io.UnboundedSource.CheckpointMark
 
NoopCheckpointMark() - Constructor for class org.apache.beam.sdk.io.UnboundedSource.CheckpointMark.NoopCheckpointMark
 
NoopCredentialFactory - Class in org.apache.beam.sdk.extensions.gcp.auth
Construct an oauth credential to be used by the SDK and the SDK workers.
NoopCredentialFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.auth.NoopCredentialFactory
 
NoOpMetricsSink() - Constructor for class org.apache.beam.sdk.metrics.MetricsOptions.NoOpMetricsSink
 
NoopPathValidator - Class in org.apache.beam.sdk.extensions.gcp.storage
For internal use only; no backwards compatibility guarantees.
NoOpStepContext - Class in org.apache.beam.runners.gearpump.translators.utils
serializable StepContext that basically does nothing.
NoOpStepContext() - Constructor for class org.apache.beam.runners.gearpump.translators.utils.NoOpStepContext
 
NoOpStepContext - Class in org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions
A StepContext for Spark Batch Runner execution.
NoOpStepContext() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.NoOpStepContext
 
normalize() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
NoSuchSchemaException - Exception in org.apache.beam.sdk.schemas
Indicates that we are missing a schema for a type.
NoSuchSchemaException() - Constructor for exception org.apache.beam.sdk.schemas.NoSuchSchemaException
 
not(SerializableMatcher<T>) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
notEqualTo(T) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
Asserts that the value in question is not equal to the provided value, according to Object.equals(java.lang.Object).
notifyOfRemovedMetric(Metric, String, MetricGroup) - Method in class org.apache.beam.runners.flink.metrics.FileReporter
 
now() - Method in interface org.apache.beam.runners.direct.Clock
Returns the current time as an Instant.
now(Matcher<Iterable<? extends Row>>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery.RowsAssertion
 
nullable() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
nullable(TableSchema.TypeName) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
nullable(String, Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.Schema.Field
Return's a nullable field with the give name and type.
NULLABLE_DATE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
NULLABLE_TIME - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
NULLABLE_TIMESTAMP - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
NullableCoder<T> - Class in org.apache.beam.sdk.coders
A NullableCoder encodes nullable values of type T using a nested Coder<T> that does not tolerate null values.
nullContext() - Static method in class org.apache.beam.sdk.state.StateContexts
Returns a fake StateContext.
NullCredentialInitializer - Class in org.apache.beam.sdk.extensions.gcp.auth
A HttpRequestInitializer for requests that don't have credentials.
NullCredentialInitializer() - Constructor for class org.apache.beam.sdk.extensions.gcp.auth.NullCredentialInitializer
 
nullRow(Schema) - Static method in class org.apache.beam.sdk.values.Row
Creates a new record filled with nulls.
nulls() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for nulls/Void.
nullValue() - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
A SerializableMatcher with identical criteria to Matchers.nullValue().
NUM_QUERY_SPLITS_MAX - Static variable in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
An upper bound on the number of splits for a query.
NUM_REDUCES - Static variable in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
MRJobConfig.NUM_REDUCES.
numberOfRecordsForRate - Variable in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
 
NUMERIC_TYPES - Static variable in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
numSupported() - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTableFilter
This is primarily used by the cost based optimization to determine the benefit of performing predicate push-down for an IOSourceRel.
numSupported() - Method in class org.apache.beam.sdk.extensions.sql.meta.DefaultTableFilter
 
numSupported() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryFilter
 
numSupported() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableFilter
 

O

OBJECT_MAPPER - Static variable in class org.apache.beam.sdk.extensions.sql.impl.JdbcDriver
 
OBJECT_TYPE_NAME - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
observe(RestrictionTracker<RestrictionT, PositionT>, RestrictionTrackers.ClaimObserver<PositionT>) - Static method in class org.apache.beam.sdk.fn.splittabledofn.RestrictionTrackers
Returns a thread safe RestrictionTracker which reports all claim attempts to the specified RestrictionTrackers.ClaimObserver.
observeTimestamp(Instant) - Method in interface org.apache.beam.sdk.transforms.splittabledofn.TimestampObservingWatermarkEstimator
Update watermark estimate with latest output timestamp.
observeTimestamp(Instant) - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.MonotonicallyIncreasing
 
of(PCollectionView<ViewT>) - Static method in class org.apache.beam.runners.apex.ApexRunner.CreateApexPCollectionView
 
of() - Static method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
 
of(BeamFnApi.ProcessBundleDescriptor, List<RemoteInputDestination>, Map<String, Coder>, Map<String, Map<String, ProcessBundleDescriptors.SideInputSpec>>, Map<String, Map<String, ProcessBundleDescriptors.BagUserStateSpec>>, Map<String, Map<String, ProcessBundleDescriptors.TimerSpec>>) - Static method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.ExecutableProcessBundleDescriptor
 
of(String, String, RunnerApi.FunctionSpec, Coder<T>, Coder<W>) - Static method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.SideInputSpec
 
of(Coder<T>, FnDataReceiver<T>) - Static method in class org.apache.beam.runners.fnexecution.control.RemoteOutputReceiver
 
of(Coder<T>, String) - Static method in class org.apache.beam.runners.fnexecution.data.RemoteInputDestination
 
of(K, Coder<K>) - Static method in class org.apache.beam.runners.local.StructuralKey
Create a new Structural Key of the provided key that can be encoded by the provided coder.
of(T, CloseableResource.Closer<T>) - Static method in class org.apache.beam.runners.portability.CloseableResource
Creates a CloseableResource with the given resource and closer.
of(JobApi.MetricResults) - Static method in class org.apache.beam.runners.portability.PortableMetrics
 
of(Coder<T>, Duration, boolean) - Static method in class org.apache.beam.runners.spark.io.CreateStream
Creates a new Spark based stream intended for test purposes.
of(Coder<T>, Duration) - Static method in class org.apache.beam.runners.spark.io.CreateStream
Creates a new Spark based stream without forced watermark sync, intended for test purposes.
of(NamedAggregators) - Static method in class org.apache.beam.runners.spark.metrics.AggregatorMetric
 
of(NamedAggregators) - Static method in class org.apache.beam.runners.spark.structuredstreaming.metrics.AggregatorMetric
 
of(Coder<TupleTag>, Map<TupleTag<?>, Coder<?>>, Coder<? extends BoundedWindow>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.MultiOuputCoder
 
of(SideInputReader) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.utils.CachedSideInputReader
Create a new cached SideInputReader.
of(SideInputReader) - Static method in class org.apache.beam.runners.spark.util.CachedSideInputReader
Create a new cached SideInputReader.
of(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.coders.AvroCoder
Returns an AvroCoder instance for the provided element type.
of(Class<T>) - Static method in class org.apache.beam.sdk.coders.AvroCoder
Returns an AvroCoder instance for the provided element class.
of(Schema) - Static method in class org.apache.beam.sdk.coders.AvroCoder
Returns an AvroGenericCoder instance for the Avro schema.
of(Class<T>, Schema) - Static method in class org.apache.beam.sdk.coders.AvroCoder
Returns an AvroCoder instance for the provided element type using the provided Avro schema.
of(Schema) - Static method in class org.apache.beam.sdk.coders.AvroGenericCoder
 
of() - Static method in class org.apache.beam.sdk.coders.BigDecimalCoder
 
of() - Static method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
 
of() - Static method in class org.apache.beam.sdk.coders.BigEndianLongCoder
 
of() - Static method in class org.apache.beam.sdk.coders.BigEndianShortCoder
 
of() - Static method in class org.apache.beam.sdk.coders.BigIntegerCoder
 
of() - Static method in class org.apache.beam.sdk.coders.BitSetCoder
 
of() - Static method in class org.apache.beam.sdk.coders.BooleanCoder
Returns the singleton instance of BooleanCoder.
of() - Static method in class org.apache.beam.sdk.coders.ByteArrayCoder
 
of() - Static method in class org.apache.beam.sdk.coders.ByteCoder
 
of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.CollectionCoder
 
of(Coder<IntermediateT>, DelegateCoder.CodingFunction<T, IntermediateT>, DelegateCoder.CodingFunction<IntermediateT, T>) - Static method in class org.apache.beam.sdk.coders.DelegateCoder
 
of(Coder<IntermediateT>, DelegateCoder.CodingFunction<T, IntermediateT>, DelegateCoder.CodingFunction<IntermediateT, T>, TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.coders.DelegateCoder
 
of() - Static method in class org.apache.beam.sdk.coders.DoubleCoder
 
of() - Static method in class org.apache.beam.sdk.coders.DurationCoder
 
of() - Static method in class org.apache.beam.sdk.coders.FloatCoder
 
of() - Static method in class org.apache.beam.sdk.coders.InstantCoder
 
of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.IterableCoder
 
of(Coder<K>, Coder<V>) - Static method in class org.apache.beam.sdk.coders.KvCoder
 
of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.LengthPrefixCoder
 
of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.ListCoder
 
of(Coder<K>, Coder<V>) - Static method in class org.apache.beam.sdk.coders.MapCoder
Produces a MapCoder with the given keyCoder and valueCoder.
of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.NullableCoder
 
of(Schema) - Static method in class org.apache.beam.sdk.coders.RowCoder
 
of(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.coders.SerializableCoder
Returns a SerializableCoder instance for the provided element type.
of(Class<T>) - Static method in class org.apache.beam.sdk.coders.SerializableCoder
Returns a SerializableCoder instance for the provided element class.
of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.SetCoder
Produces a SetCoder with the given elementCoder.
of(Coder<KeyT>) - Static method in class org.apache.beam.sdk.coders.ShardedKeyCoder
 
of(Coder<T>) - Static method in class org.apache.beam.sdk.coders.SnappyCoder
Wraps the given coder into a SnappyCoder.
of(Class<T>) - Static method in class org.apache.beam.sdk.coders.StringDelegateCoder
 
of(Class<T>, TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.coders.StringDelegateCoder
 
of() - Static method in class org.apache.beam.sdk.coders.StringUtf8Coder
 
of() - Static method in class org.apache.beam.sdk.coders.TextualIntegerCoder
 
of() - Static method in class org.apache.beam.sdk.coders.VarIntCoder
 
of() - Static method in class org.apache.beam.sdk.coders.VarLongCoder
 
of() - Static method in class org.apache.beam.sdk.coders.VoidCoder
 
of(Class<? extends InputT>) - Static method in class org.apache.beam.sdk.extensions.jackson.AsJsons
Creates a AsJsons PTransform that will transform a PCollection<InputT> into a PCollection of JSON Strings representing those objects using a Jackson ObjectMapper.
of(Class<? extends OutputT>) - Static method in class org.apache.beam.sdk.extensions.jackson.ParseJsons
Creates a ParseJsons PTransform that will parse JSON Strings into a PCollection<OutputT> using a Jackson ObjectMapper.
of() - Static method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
 
of(Descriptors.Descriptor) - Static method in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
Returns a DynamicProtoCoder for the Protocol Buffers DynamicMessage for the given Descriptors.Descriptor.
of(ProtoDomain, Descriptors.Descriptor) - Static method in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
Returns a DynamicProtoCoder for the Protocol Buffers DynamicMessage for the given Descriptors.Descriptor.
of(ProtoDomain, String) - Static method in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
Returns a DynamicProtoCoder for the Protocol Buffers DynamicMessage for the given message name in a ProtoDomain.
of(Class<T>) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
Returns a ProtoCoder for the given Protocol Buffers Message.
of(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
Returns a ProtoCoder for the Protocol Buffers Message indicated by the given TypeDescriptor.
of() - Static method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.HyperLogLogPlusCoder
 
of(BeamSqlTable) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteTable
 
of(Object...) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestBoundedTable
Convenient way to build a mocked bounded table.
of(Schema) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestBoundedTable
Build a mocked bounded table with the specified type.
of(Object...) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestUnboundedTable
Convenient way to build a mocked unbounded table.
of(FrameworkConfig, ExpressionConverter, RelOptCluster, QueryTrait) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ConversionContext
 
of() - Static method in class org.apache.beam.sdk.io.aws.dynamodb.AttributeValueCoder
 
of() - Static method in class org.apache.beam.sdk.io.aws2.dynamodb.AttributeValueCoder
 
of(String, TableSchema.ColumnType) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.Column
 
of(String, TableSchema.ColumnType, TableSchema.DefaultType, Object) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.Column
 
of(TableSchema.TypeName) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
of(TableSchema.Column...) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema
 
of() - Static method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.ParamsCoder
 
of(Coder<BoundedWindow>, Coder<DestinationT>) - Static method in class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
 
of() - Static method in class org.apache.beam.sdk.io.fs.MetadataCoder
Returns the singleton MetadataCoder instance.
of() - Static method in class org.apache.beam.sdk.io.fs.MetadataCoderV2
Returns the singleton MetadataCoderV2 instance.
of() - Static method in class org.apache.beam.sdk.io.fs.ResourceIdCoder
Creates a ResourceIdCoder.
of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertErrorCoder
 
of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoder
 
of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV2
 
of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV3
 
of() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
 
of() - Static method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteResultCoder
 
of(PubsubMessage, long, String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.OutgoingMessage
 
of(PubsubMessage, long, String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.OutgoingMessage
 
of() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessagePayloadOnlyCoder
 
of(TypeDescriptor<PubsubMessage>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesAndMessageIdCoder
 
of() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesAndMessageIdCoder
 
of(TypeDescriptor<PubsubMessage>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
 
of() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
 
of() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithMessageIdCoder
 
of(Class<T>) - Static method in class org.apache.beam.sdk.io.hadoop.WritableCoder
Returns a WritableCoder instance for the provided element class.
of(JdbcIO.DataSourceConfiguration) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceProviderFromDataSourceConfiguration
 
of(JdbcIO.DataSourceConfiguration) - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO.PoolableDataSourceProvider
 
of(String, String) - Static method in class org.apache.beam.sdk.io.kafka.ConfluentSchemaRegistryDeserializerProvider
 
of(String, String, Integer) - Static method in class org.apache.beam.sdk.io.kafka.ConfluentSchemaRegistryDeserializerProvider
 
of(Coder<K>, Coder<V>) - Static method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
 
of(Coder<K>, Coder<V>) - Static method in class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
 
of(int...) - Static method in class org.apache.beam.sdk.io.range.ByteKey
Creates a new ByteKey backed by a copy of the specified int[].
of(ByteKey, ByteKey) - Static method in class org.apache.beam.sdk.io.range.ByteKeyRange
Creates a new ByteKeyRange with the given start and end keys.
of(ByteKeyRange) - Static method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
Instantiates a new ByteKeyRangeTracker with the specified range.
of() - Static method in class org.apache.beam.sdk.io.range.OffsetRange.Coder
 
of() - Static method in class org.apache.beam.sdk.io.ReadableFileCoder
Returns the instance of ReadableFileCoder.
of(Class<T>) - Static method in class org.apache.beam.sdk.io.xml.JAXBCoder
Create a coder for a given type of JAXB annotated objects.
of(ValueProvider<X>, SerializableFunction<X, T>) - Static method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
Creates a ValueProvider.NestedValueProvider that wraps the provided value.
of(T) - Static method in class org.apache.beam.sdk.options.ValueProvider.StaticValueProvider
Creates a ValueProvider.StaticValueProvider that wraps the provided value.
of(FieldAccessDescriptor.FieldDescriptor.ListQualifier) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier
 
of(FieldAccessDescriptor.FieldDescriptor.MapQualifier) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier
 
of(int) - Static method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
 
of(String, Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.Schema.Field
Return's a field with the give name and type.
of(Schema.TypeName) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
Create a Schema.FieldType for the given type.
of(Schema.Field...) - Static method in class org.apache.beam.sdk.schemas.Schema
 
of(Schema, TypeDescriptor<T>, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Static method in class org.apache.beam.sdk.schemas.SchemaCoder
Returns a SchemaCoder for the specified class.
of(Schema) - Static method in class org.apache.beam.sdk.schemas.SchemaCoder
Returns a SchemaCoder for Row instances with the given schema.
of() - Static method in class org.apache.beam.sdk.schemas.transforms.Cast.Narrowing
 
of(Schema, Cast.Validator) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
 
of() - Static method in class org.apache.beam.sdk.schemas.transforms.Cast.Widening
 
of(Coder<T>) - Static method in class org.apache.beam.sdk.testing.TestStream.TestStreamCoder
 
of(SerializableBiFunction<V, V, V>) - Static method in class org.apache.beam.sdk.transforms.Combine.BinaryCombineFn
Returns a CombineFn that uses the given SerializableBiFunction to combine values.
of(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
Returns a CombineFn that uses the given SerializableFunction to combine values.
of(SerializableFunction<Iterable<V>, V>, int) - Static method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
Returns a CombineFn that uses the given SerializableFunction to combine values, attempting to buffer at least bufferSize values between invocations.
of(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine.SimpleCombineFn
Deprecated.
Returns a CombineFn that uses the given SerializableFunction to combine values.
of(ClosureT, Requirements) - Static method in class org.apache.beam.sdk.transforms.Contextful
Constructs a pair of the given closure and its requirements.
of(Iterable<T>) - Static method in class org.apache.beam.sdk.transforms.Create
Returns a new Create.Values transform that produces a PCollection containing elements of the provided Iterable.
of(T, T...) - Static method in class org.apache.beam.sdk.transforms.Create
Returns a new Create.Values transform that produces a PCollection containing the specified elements.
of(Map<K, V>) - Static method in class org.apache.beam.sdk.transforms.Create
Returns a new Create.Values transform that produces a PCollection of KVs corresponding to the keys and values of the specified Map.
of(DisplayData.Path, Class<?>, String) - Static method in class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
 
of(DoFn<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
of(CoGbkResultSchema, UnionCoder) - Static method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
Returns a CoGbkResult.CoGbkResultCoder for the given schema and UnionCoder.
of(TupleTag<V>, List<V>) - Static method in class org.apache.beam.sdk.transforms.join.CoGbkResult
Returns a new CoGbkResult that contains just the given tag and given data.
of(List<TupleTag<?>>) - Static method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
 
of(TupleTag<InputT>, PCollection<KV<K, InputT>>) - Static method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
Returns a new KeyedPCollectionTuple<K> with the given tag and initial PCollection.
of(String, PCollection<KV<K, InputT>>) - Static method in class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple
A version of KeyedPCollectionTuple.of(TupleTag, PCollection) that takes in a string instead of a TupleTag.
of(List<Coder<?>>) - Static method in class org.apache.beam.sdk.transforms.join.UnionCoder
Builds a union coder with the given list of element coders.
of(T, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Max
A CombineFn that computes the maximum of a collection of elements of type T using an arbitrary Comparator and identity, useful as an argument to Combine.globally(org.apache.beam.sdk.transforms.SerializableFunction<java.lang.Iterable<V>, V>) or Combine.perKey(org.apache.beam.sdk.transforms.SerializableFunction<java.lang.Iterable<V>, V>).
of(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Max
A CombineFn that computes the maximum of a collection of elements of type T using an arbitrary Comparator, useful as an argument to Combine.globally(org.apache.beam.sdk.transforms.SerializableFunction<java.lang.Iterable<V>, V>) or Combine.perKey(org.apache.beam.sdk.transforms.SerializableFunction<java.lang.Iterable<V>, V>).
of() - Static method in class org.apache.beam.sdk.transforms.Mean
A Combine.CombineFn that computes the arithmetic mean (a.k.a.
of(T, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Min
A CombineFn that computes the minimum of a collection of elements of type T using an arbitrary Comparator and an identity, useful as an argument to Combine.globally(org.apache.beam.sdk.transforms.SerializableFunction<java.lang.Iterable<V>, V>) or Combine.perKey(org.apache.beam.sdk.transforms.SerializableFunction<java.lang.Iterable<V>, V>).
of(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Min
A CombineFn that computes the minimum of a collection of elements of type T using an arbitrary Comparator, useful as an argument to Combine.globally(org.apache.beam.sdk.transforms.SerializableFunction<java.lang.Iterable<V>, V>) or Combine.perKey(org.apache.beam.sdk.transforms.SerializableFunction<java.lang.Iterable<V>, V>).
of(DoFn<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.ParDo
Creates a ParDo PTransform that will invoke the given DoFn function.
of(int, Partition.PartitionFn<? super T>) - Static method in class org.apache.beam.sdk.transforms.Partition
Returns a new Partition PTransform that divides its input PCollection into the given number of partitions, using the given partitioning function.
of() - Static method in class org.apache.beam.sdk.transforms.Reshuffle
Deprecated.
 
of(ByteKeyRange) - Static method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
 
of(RestrictionT, RestrictionT) - Static method in class org.apache.beam.sdk.transforms.splittabledofn.SplitResult
Returns a SplitResult for the specified primary and residual restrictions.
of() - Static method in class org.apache.beam.sdk.transforms.ToJson
 
of(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Top
Returns a PTransform that takes an input PCollection<T> and returns a PCollection<List<T>> with a single element containing the largest count elements of the input PCollection<T>, in decreasing order, sorted using the given Comparator<T>.
of(PCollectionView<ViewT>) - Static method in class org.apache.beam.sdk.transforms.View.CreatePCollectionView
 
of(Trigger.OnceTrigger...) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterAll
Returns an AfterAll Trigger with the given subtriggers.
of(List<Trigger>) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterAll
Returns an AfterAll Trigger with the given subtriggers.
of(Trigger.OnceTrigger...) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterFirst
Returns an AfterFirst Trigger with the given subtriggers.
of(List<Trigger>) - Static method in class org.apache.beam.sdk.transforms.windowing.AfterFirst
Returns an AfterFirst Trigger with the given subtriggers.
of() - Static method in class org.apache.beam.sdk.transforms.windowing.DefaultTrigger
Returns the default trigger.
of(Duration) - Static method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
Partitions the timestamp space into half-open intervals of the form [N * size, (N + 1) * size), where 0 is the epoch.
of() - Static method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
 
of() - Static method in class org.apache.beam.sdk.transforms.windowing.PaneInfo.PaneInfoCoder
 
of(Duration) - Static method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
Assigns timestamps into half-open intervals of the form [N * period, N * period + size), where 0 is the epoch.
of(T, Exception) - Static method in class org.apache.beam.sdk.transforms.WithFailures.ExceptionElement
 
of(OutputT, PCollection<FailureElementT>) - Static method in class org.apache.beam.sdk.transforms.WithFailures.Result
 
of(PCollection<OutputElementT>, PCollection<FailureElementT>) - Static method in class org.apache.beam.sdk.transforms.WithFailures.Result
 
of(PCollectionTuple, TupleTag<OutputElementT>, TupleTag<FailureElementT>) - Static method in class org.apache.beam.sdk.transforms.WithFailures.Result
 
of(SerializableFunction<V, K>) - Static method in class org.apache.beam.sdk.transforms.WithKeys
Returns a PTransform that takes a PCollection<V> and returns a PCollection<KV<K, V>>, where each of the values in the input PCollection has been paired with a key computed from the value by invoking the given SerializableFunction.
of(K) - Static method in class org.apache.beam.sdk.transforms.WithKeys
Returns a PTransform that takes a PCollection<V> and returns a PCollection<KV<K, V>>, where each of the values in the input PCollection has been paired with the given key.
of(SerializableFunction<T, Instant>) - Static method in class org.apache.beam.sdk.transforms.WithTimestamps
For a SerializableFunction fn from T to Instant, outputs a PTransform that takes an input PCollection<T> and outputs a PCollection<T> containing every element v in the input where each element is output with a timestamp obtained as the result of fn.apply(v).
of(K, V) - Static method in class org.apache.beam.sdk.values.KV
Returns a KV with the given key and value.
of(PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionList
Returns a singleton PCollectionList containing the given PCollection.
of(Iterable<PCollection<T>>) - Static method in class org.apache.beam.sdk.values.PCollectionList
Returns a PCollectionList containing the given PCollections, in order.
of(TupleTag<T>, PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
Returns a singleton PCollectionTuple containing the given PCollection keyed by the given TupleTag.
of(String, PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
A version of PCollectionTuple.of(TupleTag, PCollection) that takes in a String instead of a TupleTag.
of(String, PCollection<T>, String, PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
A version of PCollectionTuple.of(String, PCollection) that takes in two PCollections of the same type.
of(String, PCollection<T>, String, PCollection<T>, String, PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
A version of PCollectionTuple.of(String, PCollection) that takes in three PCollections of the same type.
of(String, PCollection<T>, String, PCollection<T>, String, PCollection<T>, String, PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
A version of PCollectionTuple.of(String, PCollection) that takes in four PCollections of the same type.
of(String, PCollection<T>, String, PCollection<T>, String, PCollection<T>, String, PCollection<T>, String, PCollection<T>) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
A version of PCollectionTuple.of(String, PCollection) that takes in five PCollections of the same type.
of(K, int) - Static method in class org.apache.beam.sdk.values.ShardedKey
 
of(TupleTag<?>, PValue) - Static method in class org.apache.beam.sdk.values.TaggedPValue
 
of(V, Instant) - Static method in class org.apache.beam.sdk.values.TimestampedValue
Returns a new TimestampedValue with the given value and timestamp.
of(Coder<T>) - Static method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
 
of(TupleTag<?>) - Static method in class org.apache.beam.sdk.values.TupleTagList
Returns a singleton TupleTagList containing the given TupleTag.
of(List<TupleTag<?>>) - Static method in class org.apache.beam.sdk.values.TupleTagList
Returns a TupleTagList containing the given TupleTags, in order.
of(Class<T>) - Static method in class org.apache.beam.sdk.values.TypeDescriptor
Returns a TypeDescriptor representing the given type.
of(Type) - Static method in class org.apache.beam.sdk.values.TypeDescriptor
Returns a TypeDescriptor representing the given type.
of(Coder<T>, Coder<? extends BoundedWindow>) - Static method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
 
of(T, Instant, BoundedWindow, PaneInfo) - Static method in class org.apache.beam.sdk.values.ValueInSingleWindow
 
of(Coder<ValueT>) - Static method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
 
of(WindowFn<T, W>) - Static method in class org.apache.beam.sdk.values.WindowingStrategy
 
ofDoubles() - Static method in class org.apache.beam.sdk.transforms.Max
ofDoubles() - Static method in class org.apache.beam.sdk.transforms.Min
ofDoubles() - Static method in class org.apache.beam.sdk.transforms.Sum
ofExpandedValue(PValue) - Static method in class org.apache.beam.sdk.values.TaggedPValue
 
offerCoders(Coder[]) - Method in interface org.apache.beam.sdk.state.StateSpec
For internal use only; no backwards-compatibility guarantees.
offeringClientsToPool(ControlClientPool.Sink, HeaderAccessor) - Static method in class org.apache.beam.runners.fnexecution.control.FnApiControlClientPoolService
Creates a new FnApiControlClientPoolService which will enqueue and vend new SDK harness connections.
ofFirstElement() - Static method in class org.apache.beam.sdk.transforms.windowing.AfterSynchronizedProcessingTime
 
offset(Duration) - Method in interface org.apache.beam.sdk.state.Timer
Offsets the target timestamp used by Timer.setRelative() by the given duration.
OFFSET_INFINITY - Static variable in class org.apache.beam.sdk.io.range.OffsetRangeTracker
Offset corresponding to infinity.
OffsetBasedReader(OffsetBasedSource<T>) - Constructor for class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
 
OffsetBasedSource<T> - Class in org.apache.beam.sdk.io
A BoundedSource that uses offsets to define starting and ending positions.
OffsetBasedSource(long, long, long) - Constructor for class org.apache.beam.sdk.io.OffsetBasedSource
 
OffsetBasedSource.OffsetBasedReader<T> - Class in org.apache.beam.sdk.io
A Source.Reader that implements code common to readers of all OffsetBasedSources.
OffsetRange - Class in org.apache.beam.sdk.io.range
A restriction represented by a range of integers [from, to).
OffsetRange(long, long) - Constructor for class org.apache.beam.sdk.io.range.OffsetRange
 
OffsetRange.Coder - Class in org.apache.beam.sdk.io.range
A coder for OffsetRanges.
OffsetRangeTracker - Class in org.apache.beam.sdk.io.range
A RangeTracker for non-negative positions of type long.
OffsetRangeTracker(long, long) - Constructor for class org.apache.beam.sdk.io.range.OffsetRangeTracker
Creates an OffsetRangeTracker for the specified range.
OffsetRangeTracker - Class in org.apache.beam.sdk.transforms.splittabledofn
A RestrictionTracker for claiming offsets in an OffsetRange in a monotonically increasing fashion.
OffsetRangeTracker(OffsetRange) - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
 
ofIntegers() - Static method in class org.apache.beam.sdk.transforms.Max
ofIntegers() - Static method in class org.apache.beam.sdk.transforms.Min
ofIntegers() - Static method in class org.apache.beam.sdk.transforms.Sum
ofLongs() - Static method in class org.apache.beam.sdk.transforms.Max
ofLongs() - Static method in class org.apache.beam.sdk.transforms.Min
ofLongs() - Static method in class org.apache.beam.sdk.transforms.Sum
ofNamed(Map<String, ?>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters
 
ofNone() - Static method in class org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters
 
ofPositional(List) - Static method in class org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters
 
ofPrimitiveOutputsInternal(Pipeline, TupleTagList, Map<TupleTag<?>, Coder<?>>, WindowingStrategy<?, ?>, PCollection.IsBounded) - Static method in class org.apache.beam.sdk.values.PCollectionTuple
For internal use only; no backwards-compatibility guarantees.
ofProvider(ValueProvider<T>, Coder<T>) - Static method in class org.apache.beam.sdk.transforms.Create
Returns an Create.OfValueProvider transform that produces a PCollection of a single element provided by the given ValueProvider.
ofSize(long) - Static method in class org.apache.beam.sdk.transforms.GroupIntoBatches
 
on(Join.FieldsEqual.Impl) - Method in class org.apache.beam.sdk.schemas.transforms.Join.Impl
Join the PCollections using the provided predicate.
on(PCollection<?>...) - Static method in class org.apache.beam.sdk.transforms.Wait
Waits on the given signal collections.
on(List<PCollection<?>>) - Static method in class org.apache.beam.sdk.transforms.Wait
Waits on the given signal collections.
ON_TIME_AND_ONLY_FIRING - Static variable in class org.apache.beam.sdk.transforms.windowing.PaneInfo
PaneInfo to use when there will be exactly one firing and it is on time.
onAdvance(int, int) - Method in class org.apache.beam.sdk.fn.stream.AdvancingPhaser
 
onBatchCompleted(JavaStreamingListenerBatchCompleted) - Method in class org.apache.beam.runners.spark.aggregators.AggregatorsAccumulator.AccumulatorCheckpointingSparkListener
 
onBatchCompleted(JavaStreamingListenerBatchCompleted) - Method in class org.apache.beam.runners.spark.metrics.MetricsAccumulator.AccumulatorCheckpointingSparkListener
 
onBatchCompleted(JavaStreamingListenerBatchCompleted) - Method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.WatermarkAdvancingStreamingListener
 
onBundleSuccess() - Method in interface org.apache.beam.sdk.transforms.DoFn.BundleFinalizer.Callback
 
OnceTrigger(List<Trigger>) - Constructor for class org.apache.beam.sdk.transforms.windowing.Trigger.OnceTrigger
 
onCheckpoint(BeamFnApi.ProcessBundleResponse) - Method in interface org.apache.beam.runners.fnexecution.control.BundleCheckpointHandler
 
onClaimed(PositionT) - Method in interface org.apache.beam.sdk.fn.splittabledofn.RestrictionTrackers.ClaimObserver
Called when RestrictionTracker.tryClaim(PositionT) returns true.
onClaimFailed(PositionT) - Method in interface org.apache.beam.sdk.fn.splittabledofn.RestrictionTrackers.ClaimObserver
Called when RestrictionTracker.tryClaim(PositionT) returns false.
onClose(Consumer<FnApiControlClient>) - Method in class org.apache.beam.runners.fnexecution.control.FnApiControlClient
 
onCompleted(BeamFnApi.ProcessBundleResponse) - Method in interface org.apache.beam.runners.fnexecution.control.BundleProgressHandler
Handles the bundle's completion report.
onCompleted() - Method in class org.apache.beam.sdk.fn.stream.BufferingStreamObserver
 
onCompleted() - Method in class org.apache.beam.sdk.fn.stream.DirectStreamObserver
 
onCompleted() - Method in class org.apache.beam.sdk.fn.stream.ForwardingClientResponseObserver
 
onCompleted() - Method in class org.apache.beam.sdk.fn.stream.SynchronizedStreamObserver
 
OneOfType - Class in org.apache.beam.sdk.schemas.logicaltypes
A logical type representing a union of fields.
OneOfType.Value - Class in org.apache.beam.sdk.schemas.logicaltypes
Represents a single OneOf value.
onError(Throwable) - Method in class org.apache.beam.sdk.fn.stream.BufferingStreamObserver
 
onError(Throwable) - Method in class org.apache.beam.sdk.fn.stream.DirectStreamObserver
 
onError(Throwable) - Method in class org.apache.beam.sdk.fn.stream.ForwardingClientResponseObserver
 
onError(Throwable) - Method in class org.apache.beam.sdk.fn.stream.SynchronizedStreamObserver
 
onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamAggregateProjectMergeRule
 
onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamAggregationRule
 
onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamBasicAggregationRule
 
onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamCoGBKJoinRule
 
onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamIOPushDownRule
 
onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamJoinAssociateRule
 
onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamJoinPushThroughJoinRule
 
onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamSideInputJoinRule
 
onMatch(RelOptRuleCall) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamUnnestRule
 
onNext(T) - Method in class org.apache.beam.sdk.fn.stream.BufferingStreamObserver
 
onNext(T) - Method in class org.apache.beam.sdk.fn.stream.DirectStreamObserver
 
onNext(ReqT) - Method in class org.apache.beam.sdk.fn.stream.ForwardingClientResponseObserver
 
onNext(V) - Method in class org.apache.beam.sdk.fn.stream.SynchronizedStreamObserver
 
onProgress(BeamFnApi.ProcessBundleProgressResponse) - Method in interface org.apache.beam.runners.fnexecution.control.BundleProgressHandler
Handles a progress report from the bundle while it is executing.
onSeenNewOutput(Instant, StateT) - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
Called by the Watch transform to compute a new termination state, in case after calling the Watch.Growth.PollFn for the current input, the Watch.Growth.PollResult included a previously unseen OutputT.
onStartup() - Method in interface org.apache.beam.sdk.harness.JvmInitializer
Implement onStartup to run some custom initialization immediately after the JVM is launched for pipeline execution.
onSuccess(List<KinesisRecord>) - Method in interface org.apache.beam.sdk.io.kinesis.RateLimitPolicy
Called after Kinesis records are successfully retrieved.
onSuccess(List<KinesisRecord>) - Method in class org.apache.beam.sdk.io.kinesis.RateLimitPolicyFactory.DelayIntervalRateLimiter
 
onThrottle(KinesisClientThrottledException) - Method in interface org.apache.beam.sdk.io.kinesis.RateLimitPolicy
Called after the Kinesis client is throttled.
onTimer(String, String, BoundedWindow, Instant, Instant, TimeDomain) - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
 
OnTimerContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.OnTimerContext
 
open(MetricConfig) - Method in class org.apache.beam.runners.flink.metrics.FileReporter
 
open(TaskContext, Instant) - Method in class org.apache.beam.runners.gearpump.translators.io.GearpumpSource
 
open(GcsPath) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
Opens an object in GCS.
open(WritableByteChannel) - Method in class org.apache.beam.sdk.io.AvroIO.Sink
 
open(String) - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
Opens a uniquely named temporary file and initializes the writer using FileBasedSink.Writer.prepareWrite(java.nio.channels.WritableByteChannel).
open() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
Returns a ReadableByteChannel reading the data from this file, potentially decompressing it using FileIO.ReadableFile.getCompression().
open(WritableByteChannel) - Method in interface org.apache.beam.sdk.io.FileIO.Sink
Initializes writing to the given channel.
open(ResourceIdT) - Method in class org.apache.beam.sdk.io.FileSystem
Returns a read channel for the given ResourceIdT.
open(ResourceId) - Static method in class org.apache.beam.sdk.io.FileSystems
Returns a read channel for the given ResourceId.
open(WritableByteChannel) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Sink
 
open(WritableByteChannel) - Method in class org.apache.beam.sdk.io.TextIO.Sink
 
open(WritableByteChannel) - Method in class org.apache.beam.sdk.io.TFRecordIO.Sink
 
open(WritableByteChannel) - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.Sink
 
open(WritableByteChannel) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Sink
 
openManifest(String) - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactRetrievalService
 
openManifest(String) - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactStagingService
 
openManifest(String) - Method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactRetrievalService
 
openManifest(String) - Method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService
 
openManifest(String) - Method in class org.apache.beam.runners.fnexecution.artifact.ClassLoaderLegacyArtifactRetrievalService
 
openManifest(String) - Method in class org.apache.beam.runners.fnexecution.artifact.JavaFilesystemLegacyArtifactStagingService
 
openSeekable() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
Returns a SeekableByteChannel equivalent to FileIO.ReadableFile.open(), but fails if this file is not seekable.
openUri(String, String) - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactRetrievalService
 
openUri(String) - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactStagingService
 
openUri(String, String) - Method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactRetrievalService
 
openUri(String) - Method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService
 
openUri(String, String) - Method in class org.apache.beam.runners.fnexecution.artifact.ClassLoaderLegacyArtifactRetrievalService
 
openUri(String) - Method in class org.apache.beam.runners.fnexecution.artifact.JavaFilesystemLegacyArtifactStagingService
 
optimizedWrites() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
If true, enables new codepaths that are expected to use less resources while writing to BigQuery.
Options() - Constructor for class org.apache.beam.runners.apex.ApexRunnerRegistrar.Options
 
Options() - Constructor for class org.apache.beam.runners.dataflow.DataflowPipelineRegistrar.Options
 
Options() - Constructor for class org.apache.beam.runners.direct.DirectRegistrar.Options
 
Options() - Constructor for class org.apache.beam.runners.flink.FlinkRunnerRegistrar.Options
 
options() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobPreparation
 
Options() - Constructor for class org.apache.beam.runners.gearpump.GearpumpRunnerRegistrar.Options
 
Options() - Constructor for class org.apache.beam.runners.jet.JetRunnerRegistrar.Options
 
Options() - Constructor for class org.apache.beam.runners.spark.SparkRunnerRegistrar.Options
 
options() - Static method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter
 
options - Variable in class org.apache.beam.sdk.extensions.sorter.ExternalSorter
 
Options() - Constructor for class org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options
 
Options() - Constructor for class org.apache.beam.sdk.options.ManualDockerEnvironmentOptions.Options
 
Options() - Constructor for class org.apache.beam.sdk.options.RemoteEnvironmentOptions.Options
 
Order - Class in org.apache.beam.sdk.extensions.sql.example.model
Describes an order.
Order(int, int) - Constructor for class org.apache.beam.sdk.extensions.sql.example.model.Order
 
Order() - Constructor for class org.apache.beam.sdk.extensions.sql.example.model.Order
 
OrderByKey() - Constructor for class org.apache.beam.sdk.values.KV.OrderByKey
 
OrderByValue() - Constructor for class org.apache.beam.sdk.values.KV.OrderByValue
 
orFinally(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
Specify an ending condition for this trigger.
OrFinallyTrigger - Class in org.apache.beam.sdk.transforms.windowing
A Trigger that executes according to its main trigger until its "finally" trigger fires.
org.apache.beam.runners.apex - package org.apache.beam.runners.apex
Implementation of the Beam runner for Apache Apex.
org.apache.beam.runners.dataflow - package org.apache.beam.runners.dataflow
Provides a Beam runner that executes pipelines on the Google Cloud Dataflow service.
org.apache.beam.runners.dataflow.options - package org.apache.beam.runners.dataflow.options
Provides PipelineOptions specific to Google Cloud Dataflow.
org.apache.beam.runners.dataflow.util - package org.apache.beam.runners.dataflow.util
Provides miscellaneous internal utilities used by the Google Cloud Dataflow runner.
org.apache.beam.runners.direct - package org.apache.beam.runners.direct
Defines the PipelineOptions.DirectRunner which executes both Bounded and Unbounded Pipelines on the local machine.
org.apache.beam.runners.flink - package org.apache.beam.runners.flink
Internal implementation of the Beam runner for Apache Flink.
org.apache.beam.runners.flink.metrics - package org.apache.beam.runners.flink.metrics
Internal metrics implementation of the Beam runner for Apache Flink.
org.apache.beam.runners.fnexecution - package org.apache.beam.runners.fnexecution
Utilities used by runners to interact with the fn execution components of the Beam Portability Framework.
org.apache.beam.runners.fnexecution.artifact - package org.apache.beam.runners.fnexecution.artifact
Pipeline execution-time artifact-management services, including abstract implementations of the Artifact Retrieval Service.
org.apache.beam.runners.fnexecution.control - package org.apache.beam.runners.fnexecution.control
Utilities for a Beam runner to interact with the Fn API Control Service via java abstractions.
org.apache.beam.runners.fnexecution.data - package org.apache.beam.runners.fnexecution.data
Utilities for a Beam runner to interact with the Fn API Data Service via java abstractions.
org.apache.beam.runners.fnexecution.environment - package org.apache.beam.runners.fnexecution.environment
Classes used to instantiate and manage SDK harness environments.
org.apache.beam.runners.fnexecution.environment.testing - package org.apache.beam.runners.fnexecution.environment.testing
Test utilities for the environment management package.
org.apache.beam.runners.fnexecution.jobsubmission - package org.apache.beam.runners.fnexecution.jobsubmission
Job management services for use in beam runners.
org.apache.beam.runners.fnexecution.logging - package org.apache.beam.runners.fnexecution.logging
Classes used to log informational messages over the Beam Fn Logging Service.
org.apache.beam.runners.fnexecution.provisioning - package org.apache.beam.runners.fnexecution.provisioning
Provision api services.
org.apache.beam.runners.fnexecution.splittabledofn - package org.apache.beam.runners.fnexecution.splittabledofn
Utilities for a Beam runner to interact with a remotely running splittable DoFn.
org.apache.beam.runners.fnexecution.state - package org.apache.beam.runners.fnexecution.state
State API services.
org.apache.beam.runners.fnexecution.status - package org.apache.beam.runners.fnexecution.status
Worker Status API services.
org.apache.beam.runners.fnexecution.translation - package org.apache.beam.runners.fnexecution.translation
Shared utilities for a Beam runner to translate portable pipelines.
org.apache.beam.runners.fnexecution.wire - package org.apache.beam.runners.fnexecution.wire
Wire coders for communications between runner and SDK harness.
org.apache.beam.runners.gearpump - package org.apache.beam.runners.gearpump
Internal implementation of the Beam runner for Apache Gearpump.
org.apache.beam.runners.gearpump.translators - package org.apache.beam.runners.gearpump.translators
Gearpump specific translators.
org.apache.beam.runners.gearpump.translators.functions - package org.apache.beam.runners.gearpump.translators.functions
Gearpump specific wrappers for Beam DoFn.
org.apache.beam.runners.gearpump.translators.io - package org.apache.beam.runners.gearpump.translators.io
Gearpump specific wrappers for Beam I/O.
org.apache.beam.runners.gearpump.translators.utils - package org.apache.beam.runners.gearpump.translators.utils
Utilities for translators.
org.apache.beam.runners.jet - package org.apache.beam.runners.jet
Implementation of the Beam runner for Hazelcast Jet.
org.apache.beam.runners.jet.metrics - package org.apache.beam.runners.jet.metrics
Helper classes for implementing metrics in the Hazelcast Jet based runner.
org.apache.beam.runners.jet.processors - package org.apache.beam.runners.jet.processors
Individual DAG node processors used by the Beam runner for Hazelcast Jet.
org.apache.beam.runners.local - package org.apache.beam.runners.local
Utilities useful when executing a pipeline on a single machine.
org.apache.beam.runners.portability - package org.apache.beam.runners.portability
Support for executing a pipeline locally over the Beam fn API.
org.apache.beam.runners.portability.testing - package org.apache.beam.runners.portability.testing
Testing utilities for the reference runner.
org.apache.beam.runners.spark - package org.apache.beam.runners.spark
Internal implementation of the Beam runner for Apache Spark.
org.apache.beam.runners.spark.aggregators - package org.apache.beam.runners.spark.aggregators
Provides internal utilities for implementing Beam aggregators using Spark accumulators.
org.apache.beam.runners.spark.coders - package org.apache.beam.runners.spark.coders
Beam coders and coder-related utilities for running on Apache Spark.
org.apache.beam.runners.spark.io - package org.apache.beam.runners.spark.io
Spark-specific transforms for I/O.
org.apache.beam.runners.spark.metrics - package org.apache.beam.runners.spark.metrics
Provides internal utilities for implementing Beam metrics using Spark accumulators.
org.apache.beam.runners.spark.metrics.sink - package org.apache.beam.runners.spark.metrics.sink
Spark sinks that supports beam metrics and aggregators.
org.apache.beam.runners.spark.stateful - package org.apache.beam.runners.spark.stateful
Spark-specific stateful operators.
org.apache.beam.runners.spark.structuredstreaming - package org.apache.beam.runners.spark.structuredstreaming
Internal implementation of the Beam runner for Apache Spark.
org.apache.beam.runners.spark.structuredstreaming.aggregators - package org.apache.beam.runners.spark.structuredstreaming.aggregators
Provides internal utilities for implementing Beam aggregators using Spark accumulators.
org.apache.beam.runners.spark.structuredstreaming.examples - package org.apache.beam.runners.spark.structuredstreaming.examples
 
org.apache.beam.runners.spark.structuredstreaming.metrics - package org.apache.beam.runners.spark.structuredstreaming.metrics
Provides internal utilities for implementing Beam metrics using Spark accumulators.
org.apache.beam.runners.spark.structuredstreaming.metrics.sink - package org.apache.beam.runners.spark.structuredstreaming.metrics.sink
Spark sinks that supports beam metrics and aggregators.
org.apache.beam.runners.spark.structuredstreaming.translation - package org.apache.beam.runners.spark.structuredstreaming.translation
Internal translators for running Beam pipelines on Spark.
org.apache.beam.runners.spark.structuredstreaming.translation.batch - package org.apache.beam.runners.spark.structuredstreaming.translation.batch
Internal utilities to translate Beam pipelines to Spark batching.
org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions - package org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions
Internal implementation of the Beam runner for Apache Spark.
org.apache.beam.runners.spark.structuredstreaming.translation.helpers - package org.apache.beam.runners.spark.structuredstreaming.translation.helpers
Internal helpers to translate Beam pipelines to Spark streaming.
org.apache.beam.runners.spark.structuredstreaming.translation.streaming - package org.apache.beam.runners.spark.structuredstreaming.translation.streaming
Internal utilities to translate Beam pipelines to Spark streaming.
org.apache.beam.runners.spark.structuredstreaming.translation.utils - package org.apache.beam.runners.spark.structuredstreaming.translation.utils
Internal utils to translate Beam pipelines to Spark streaming.
org.apache.beam.runners.spark.util - package org.apache.beam.runners.spark.util
Internal utilities to translate Beam pipelines to Spark.
org.apache.beam.sdk - package org.apache.beam.sdk
Provides a simple, powerful model for building both batch and streaming parallel data processing Pipelines.
org.apache.beam.sdk.annotations - package org.apache.beam.sdk.annotations
Defines annotations used across the SDK.
org.apache.beam.sdk.coders - package org.apache.beam.sdk.coders
Defines Coders to specify how data is encoded to and decoded from byte strings.
org.apache.beam.sdk.expansion - package org.apache.beam.sdk.expansion
Contains classes needed to expose transforms to other SDKs.
org.apache.beam.sdk.expansion.service - package org.apache.beam.sdk.expansion.service
Classes used to expand cross-language transforms.
org.apache.beam.sdk.extensions.gcp.auth - package org.apache.beam.sdk.extensions.gcp.auth
Defines classes related to interacting with Credentials for pipeline creation and execution containing Google Cloud Platform components.
org.apache.beam.sdk.extensions.gcp.options - package org.apache.beam.sdk.extensions.gcp.options
Defines PipelineOptions for configuring pipeline execution for Google Cloud Platform components.
org.apache.beam.sdk.extensions.gcp.storage - package org.apache.beam.sdk.extensions.gcp.storage
Defines IO connectors for Google Cloud Storage.
org.apache.beam.sdk.extensions.gcp.util - package org.apache.beam.sdk.extensions.gcp.util
Defines Google Cloud Platform component utilities that can be used by Beam runners.
org.apache.beam.sdk.extensions.gcp.util.gcsfs - package org.apache.beam.sdk.extensions.gcp.util.gcsfs
Defines utilities used to interact with Google Cloud Storage.
org.apache.beam.sdk.extensions.jackson - package org.apache.beam.sdk.extensions.jackson
Utilities for parsing and creating JSON serialized objects.
org.apache.beam.sdk.extensions.joinlibrary - package org.apache.beam.sdk.extensions.joinlibrary
Utilities for performing SQL-style joins of keyed PCollections.
org.apache.beam.sdk.extensions.ml - package org.apache.beam.sdk.extensions.ml
Provides DoFns for integration with Google Cloud AI Video Intelligence service.
org.apache.beam.sdk.extensions.protobuf - package org.apache.beam.sdk.extensions.protobuf
Defines a Coder for Protocol Buffers messages, ProtoCoder.
org.apache.beam.sdk.extensions.sketching - package org.apache.beam.sdk.extensions.sketching
Utilities for computing statistical indicators using probabilistic sketches.
org.apache.beam.sdk.extensions.sorter - package org.apache.beam.sdk.extensions.sorter
Utility for performing local sort of potentially large sets of values.
org.apache.beam.sdk.extensions.sql - package org.apache.beam.sdk.extensions.sql
BeamSQL provides a new interface to run a SQL statement with Beam.
org.apache.beam.sdk.extensions.sql.example - package org.apache.beam.sdk.extensions.sql.example
Example how to use Data Catalog table provider.
org.apache.beam.sdk.extensions.sql.example.model - package org.apache.beam.sdk.extensions.sql.example.model
Java classes used to for modeling the examples.
org.apache.beam.sdk.extensions.sql.impl - package org.apache.beam.sdk.extensions.sql.impl
Implementation classes of BeamSql.
org.apache.beam.sdk.extensions.sql.impl.parser - package org.apache.beam.sdk.extensions.sql.impl.parser
Beam SQL parsing additions to Calcite SQL.
org.apache.beam.sdk.extensions.sql.impl.planner - package org.apache.beam.sdk.extensions.sql.impl.planner
BeamQueryPlanner is the main interface.
org.apache.beam.sdk.extensions.sql.impl.rel - package org.apache.beam.sdk.extensions.sql.impl.rel
BeamSQL specified nodes, to replace RelNode.
org.apache.beam.sdk.extensions.sql.impl.rule - package org.apache.beam.sdk.extensions.sql.impl.rule
RelOptRule to generate BeamRelNode.
org.apache.beam.sdk.extensions.sql.impl.schema - package org.apache.beam.sdk.extensions.sql.impl.schema
define table schema, to map with Beam IO components.
org.apache.beam.sdk.extensions.sql.impl.transform - package org.apache.beam.sdk.extensions.sql.impl.transform
PTransform used in a BeamSql pipeline.
org.apache.beam.sdk.extensions.sql.impl.transform.agg - package org.apache.beam.sdk.extensions.sql.impl.transform.agg
Implementation of standard SQL aggregation functions, e.g.
org.apache.beam.sdk.extensions.sql.impl.udf - package org.apache.beam.sdk.extensions.sql.impl.udf
UDF classes.
org.apache.beam.sdk.extensions.sql.impl.utils - package org.apache.beam.sdk.extensions.sql.impl.utils
Utility classes.
org.apache.beam.sdk.extensions.sql.meta - package org.apache.beam.sdk.extensions.sql.meta
Metadata related classes.
org.apache.beam.sdk.extensions.sql.meta.provider - package org.apache.beam.sdk.extensions.sql.meta.provider
Table providers.
org.apache.beam.sdk.extensions.sql.meta.provider.avro - package org.apache.beam.sdk.extensions.sql.meta.provider.avro
Table schema for AvroIO.
org.apache.beam.sdk.extensions.sql.meta.provider.bigquery - package org.apache.beam.sdk.extensions.sql.meta.provider.bigquery
Table schema for BigQuery.
org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog - package org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog
Table schema for Google Cloud Data Catalog.
org.apache.beam.sdk.extensions.sql.meta.provider.datastore - package org.apache.beam.sdk.extensions.sql.meta.provider.datastore
Table schema for DataStore.
org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog - package org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog
Table schema for HCatalog.
org.apache.beam.sdk.extensions.sql.meta.provider.kafka - package org.apache.beam.sdk.extensions.sql.meta.provider.kafka
Table schema for KafkaIO.
org.apache.beam.sdk.extensions.sql.meta.provider.mongodb - package org.apache.beam.sdk.extensions.sql.meta.provider.mongodb
Table schema for MongoDb.
org.apache.beam.sdk.extensions.sql.meta.provider.parquet - package org.apache.beam.sdk.extensions.sql.meta.provider.parquet
Table schema for ParquetIO.
org.apache.beam.sdk.extensions.sql.meta.provider.pubsub - package org.apache.beam.sdk.extensions.sql.meta.provider.pubsub
Table schema for PubsubIO.
org.apache.beam.sdk.extensions.sql.meta.provider.seqgen - package org.apache.beam.sdk.extensions.sql.meta.provider.seqgen
Table schema for streaming sequence generator.
org.apache.beam.sdk.extensions.sql.meta.provider.test - package org.apache.beam.sdk.extensions.sql.meta.provider.test
Table schema for in-memory test data.
org.apache.beam.sdk.extensions.sql.meta.provider.text - package org.apache.beam.sdk.extensions.sql.meta.provider.text
Table schema for text files.
org.apache.beam.sdk.extensions.sql.meta.store - package org.apache.beam.sdk.extensions.sql.meta.store
Meta stores.
org.apache.beam.sdk.extensions.sql.zetasql - package org.apache.beam.sdk.extensions.sql.zetasql
ZetaSQL Dialect package.
org.apache.beam.sdk.extensions.sql.zetasql.translation - package org.apache.beam.sdk.extensions.sql.zetasql.translation
Conversion logic between ZetaSQL resolved query nodes and Calcite rel nodes.
org.apache.beam.sdk.extensions.zetasketch - package org.apache.beam.sdk.extensions.zetasketch
PTransforms to compute statistical sketches on data streams based on the ZetaSketch implementation.
org.apache.beam.sdk.fn - package org.apache.beam.sdk.fn
The top level package for the Fn Execution Java libraries.
org.apache.beam.sdk.fn.channel - package org.apache.beam.sdk.fn.channel
gRPC channel management.
org.apache.beam.sdk.fn.data - package org.apache.beam.sdk.fn.data
Classes to interact with the portability framework data plane.
org.apache.beam.sdk.fn.splittabledofn - package org.apache.beam.sdk.fn.splittabledofn
Defines utilities related to executing splittable DoFn.
org.apache.beam.sdk.fn.stream - package org.apache.beam.sdk.fn.stream
gRPC stream management.
org.apache.beam.sdk.fn.test - package org.apache.beam.sdk.fn.test
Utilities for testing use of this package.
org.apache.beam.sdk.fn.windowing - package org.apache.beam.sdk.fn.windowing
Common utilities related to windowing during execution of a pipeline.
org.apache.beam.sdk.function - package org.apache.beam.sdk.function
Java 8 functional interface extensions.
org.apache.beam.sdk.harness - package org.apache.beam.sdk.harness
Utilities for configuring worker environment.
org.apache.beam.sdk.io - package org.apache.beam.sdk.io
Defines transforms for reading and writing common storage formats, including AvroIO, and TextIO.
org.apache.beam.sdk.io.amqp - package org.apache.beam.sdk.io.amqp
Transforms for reading and writing using AMQP 1.0 protocol.
org.apache.beam.sdk.io.aws.coders - package org.apache.beam.sdk.io.aws.coders
Defines common coders for Amazon Web Services.
org.apache.beam.sdk.io.aws.dynamodb - package org.apache.beam.sdk.io.aws.dynamodb
Defines IO connectors for Amazon Web Services DynamoDB.
org.apache.beam.sdk.io.aws.options - package org.apache.beam.sdk.io.aws.options
Defines PipelineOptions for configuring pipeline execution for Amazon Web Services components.
org.apache.beam.sdk.io.aws.s3 - package org.apache.beam.sdk.io.aws.s3
Defines IO connectors for Amazon Web Services S3.
org.apache.beam.sdk.io.aws.sns - package org.apache.beam.sdk.io.aws.sns
Defines IO connectors for Amazon Web Services SNS.
org.apache.beam.sdk.io.aws.sqs - package org.apache.beam.sdk.io.aws.sqs
Defines IO connectors for Amazon Web Services SQS.
org.apache.beam.sdk.io.aws2.dynamodb - package org.apache.beam.sdk.io.aws2.dynamodb
Defines IO connectors for Amazon Web Services DynamoDB.
org.apache.beam.sdk.io.aws2.options - package org.apache.beam.sdk.io.aws2.options
Defines PipelineOptions for configuring pipeline execution for Amazon Web Services components.
org.apache.beam.sdk.io.aws2.sns - package org.apache.beam.sdk.io.aws2.sns
Defines IO connectors for Amazon Web Services SNS.
org.apache.beam.sdk.io.aws2.sqs - package org.apache.beam.sdk.io.aws2.sqs
Defines IO connectors for Amazon Web Services SQS.
org.apache.beam.sdk.io.cassandra - package org.apache.beam.sdk.io.cassandra
Transforms for reading and writing from/to Apache Cassandra.
org.apache.beam.sdk.io.clickhouse - package org.apache.beam.sdk.io.clickhouse
Transform for writing to ClickHouse.
org.apache.beam.sdk.io.elasticsearch - package org.apache.beam.sdk.io.elasticsearch
Transforms for reading and writing from Elasticsearch.
org.apache.beam.sdk.io.fs - package org.apache.beam.sdk.io.fs
Apache Beam FileSystem interfaces and their default implementations.
org.apache.beam.sdk.io.gcp.bigquery - package org.apache.beam.sdk.io.gcp.bigquery
Defines transforms for reading and writing from Google BigQuery.
org.apache.beam.sdk.io.gcp.bigtable - package org.apache.beam.sdk.io.gcp.bigtable
Defines transforms for reading and writing from Google Cloud Bigtable.
org.apache.beam.sdk.io.gcp.common - package org.apache.beam.sdk.io.gcp.common
Defines common Google Cloud Platform IO support classes.
org.apache.beam.sdk.io.gcp.datastore - package org.apache.beam.sdk.io.gcp.datastore
Provides an API for reading from and writing to Google Cloud Datastore over different versions of the Cloud Datastore Client libraries.
org.apache.beam.sdk.io.gcp.pubsub - package org.apache.beam.sdk.io.gcp.pubsub
Defines transforms for reading and writing from Google Cloud Pub/Sub.
org.apache.beam.sdk.io.gcp.spanner - package org.apache.beam.sdk.io.gcp.spanner
Provides an API for reading from and writing to Google Cloud Spanner.
org.apache.beam.sdk.io.gcp.testing - package org.apache.beam.sdk.io.gcp.testing
Defines utilities for unit testing Google Cloud Platform components of Apache Beam pipelines.
org.apache.beam.sdk.io.hadoop - package org.apache.beam.sdk.io.hadoop
Classes shared by Hadoop based IOs.
org.apache.beam.sdk.io.hadoop.format - package org.apache.beam.sdk.io.hadoop.format
Defines transforms for writing to Data sinks that implement HadoopFormatIO .
org.apache.beam.sdk.io.hbase - package org.apache.beam.sdk.io.hbase
Transforms for reading and writing from/to Apache HBase.
org.apache.beam.sdk.io.hcatalog - package org.apache.beam.sdk.io.hcatalog
Transforms for reading and writing using HCatalog.
org.apache.beam.sdk.io.hcatalog.test - package org.apache.beam.sdk.io.hcatalog.test
Test utilities for HCatalog IO.
org.apache.beam.sdk.io.hdfs - package org.apache.beam.sdk.io.hdfs
FileSystem implementation for any Hadoop FileSystem.
org.apache.beam.sdk.io.jdbc - package org.apache.beam.sdk.io.jdbc
Transforms for reading and writing from JDBC.
org.apache.beam.sdk.io.jms - package org.apache.beam.sdk.io.jms
Transforms for reading and writing from JMS (Java Messaging Service).
org.apache.beam.sdk.io.kafka - package org.apache.beam.sdk.io.kafka
Transforms for reading and writing from Apache Kafka.
org.apache.beam.sdk.io.kafka.serialization - package org.apache.beam.sdk.io.kafka.serialization
Kafka serializers and deserializers.
org.apache.beam.sdk.io.kinesis - package org.apache.beam.sdk.io.kinesis
Transforms for reading and writing from Amazon Kinesis.
org.apache.beam.sdk.io.mongodb - package org.apache.beam.sdk.io.mongodb
Transforms for reading and writing from MongoDB.
org.apache.beam.sdk.io.mqtt - package org.apache.beam.sdk.io.mqtt
Transforms for reading and writing from MQTT.
org.apache.beam.sdk.io.parquet - package org.apache.beam.sdk.io.parquet
Transforms for reading and writing from Parquet.
org.apache.beam.sdk.io.range - package org.apache.beam.sdk.io.range
Provides thread-safe helpers for implementing dynamic work rebalancing in position-based bounded sources.
org.apache.beam.sdk.io.redis - package org.apache.beam.sdk.io.redis
Transforms for reading and writing from Redis.
org.apache.beam.sdk.io.solr - package org.apache.beam.sdk.io.solr
Transforms for reading and writing from/to Solr.
org.apache.beam.sdk.io.thrift - package org.apache.beam.sdk.io.thrift
Transforms for reading and writing to Thrift files.
org.apache.beam.sdk.io.tika - package org.apache.beam.sdk.io.tika
Transform for reading and parsing files with Apache Tika.
org.apache.beam.sdk.io.xml - package org.apache.beam.sdk.io.xml
Transforms for reading and writing Xml files.
org.apache.beam.sdk.metrics - package org.apache.beam.sdk.metrics
Metrics allow exporting information about the execution of a pipeline.
org.apache.beam.sdk.options - package org.apache.beam.sdk.options
Defines PipelineOptions for configuring pipeline execution.
org.apache.beam.sdk.schemas - package org.apache.beam.sdk.schemas
Defines Schema and other classes for representing schema'd data in a Pipeline.
org.apache.beam.sdk.schemas.annotations - package org.apache.beam.sdk.schemas.annotations
Defines Schema and other classes for representing schema'd data in a Pipeline.
org.apache.beam.sdk.schemas.logicaltypes - package org.apache.beam.sdk.schemas.logicaltypes
A set of common LogicalTypes for use with schemas.
org.apache.beam.sdk.schemas.parser - package org.apache.beam.sdk.schemas.parser
Defines utilities for deailing with schemas.
org.apache.beam.sdk.schemas.parser.generated - package org.apache.beam.sdk.schemas.parser.generated
Defines utilities for deailing with schemas.
org.apache.beam.sdk.schemas.transforms - package org.apache.beam.sdk.schemas.transforms
Defines transforms that work on PCollections with schemas..
org.apache.beam.sdk.schemas.utils - package org.apache.beam.sdk.schemas.utils
Defines utilities for deailing with schemas.
org.apache.beam.sdk.state - package org.apache.beam.sdk.state
Classes and interfaces for interacting with state.
org.apache.beam.sdk.testing - package org.apache.beam.sdk.testing
Defines utilities for unit testing Apache Beam pipelines.
org.apache.beam.sdk.transforms - package org.apache.beam.sdk.transforms
Defines PTransforms for transforming data in a pipeline.
org.apache.beam.sdk.transforms.display - package org.apache.beam.sdk.transforms.display
Defines HasDisplayData for annotating components which provide display data used within UIs and diagnostic tools.
org.apache.beam.sdk.transforms.join - package org.apache.beam.sdk.transforms.join
Defines the CoGroupByKey transform for joining multiple PCollections.
org.apache.beam.sdk.transforms.splittabledofn - package org.apache.beam.sdk.transforms.splittabledofn
Defines utilities related to splittable DoFn.
org.apache.beam.sdk.transforms.windowing - package org.apache.beam.sdk.transforms.windowing
Defines the Window transform for dividing the elements in a PCollection into windows, and the Trigger for controlling when those elements are output.
org.apache.beam.sdk.values - package org.apache.beam.sdk.values
Defines PCollection and other classes for representing data in a Pipeline.
OTHER - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
out() - Static method in class org.apache.beam.runners.spark.io.ConsoleIO.Write
Prints elements from the PCollection to the console.
out(int) - Static method in class org.apache.beam.runners.spark.io.ConsoleIO.Write
Prints num elements from the PCollection to stdout.
outbound(DataStreams.OutputChunkConsumer<ByteString>) - Static method in class org.apache.beam.sdk.fn.stream.DataStreams
Converts a single element delimited OutputStream into multiple ByteStrings.
outbound(DataStreams.OutputChunkConsumer<ByteString>, int) - Static method in class org.apache.beam.sdk.fn.stream.DataStreams
Converts a single element delimited OutputStream into multiple ByteStrings using the specified maximum chunk size.
OutboundObserverFactory - Class in org.apache.beam.sdk.fn.stream
Creates factories which determine an underlying StreamObserver implementation to use in to interact with fn execution APIs.
OutboundObserverFactory() - Constructor for class org.apache.beam.sdk.fn.stream.OutboundObserverFactory
 
OutboundObserverFactory.BasicFactory<ReqT,RespT> - Interface in org.apache.beam.sdk.fn.stream
Creates an outbound observer for the given inbound observer.
outboundObserverFor(StreamObserver<ReqT>) - Method in interface org.apache.beam.sdk.fn.stream.OutboundObserverFactory.BasicFactory
 
outboundObserverFor(OutboundObserverFactory.BasicFactory<ReqT, RespT>, StreamObserver<ReqT>) - Method in class org.apache.beam.sdk.fn.stream.OutboundObserverFactory
Creates an outbound observer for the given inbound observer by potentially inserting hooks into the inbound and outbound observers.
OUTER - Static variable in class org.apache.beam.sdk.coders.Coder.Context
Deprecated.
The outer context: the value being encoded or decoded takes up the remainder of the record/stream contents.
OutgoingMessage() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.OutgoingMessage
 
OUTPUT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
output(T) - Method in interface org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.ParserCallback
Output the object.
output(T, Instant) - Method in interface org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.ParserCallback
Output the object using the specified timestamp.
output(OutputT, Instant, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFn.FinishBundleContext
Adds the given element to the main output PCollection at the given timestamp in the given window.
output(TupleTag<T>, T, Instant, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFn.FinishBundleContext
Adds the given element to the output PCollection with the given tag at the given timestamp in the given window.
output(T) - Method in interface org.apache.beam.sdk.transforms.DoFn.OutputReceiver
 
output(OutputT) - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
Adds the given element to the main output PCollection.
output(TupleTag<T>, T) - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
Adds the given element to the output PCollection with the given tag.
output() - Method in class org.apache.beam.sdk.transforms.WithFailures.Result
 
OUTPUT_DIR - Static variable in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
MRJobConfig.MAPREDUCE_JOB_DIR.
OUTPUT_FORMAT_CLASS_ATTR - Static variable in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
MRJobConfig.OUTPUT_FORMAT_CLASS_ATTR.
OUTPUT_INFO - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
OUTPUT_KEY_CLASS - Static variable in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
MRJobConfig.OUTPUT_KEY_CLASS.
OUTPUT_NAME - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
OUTPUT_VALUE_CLASS - Static variable in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
MRJobConfig.OUTPUT_VALUE_CLASS.
outputColumnMap - Variable in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
 
outputOf(ProcessFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
Returns a type descriptor for the output of the given ProcessFunction, subject to Java type erasure: may contain unresolved type variables if the type was erased.
outputOf(SerializableFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
Binary compatibility adapter for TypeDescriptors.outputOf(ProcessFunction).
outputOf(Contextful.Fn<InputT, OutputT>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
Like TypeDescriptors.outputOf(ProcessFunction) but for Contextful.Fn.
OutputReceiverFactory - Interface in org.apache.beam.runners.fnexecution.control
A factory that can create output receivers during an executable stage.
OutputReference - Class in org.apache.beam.runners.dataflow.util
A representation used by Steps to reference the output of other Steps.
OutputReference(String, String) - Constructor for class org.apache.beam.runners.dataflow.util.OutputReference
 
outputRuntimeOptions() - Method in interface org.apache.beam.sdk.options.PipelineOptions
Returns a map of properties which correspond to ValueProvider.RuntimeValueProvider, keyed by the property name.
outputSchema() - Method in class org.apache.beam.sdk.schemas.transforms.Cast
 
outputSchemaCoder - Variable in class org.apache.beam.sdk.schemas.utils.ConvertHelpers.ConvertedSchemaInformation
 
outputWithTimestamp(T, Instant) - Method in interface org.apache.beam.sdk.transforms.DoFn.OutputReceiver
 
outputWithTimestamp(OutputT, Instant) - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
Adds the given element to the main output PCollection, with the given timestamp.
outputWithTimestamp(TupleTag<T>, T, Instant) - Method in class org.apache.beam.sdk.transforms.DoFn.WindowedContext
Adds the given element to the specified output PCollection, with the given timestamp.
overlaps(ByteKeyRange) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
Returns true if the specified ByteKeyRange overlaps this range.

P

pane() - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContext
Returns information about the pane within this window into which the input element has been assigned.
PaneInfo - Class in org.apache.beam.sdk.transforms.windowing
Provides information about the pane an element belongs to.
PaneInfo.PaneInfoCoder - Class in org.apache.beam.sdk.transforms.windowing
A Coder for encoding PaneInfo instances.
PaneInfo.Timing - Enum in org.apache.beam.sdk.transforms.windowing
Enumerates the possibilities for the timing of this pane firing related to the input and output watermarks for its computation.
PARALLEL_INPUT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
ParameterListBuilder() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase.ParameterListBuilder
 
parameters - Variable in class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase
Types of parameter for the function call.
Params() - Constructor for class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
Construct a default Params object.
ParamsCoder() - Constructor for class org.apache.beam.sdk.io.DefaultFilenamePolicy.ParamsCoder
 
ParDo - Class in org.apache.beam.sdk.transforms
ParDo is the core element-wise transform in Apache Beam, invoking a user-specified function on each of the elements of the input PCollection to produce zero or more output elements, all of which are collected into the output PCollection.
ParDo() - Constructor for class org.apache.beam.sdk.transforms.ParDo
 
ParDo.MultiOutput<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
A PTransform that, when applied to a PCollection<InputT>, invokes a user-specified DoFn<InputT, OutputT> on all its elements, which can emit elements to any of the PTransform's output PCollections, which are bundled into a result PCollectionTuple.
ParDo.SingleOutput<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
A PTransform that, when applied to a PCollection<InputT>, invokes a user-specified DoFn<InputT, OutputT> on all its elements, with all its outputs collected into an output PCollection<OutputT>.
ParDoMultiOutputTranslator<InputT,OutputT> - Class in org.apache.beam.runners.gearpump.translators
ParDo.MultiOutput is translated to Gearpump flatMap function with DoFn wrapped in DoFnFunction.
ParDoMultiOutputTranslator() - Constructor for class org.apache.beam.runners.gearpump.translators.ParDoMultiOutputTranslator
 
ParDoMultiOverrideFactory<InputT,OutputT> - Class in org.apache.beam.runners.direct
A PTransformOverrideFactory that provides overrides for applications of a ParDo in the direct runner.
ParDoMultiOverrideFactory() - Constructor for class org.apache.beam.runners.direct.ParDoMultiOverrideFactory
 
ParDoP<InputT,OutputT> - Class in org.apache.beam.runners.jet.processors
Jet Processor implementation for Beam's ParDo primitive (when no user-state is being used).
ParDoP.Supplier<InputT,OutputT> - Class in org.apache.beam.runners.jet.processors
Jet Processor supplier that will provide instances of ParDoP.
parent() - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold.Context
Type of parent node in a tree.
ParquetIO - Class in org.apache.beam.sdk.io.parquet
IO to read and write Parquet files.
ParquetIO.Read - Class in org.apache.beam.sdk.io.parquet
Implementation of ParquetIO.read(Schema).
ParquetIO.ReadFiles - Class in org.apache.beam.sdk.io.parquet
Implementation of ParquetIO.readFiles(Schema).
ParquetIO.Sink - Class in org.apache.beam.sdk.io.parquet
ParquetTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.parquet
ParquetTable(Schema, String) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.parquet.ParquetTable
 
ParquetTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.parquet
ParquetTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.parquet.ParquetTableProvider
 
parse(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner
Parse input SQL query, and return a SqlNode as grammar tree.
parse(String) - Method in interface org.apache.beam.sdk.extensions.sql.impl.QueryPlanner
Parse input SQL query, and return a SqlNode as grammar tree.
parse(String) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl
 
parse(Reader) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl
 
parse(String) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
 
Parse() - Constructor for class org.apache.beam.sdk.io.AvroIO.Parse
 
parse(String) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
Parse string with ClickHouse type to TableSchema.ColumnType.
parse(String) - Static method in enum org.apache.beam.sdk.io.clickhouse.TableSchema.DefaultType
 
parse(GridFSDBFile, MongoDbGridFSIO.ParserCallback<T>) - Method in interface org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Parser
 
parse() - Static method in class org.apache.beam.sdk.io.tika.TikaIO
Parses files matching a given filepattern.
Parse() - Constructor for class org.apache.beam.sdk.io.tika.TikaIO.Parse
 
parse(String) - Static method in class org.apache.beam.sdk.schemas.parser.FieldAccessDescriptorParser
 
parse(T) - Method in class org.apache.beam.sdk.testing.JsonMatcher
 
ParseAll() - Constructor for class org.apache.beam.sdk.io.AvroIO.ParseAll
Deprecated.
 
parseAllGenericRecords(SerializableFunction<GenericRecord, T>) - Static method in class org.apache.beam.sdk.io.AvroIO
Deprecated.
You can achieve The functionality of AvroIO.parseAllGenericRecords(SerializableFunction) using FileIO matching plus AvroIO.parseFilesGenericRecords(SerializableFunction) ()}. This is the preferred method to make composition explicit. AvroIO.ParseAll will not receive upgrades and will be removed in a future version of Beam.
parseArgs(String[]) - Static method in class org.apache.beam.runners.flink.FlinkJobServerDriver
 
parseDate(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
parseDateToValue(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
parseDefaultExpression(TableSchema.ColumnType, String) - Static method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
Get default value of a column based on expression.
ParseException - Exception in org.apache.beam.sdk.extensions.sql.impl
Exception thrown when Beam SQL is unable to parse the statement.
ParseException(Throwable) - Constructor for exception org.apache.beam.sdk.extensions.sql.impl.ParseException
 
ParseException(String, Throwable) - Constructor for exception org.apache.beam.sdk.extensions.sql.impl.ParseException
 
ParseFiles() - Constructor for class org.apache.beam.sdk.io.AvroIO.ParseFiles
 
parseFiles() - Static method in class org.apache.beam.sdk.io.tika.TikaIO
Parses files in a PCollection of FileIO.ReadableFile.
ParseFiles() - Constructor for class org.apache.beam.sdk.io.tika.TikaIO.ParseFiles
 
parseFilesGenericRecords(SerializableFunction<GenericRecord, T>) - Static method in class org.apache.beam.sdk.io.AvroIO
parseGenericRecords(SerializableFunction<GenericRecord, T>) - Static method in class org.apache.beam.sdk.io.AvroIO
Reads Avro file(s) containing records of an unspecified schema and converting each record to a custom type.
ParseJsons<OutputT> - Class in org.apache.beam.sdk.extensions.jackson
PTransform for parsing JSON Strings.
ParseJsons.ParseJsonsWithFailures<FailureT> - Class in org.apache.beam.sdk.extensions.jackson
A PTransform that adds exception handling to ParseJsons.
parseQuery(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
 
parseQuery(String, QueryPlanner.QueryParameters) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
 
ParseResult - Class in org.apache.beam.sdk.io.tika
The result of parsing a single file with Tika: contains the file's location, metadata, extracted text, and optionally an error.
parseTableSpec(String) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
Parse a table specification in the form "[project_id]:[dataset_id].[table_id]" or "[dataset_id].[table_id]".
parseTime(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
parseTimestampWithLocalTimeZone(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
parseTimestampWithoutTimeZone(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
parseTimestampWithTimeZone(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
parseTimestampWithTZToValue(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
parseTimestampWithUTCTimeZone(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
parseTimeToValue(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
Partition<T> - Class in org.apache.beam.sdk.transforms
Partition takes a PCollection<T> and a PartitionFn, uses the PartitionFn to split the elements of the input PCollection into N partitions, and returns a PCollectionList<T> that bundles N PCollection<T>s containing the split elements.
Partition.PartitionFn<T> - Interface in org.apache.beam.sdk.transforms
A function object that chooses an output partition for an element.
PartitionContext() - Constructor for class org.apache.beam.sdk.io.kafka.TimestampPolicy.PartitionContext
 
partitioner() - Method in class org.apache.beam.runners.spark.io.SourceRDD.Unbounded
 
PARTITIONER_CLASS_ATTR - Static variable in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
MRJobConfig.PARTITIONER_CLASS_ATTR.
partitionFor(T, int) - Method in interface org.apache.beam.sdk.transforms.Partition.PartitionFn
Chooses the partition into which to put the given element.
PartitioningWindowFn<T,W extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
A WindowFn that places each value into exactly one window based on its timestamp and never merges windows.
PartitioningWindowFn() - Constructor for class org.apache.beam.sdk.transforms.windowing.PartitioningWindowFn
 
PartitionMark(String, int, long, long) - Constructor for class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
 
PAssert - Class in org.apache.beam.sdk.testing
An assertion on the contents of a PCollection incorporated into the pipeline.
PAssert.DefaultConcludeTransform - Class in org.apache.beam.sdk.testing
Default transform to check that a PAssert was successful.
PAssert.GroupThenAssert<T> - Class in org.apache.beam.sdk.testing
A transform that applies an assertion-checking function over iterables of ActualT to the entirety of the contents of its input.
PAssert.GroupThenAssertForSingleton<T> - Class in org.apache.beam.sdk.testing
A transform that applies an assertion-checking function to the sole element of a PCollection.
PAssert.IterableAssert<T> - Interface in org.apache.beam.sdk.testing
Builder interface for assertions applicable to iterables and PCollection contents.
PAssert.OneSideInputAssert<ActualT> - Class in org.apache.beam.sdk.testing
An assertion checker that takes a single PCollectionView<ActualT> and an assertion over ActualT, and checks it within a Beam pipeline.
PAssert.PAssertionSite - Class in org.apache.beam.sdk.testing
Track the place where an assertion is defined.
PAssert.PCollectionContentsAssert<T> - Class in org.apache.beam.sdk.testing
An PAssert.IterableAssert about the contents of a PCollection.
PAssert.PCollectionContentsAssert.MatcherCheckerFn<T> - Class in org.apache.beam.sdk.testing
Check that the passed-in matchers match the existing data.
PAssert.SingletonAssert<T> - Interface in org.apache.beam.sdk.testing
Builder interface for assertions applicable to a single value.
PassThroughLogicalType<T> - Class in org.apache.beam.sdk.schemas.logicaltypes
A base class for LogicalTypes that use the same Java type as the underlying base type.
PassThroughLogicalType(String, Schema.FieldType, Object, Schema.FieldType) - Constructor for class org.apache.beam.sdk.schemas.logicaltypes.PassThroughLogicalType
 
pastEndOfWindow() - Static method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark
Creates a trigger that fires when the watermark passes the end of the window.
pastFirstElementInPane() - Static method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
Creates a trigger that fires when the current processing time passes the processing time at which this trigger saw the first element in a pane.
patchTableDescription(TableReference, String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.DatasetService
Patch BigQuery Table description.
patchTableDescription(TableReference, String) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
path() - Method in class org.apache.beam.sdk.schemas.transforms.Cast.CompatibilityError
 
path() - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold.Context
Field path from a root of a schema.
PathValidator - Interface in org.apache.beam.sdk.extensions.gcp.storage
For internal use only; no backwards compatibility guarantees.
PathValidatorFactory() - Constructor for class org.apache.beam.sdk.extensions.gcp.options.GcsOptions.PathValidatorFactory
 
PBegin - Class in org.apache.beam.sdk.values
PBegin is the "input" to a root PTransform, such as Read or Create.
PBegin(Pipeline) - Constructor for class org.apache.beam.sdk.values.PBegin
Constructs a PBegin in the given Pipeline.
PCollection<T> - Class in org.apache.beam.sdk.values
A PCollection<T> is an immutable collection of values of type T.
PCollection.IsBounded - Enum in org.apache.beam.sdk.values
The enumeration of cases for whether a PCollection is bounded.
PCollectionContentsAssert(PCollection<T>, PAssert.PAssertionSite) - Constructor for class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
 
PCollectionContentsAssert(PCollection<T>, PAssert.AssertionWindows, SimpleFunction<Iterable<ValueInSingleWindow<T>>, Iterable<T>>, PAssert.PAssertionSite) - Constructor for class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
 
PCollectionList<T> - Class in org.apache.beam.sdk.values
A PCollectionList<T> is an immutable list of homogeneously typed PCollection<T>s.
pCollections() - Static method in class org.apache.beam.sdk.transforms.Flatten
Returns a PTransform that flattens a PCollectionList into a PCollection containing all the elements of all the PCollections in its input.
PCollectionTuple - Class in org.apache.beam.sdk.values
A PCollectionTuple is an immutable tuple of heterogeneously-typed PCollections, "keyed" by TupleTags.
PCollectionView<T> - Interface in org.apache.beam.sdk.values
A PCollectionView<T> is an immutable view of a PCollection as a value of type T that can be accessed as a side input to a ParDo transform.
PCollectionViews - Class in org.apache.beam.sdk.values
For internal use only; no backwards compatibility guarantees.
PCollectionViews() - Constructor for class org.apache.beam.sdk.values.PCollectionViews
 
PCollectionViews.IterableViewFn<T> - Class in org.apache.beam.sdk.values
Implementation which is able to adapt a multimap materialization to a Iterable<T>.
PCollectionViews.ListViewFn<T> - Class in org.apache.beam.sdk.values
Implementation which is able to adapt a multimap materialization to a List<T>.
PCollectionViews.MapViewFn<K,V> - Class in org.apache.beam.sdk.values
Implementation which is able to adapt a multimap materialization to a Map<K, V>.
PCollectionViews.MultimapViewFn<K,V> - Class in org.apache.beam.sdk.values
Implementation which is able to adapt a multimap materialization to a Map<K, Iterable<V>>.
PCollectionViews.SimplePCollectionView<ElemT,PrimitiveViewT,ViewT,W extends BoundedWindow> - Class in org.apache.beam.sdk.values
A class for PCollectionView implementations, with additional type parameters that are not visible at pipeline assembly time when the view is used as a side input.
PCollectionViews.SingletonViewFn<T> - Class in org.apache.beam.sdk.values
Implementation which is able to adapt a multimap materialization to a T.
PCollectionViews.TypeDescriptorSupplier<T> - Interface in org.apache.beam.sdk.values
 
PDone - Class in org.apache.beam.sdk.values
PDone is the output of a PTransform that has a trivial result, such as a WriteFiles.
peekOutputElements() - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
peekOutputElements(TupleTag<T>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
peekOutputElementsInWindow(BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
peekOutputElementsInWindow(TupleTag<OutputT>, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
peekOutputElementsWithTimestamp() - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
perElement() - Static method in class org.apache.beam.sdk.transforms.Count
Returns a PTransform that counts the number of occurrences of each element in its input PCollection.
perKey() - Static method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct
Like ApproximateDistinct.globally() but per key, i.e computes the approximate number of distinct values per key in a PCollection<KV<K, V>> and returns PCollection<KV<K, Long>>.
perKey() - Static method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies
Like SketchFrequencies.globally() but per key, i.e a Count-min sketch per key in PCollection<KV<K, V>> and returns a PCollection<KV<K, {@link CountMinSketch}>>.
perKey() - Static method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles
Like TDigestQuantiles.globally(), but builds a digest for each key in the stream.
perKey() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Extract
Returns a PTransform that takes an input PCollection<KV<K, byte[]>> of (key, HLL++ sketch) pairs and returns a PCollection<KV<K, Long>> of (key, estimated count of distinct elements extracted from each sketch).
perKey() - Method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Init.Builder
Returns a Combine.PerKey PTransform that takes an input PCollection<KV<K, InputT>> and returns a PCollection<KV<K, byte[]>> which consists of the per-key HLL++ sketch computed from the values matching each key in the input PCollection.
perKey() - Static method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.MergePartial
Returns a Combine.PerKey PTransform that takes an input PCollection<KV<K, byte[]>> of (key, HLL++ sketch) pairs and returns a PCollection<KV<K, byte[]>> of (key, new sketch merged from the input sketches under the key).
perKey(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles
Returns a PTransform that takes a PCollection<KV<K, V>> and returns a PCollection<KV<K, List<V>>> that contains an output element mapping each distinct key in the input PCollection to a List of the approximate N-tiles of the values associated with that key in the input PCollection.
perKey(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateQuantiles
Like ApproximateQuantiles.perKey(int, Comparator), but sorts values using the their natural ordering.
perKey(int) - Static method in class org.apache.beam.sdk.transforms.ApproximateUnique
Returns a PTransform that takes a PCollection<KV<K, V>> and returns a PCollection<KV<K, Long>> that contains an output element mapping each distinct key in the input PCollection to an estimate of the number of distinct values associated with that key in the input PCollection.
perKey(double) - Static method in class org.apache.beam.sdk.transforms.ApproximateUnique
Like ApproximateUnique.perKey(int), but specifies the desired maximum estimation error instead of the sample size.
PerKey(int) - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique.PerKey
 
PerKey(double) - Constructor for class org.apache.beam.sdk.transforms.ApproximateUnique.PerKey
 
perKey(SerializableFunction<Iterable<V>, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
Returns a Combine.PerKey PTransform that first groups its input PCollection of KVs by keys and windows, then invokes the given function on each of the values lists to produce a combined value, and then returns a PCollection of KVs mapping each distinct key to its combined value for each window.
perKey(SerializableBiFunction<V, V, V>) - Static method in class org.apache.beam.sdk.transforms.Combine
Returns a Combine.PerKey PTransform that first groups its input PCollection of KVs by keys and windows, then invokes the given function on each of the values lists to produce a combined value, and then returns a PCollection of KVs mapping each distinct key to its combined value for each window.
perKey(CombineFnBase.GlobalCombineFn<? super InputT, ?, OutputT>) - Static method in class org.apache.beam.sdk.transforms.Combine
Returns a Combine.PerKey PTransform that first groups its input PCollection of KVs by keys and windows, then invokes the given function on each of the values lists to produce a combined value, and then returns a PCollection of KVs mapping each distinct key to its combined value for each window.
perKey() - Static method in class org.apache.beam.sdk.transforms.Count
Returns a PTransform that counts the number of elements associated with each key of its input PCollection.
perKey() - Static method in class org.apache.beam.sdk.transforms.Latest
Returns a PTransform that takes as input a PCollection<KV<K, V>> and returns a PCollection<KV<K, V>> whose contents is the latest element per-key according to its event time.
perKey() - Static method in class org.apache.beam.sdk.transforms.Max
Returns a PTransform that takes an input PCollection<KV<K, T>> and returns a PCollection<KV<K, T>> that contains an output element mapping each distinct key in the input PCollection to the maximum according to the natural ordering of T of the values associated with that key in the input PCollection.
perKey(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Max
Returns a PTransform that takes an input PCollection<KV<K, T>> and returns a PCollection<KV<K, T>> that contains one output element per key mapping each to the maximum of the values associated with that key in the input PCollection.
perKey() - Static method in class org.apache.beam.sdk.transforms.Mean
Returns a PTransform that takes an input PCollection<KV<K, N>> and returns a PCollection<KV<K, Double>> that contains an output element mapping each distinct key in the input PCollection to the mean of the values associated with that key in the input PCollection.
perKey() - Static method in class org.apache.beam.sdk.transforms.Min
Returns a PTransform that takes an input PCollection<KV<K, T>> and returns a PCollection<KV<K, T>> that contains an output element mapping each distinct key in the input PCollection to the minimum according to the natural ordering of T of the values associated with that key in the input PCollection.
perKey(ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Min
Returns a PTransform that takes an input PCollection<KV<K, T>> and returns a PCollection<KV<K, T>> that contains one output element per key mapping each to the minimum of the values associated with that key in the input PCollection.
perKey(int, ComparatorT) - Static method in class org.apache.beam.sdk.transforms.Top
Returns a PTransform that takes an input PCollection<KV<K, V>> and returns a PCollection<KV<K, List<V>>> that contains an output element mapping each distinct key in the input PCollection to the largest count values associated with that key in the input PCollection<KV<K, V>>, in decreasing order, sorted using the given Comparator<V>.
PerKeyDigest() - Constructor for class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.PerKeyDigest
 
PerKeyDistinct() - Constructor for class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.PerKeyDistinct
 
PerKeySketch() - Constructor for class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.PerKeySketch
 
PInput - Interface in org.apache.beam.sdk.values
The interface for things that might be input to a PTransform.
pipeline() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobPreparation
 
Pipeline - Class in org.apache.beam.sdk
A Pipeline manages a directed acyclic graph of PTransforms, and the PCollections that the PTransforms consume and produce.
Pipeline(PipelineOptions) - Constructor for class org.apache.beam.sdk.Pipeline
 
Pipeline.PipelineExecutionException - Exception in org.apache.beam.sdk
Thrown during execution of a Pipeline, whenever user code within that Pipeline throws an exception.
Pipeline.PipelineVisitor - Interface in org.apache.beam.sdk
For internal use only; no backwards-compatibility guarantees.
Pipeline.PipelineVisitor.CompositeBehavior - Enum in org.apache.beam.sdk
Control enum for indicating whether or not a traversal should process the contents of a composite transform or not.
Pipeline.PipelineVisitor.Defaults - Class in org.apache.beam.sdk
Default no-op Pipeline.PipelineVisitor that enters all composite transforms.
PIPELINE_PROTO_CODER_ID - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
PIPELINED - Static variable in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
PipelineExecutionException(Throwable) - Constructor for exception org.apache.beam.sdk.Pipeline.PipelineExecutionException
PipelineMessageReceiver - Interface in org.apache.beam.runners.local
Handles failures in the form of exceptions.
pipelineOptions() - Method in class org.apache.beam.runners.fnexecution.provisioning.JobInfo
 
PipelineOptions - Interface in org.apache.beam.sdk.options
PipelineOptions are used to configure Pipelines.
PipelineOptions.AtomicLongFactory - Class in org.apache.beam.sdk.options
DefaultValueFactory which supplies an ID that is guaranteed to be unique within the given process.
PipelineOptions.CheckEnabled - Enum in org.apache.beam.sdk.options
Enumeration of the possible states for a given check.
PipelineOptions.DirectRunner - Class in org.apache.beam.sdk.options
A DefaultValueFactory that obtains the class of the DirectRunner if it exists on the classpath, and throws an exception otherwise.
PipelineOptions.JobNameFactory - Class in org.apache.beam.sdk.options
Returns a normalized job name constructed from ApplicationNameOptions.getAppName(), the local system user name (if available), the current time, and a random integer.
PipelineOptions.UserAgentFactory - Class in org.apache.beam.sdk.options
Returns a user agent string constructed from ReleaseInfo.getName() and ReleaseInfo.getVersion(), in the format [name]/[version].
PipelineOptionsFactory - Class in org.apache.beam.sdk.options
Constructs a PipelineOptions or any derived interface that is composable to any other derived interface of PipelineOptions via the PipelineOptions.as(java.lang.Class<T>) method.
PipelineOptionsFactory() - Constructor for class org.apache.beam.sdk.options.PipelineOptionsFactory
 
PipelineOptionsFactory.Builder - Class in org.apache.beam.sdk.options
A fluent PipelineOptions builder.
PipelineOptionsRegistrar - Interface in org.apache.beam.sdk.options
PipelineOptions creators have the ability to automatically have their PipelineOptions registered with this SDK by creating a ServiceLoader entry and a concrete implementation of this interface.
PipelineOptionsValidator - Class in org.apache.beam.sdk.options
Validates that the PipelineOptions conforms to all the Validation criteria.
PipelineOptionsValidator() - Constructor for class org.apache.beam.sdk.options.PipelineOptionsValidator
 
PipelineResult - Interface in org.apache.beam.sdk
Result of Pipeline.run().
PipelineResult.State - Enum in org.apache.beam.sdk
Possible job states, for both completed and ongoing jobs.
PipelineRunner<ResultT extends PipelineResult> - Class in org.apache.beam.sdk
PipelineRunner() - Constructor for class org.apache.beam.sdk.PipelineRunner
 
PipelineTranslator - Class in org.apache.beam.runners.spark.structuredstreaming.translation
Pipeline.PipelineVisitor that translates the Beam operators to their Spark counterparts.
PipelineTranslator() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
 
PipelineTranslatorBatch - Class in org.apache.beam.runners.spark.structuredstreaming.translation.batch
PipelineTranslator for executing a Pipeline in Spark in batch mode.
PipelineTranslatorBatch(SparkStructuredStreamingPipelineOptions) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.batch.PipelineTranslatorBatch
 
PipelineTranslatorStreaming - Class in org.apache.beam.runners.spark.structuredstreaming.translation.streaming
PipelineTranslator for executing a Pipeline in Spark in streaming mode.
PipelineTranslatorStreaming(SparkStructuredStreamingPipelineOptions) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.streaming.PipelineTranslatorStreaming
 
PipelineTranslatorUtils - Class in org.apache.beam.runners.fnexecution.translation
Utilities for pipeline translation.
plus(RelOptCost) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
plus(NodeStats) - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
 
plusDelayOf(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
Adds some delay to the original target time.
POJOUtils - Class in org.apache.beam.sdk.schemas.utils
A set of utilities to generate getter and setter classes for POJOs.
POJOUtils() - Constructor for class org.apache.beam.sdk.schemas.utils.POJOUtils
 
PollFn() - Constructor for class org.apache.beam.sdk.transforms.Watch.Growth.PollFn
 
pollFor(Duration) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery.PollingAssertion
 
pollJob(JobReference, int) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.JobService
Waits for the job is Done, and returns the job.
pollJob(JobReference, int) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.ApproximateDistinctFn
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.CountMinSketchFn
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.TDigestQuantilesFn
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.amqp.AmqpIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.AvroIO.ParseAll
Deprecated.
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.AvroIO.ParseFiles
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.AvroIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.AvroIO.ReadAll
Deprecated.
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.AvroIO.ReadFiles
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.AvroIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.BoundedReadFromUnboundedSource
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.CompressedSource
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
Populates the display data.
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileBasedSink.FilenamePolicy
Populates the display data.
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileBasedSink
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileBasedSource
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileIO.Match
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileIO.MatchAll
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.FileIO.ReadMatches
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageQuerySource
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageTableSource
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.WriteWithResults
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write.PubsubBoundedWriter
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.GenerateSequence
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceProviderFromDataSourceConfiguration
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.PoolableDataSourceProvider
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.TypedWithoutMetadata
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.Read.Bounded
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.Read.Unbounded
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
Populate the display data with connectionConfiguration details.
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.Source
Register display data for the given transform or component.
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TextIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
Deprecated.
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TextIO.ReadFiles
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TextIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.ReadFiles
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.tika.TikaIO.Parse
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.tika.TikaIO.ParseFiles
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.WriteFiles
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.Globally
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.ApproximateUnique.PerKey
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.Globally
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.GloballyAsSingletonView
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.IterableCombineFn
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Combine.PerKeyWithHotKeyFanout
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
 
populateDisplayData(DisplayData.Builder) - Method in interface org.apache.beam.sdk.transforms.display.HasDisplayData
Register display data for the given transform or component.
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.DoFn
Register display data for the given transform or component.
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Filter
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.FlatMapElements.FlatMapWithFailures
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.FlatMapElements
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.GroupByKey
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.InferableFunction
Register display data for the given transform or component.
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.MapElements.MapWithFailures
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.MapElements
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
Register display data for the given transform or component.
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Partition
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.PTransform
Register display data for the given transform or component.
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Sample.FixedSizedSampleFn
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.Top.TopCombineFn
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.Window.Assign
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.Window
 
populateDisplayData(DisplayData.Builder) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
Register display data for the given transform or component.
portableMetrics() - Method in class org.apache.beam.runners.flink.FlinkPortableRunnerResult
 
portableMetrics() - Method in interface org.apache.beam.runners.fnexecution.jobsubmission.PortablePipelineResult
Returns the object to access monitoring infos from the pipeline.
PortableMetrics - Class in org.apache.beam.runners.portability
 
PortablePipelineJarCreator - Class in org.apache.beam.runners.fnexecution.jobsubmission
PortablePipelineRunner that bundles the input pipeline along with all dependencies, artifacts, etc.
PortablePipelineJarCreator(Class) - Constructor for class org.apache.beam.runners.fnexecution.jobsubmission.PortablePipelineJarCreator
 
PortablePipelineJarUtils - Class in org.apache.beam.runners.fnexecution.jobsubmission
Contains common code for writing and reading portable pipeline jars.
PortablePipelineJarUtils() - Constructor for class org.apache.beam.runners.fnexecution.jobsubmission.PortablePipelineJarUtils
 
PortablePipelineOptions - Interface in org.apache.beam.sdk.options
Pipeline options common to all portable runners.
PortablePipelineOptions.RetrievalServiceType - Enum in org.apache.beam.sdk.options
Enumeration of the different implementations of the artifact retrieval service.
PortablePipelineResult - Interface in org.apache.beam.runners.fnexecution.jobsubmission
PortablePipelineRunner - Interface in org.apache.beam.runners.fnexecution.jobsubmission
Runs a portable Beam pipeline on some execution engine.
PortableRunner - Class in org.apache.beam.runners.portability
A PipelineRunner a Pipeline against a JobService.
PortableRunnerRegistrar - Class in org.apache.beam.runners.portability
Registrar for the portable runner.
PortableRunnerRegistrar() - Constructor for class org.apache.beam.runners.portability.PortableRunnerRegistrar
 
positional() - Method in class org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters
 
POutput - Interface in org.apache.beam.sdk.values
The interface for things that might be output from a PTransform.
precisionForRelativeError(double) - Static method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct
Computes the precision based on the desired relative error.
prepare(JobApi.PrepareJobRequest, StreamObserver<JobApi.PrepareJobResponse>) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService
 
prepare(JobApi.PrepareJobRequest, StreamObserver<JobApi.PrepareJobResponse>) - Method in class org.apache.beam.runners.portability.testing.TestJobService
 
prepareCall(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
prepareCall(String, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
prepareCall(String, int, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
prepareFilesToStage(SparkPipelineOptions) - Static method in interface org.apache.beam.runners.spark.SparkPipelineOptions
Classpath contains non jar files (eg.
prepareFilesToStageForRemoteClusterExecution(SparkStructuredStreamingPipelineOptions) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
Local configurations work in the same JVM and have no problems with improperly formatted files on classpath (eg.
prepareForProcessing() - Method in class org.apache.beam.sdk.transforms.DoFn
Deprecated.
use DoFn.Setup or DoFn.StartBundle instead. This method will be removed in a future release.
prepareStatement(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
prepareStatement(String, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
prepareStatement(String, int, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
prepareStatement(String, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
prepareStatement(String, int[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
prepareStatement(String, String[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
prepareWrite(WritableByteChannel) - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
Called with the channel that a subclass will write its header, footer, and values to.
PrepareWrite<InputT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io.gcp.bigquery
Prepare an input PCollection for writing to BigQuery.
PrepareWrite(DynamicDestinations<InputT, DestinationT>, SerializableFunction<InputT, OutputT>) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.PrepareWrite
 
primary() - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
 
PrimitiveParDoSingleFactory<InputT,OutputT> - Class in org.apache.beam.runners.dataflow
A PTransformOverrideFactory that produces PrimitiveParDoSingleFactory.ParDoSingle instances from ParDo.SingleOutput instances.
PrimitiveParDoSingleFactory() - Constructor for class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory
 
PrimitiveParDoSingleFactory.ParDoSingle<InputT,OutputT> - Class in org.apache.beam.runners.dataflow
A single-output primitive ParDo.
PrimitiveParDoSingleFactory.PayloadTranslator - Class in org.apache.beam.runners.dataflow
PrimitiveParDoSingleFactory.Registrar - Class in org.apache.beam.runners.dataflow
printDatasetContent(Dataset<WindowedValue>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
printHelp(PrintStream) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
Outputs the set of registered options with the PipelineOptionsFactory with a description for each one if available to the output stream.
printHelp(PrintStream, Class<? extends PipelineOptions>) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
Outputs the set of options available to be set for the passed in PipelineOptions interface.
process(List<JobMessage>) - Method in interface org.apache.beam.runners.dataflow.util.MonitoringUtil.JobMessagesHandler
Process the rows.
process(List<JobMessage>) - Method in class org.apache.beam.runners.dataflow.util.MonitoringUtil.LoggingHandler
 
processBundle(Iterable<? extends InputT>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
processBundle(InputT...) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
ProcessBundleDescriptors - Class in org.apache.beam.runners.fnexecution.control
Utility methods for creating BeamFnApi.ProcessBundleDescriptor instances.
ProcessBundleDescriptors() - Constructor for class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors
 
ProcessBundleDescriptors.BagUserStateSpec<K,V,W extends BoundedWindow> - Class in org.apache.beam.runners.fnexecution.control
A container type storing references to the key, value, and window Coder used when handling bag user state requests.
ProcessBundleDescriptors.ExecutableProcessBundleDescriptor - Class in org.apache.beam.runners.fnexecution.control
 
ProcessBundleDescriptors.SideInputSpec<T,W extends BoundedWindow> - Class in org.apache.beam.runners.fnexecution.control
A container type storing references to the value, and window Coder used when handling side input state requests.
ProcessBundleDescriptors.TimerSpec<K,V,W extends BoundedWindow> - Class in org.apache.beam.runners.fnexecution.control
A container type storing references to the key, timer and payload coders and the remote input destination used when handling timer requests.
ProcessContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.ProcessContext
 
ProcessContinuation() - Constructor for class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
 
processElement(DoFn<KV<K, Iterable<KV<Instant, WindowedValue<KV<K, V>>>>>, OutputT>.ProcessContext, BoundedWindow) - Method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides.BatchStatefulDoFn
 
processElement(DoFn<Iterable<T>, T>.ProcessContext, BoundedWindow) - Method in class org.apache.beam.runners.dataflow.DataflowRunner.StreamingPCollectionViewWriterFn
 
processElement(WindowedValue<InputT>) - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
 
processElement(DoFn<T, List<VideoAnnotationResults>>.ProcessContext) - Method in class org.apache.beam.sdk.extensions.ml.AnnotateVideo
Process element implementation required.
processElement(DoFn<KV<ByteString, VideoContext>, List<VideoAnnotationResults>>.ProcessContext) - Method in class org.apache.beam.sdk.extensions.ml.VideoIntelligence.AnnotateVideoBytesWithContext
ProcessElement implementation.
processElement(DoFn<ByteString, List<VideoAnnotationResults>>.ProcessContext) - Method in class org.apache.beam.sdk.extensions.ml.VideoIntelligence.AnnotateVideoFromBytes
Implementation of ProcessElement.
processElement(DoFn<String, List<VideoAnnotationResults>>.ProcessContext) - Method in class org.apache.beam.sdk.extensions.ml.VideoIntelligence.AnnotateVideoFromURI
ProcessElement implementation.
processElement(DoFn<KV<String, VideoContext>, List<VideoAnnotationResults>>.ProcessContext) - Method in class org.apache.beam.sdk.extensions.ml.VideoIntelligence.AnnotateVideoURIWithContext
ProcessElement implementation.
processElement(Row, DoFn.OutputReceiver<Row>) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSetOperatorsTransforms.SetOperatorFilteringDoFn
 
processElement(DoFn<Row, Void>.ProcessContext) - Method in class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSqlOutputToConsoleFn
 
processElement(DoFn<T, Void>.ProcessContext) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write.PubsubBoundedWriter
 
processElement(InputT) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
processElement(DoFn<ValueWithRecordId<T>, T>.ProcessContext) - Method in class org.apache.beam.sdk.values.ValueWithRecordId.StripIdsDoFn
 
processElementWithRunner(DoFnRunner<KV<?, ?>, OutputT>, WindowedValue<KV<?, ?>>) - Method in class org.apache.beam.runners.jet.processors.StatefulParDoP
 
ProcessEnvironment - Class in org.apache.beam.runners.fnexecution.environment
Environment for process-based execution.
ProcessEnvironmentFactory - Class in org.apache.beam.runners.fnexecution.environment
An EnvironmentFactory which forks processes based on the parameters in the Environment.
ProcessEnvironmentFactory.Provider - Class in org.apache.beam.runners.fnexecution.environment
Provider of ProcessEnvironmentFactory.
ProcessFunction<InputT,OutputT> - Interface in org.apache.beam.sdk.transforms
A function that computes an output value of type OutputT from an input value of type InputT and is Serializable.
ProcessingTimeEvent() - Constructor for class org.apache.beam.sdk.testing.TestStream.ProcessingTimeEvent
 
ProcessingTimePolicy() - Constructor for class org.apache.beam.sdk.io.kafka.TimestampPolicyFactory.ProcessingTimePolicy
 
ProcessingTimeWatermarkPolicy() - Constructor for class org.apache.beam.sdk.io.kinesis.WatermarkPolicyFactory.ProcessingTimeWatermarkPolicy
 
ProcessManager - Class in org.apache.beam.runners.fnexecution.environment
A simple process manager which forks processes and kills them if necessary.
ProcessManager.RunningProcess - Class in org.apache.beam.runners.fnexecution.environment
 
processTimestampedElement(TimestampedValue<InputT>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
ProcessWatcher(Process) - Constructor for class org.apache.beam.runners.apex.ApexYarnLauncher.ProcessWatcher
 
processWindowedElement(InputT, Instant, BoundedWindow) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
ProducerRecordCoder<K,V> - Class in org.apache.beam.sdk.io.kafka
ProducerRecordCoder(Coder<K>, Coder<V>) - Constructor for class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
 
productArity() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.DecodeUsingBeamCoder
 
productArity() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.EncodeUsingBeamCoder
 
productElement(int) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.DecodeUsingBeamCoder
 
productElement(int) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.EncoderHelpers.EncodeUsingBeamCoder
 
PROJECT_ID_REGEXP - Static variable in class org.apache.beam.runners.dataflow.DataflowRunner
Project IDs must contain lowercase letters, digits, or dashes.
projectPathFromId(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
 
projectPathFromPath(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
 
ProjectSupport - Enum in org.apache.beam.sdk.extensions.sql.meta
 
properties(JSONObject) - Method in class org.apache.beam.sdk.extensions.sql.meta.Table.Builder
 
properties() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
 
PROPERTY_BEAM_TEST_PIPELINE_OPTIONS - Static variable in class org.apache.beam.sdk.testing.TestPipeline
System property used to set TestPipelineOptions.
propertyName() - Method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
Returns the property name associated with this provider.
propertyName() - Method in class org.apache.beam.sdk.options.ValueProvider.RuntimeValueProvider
Returns the property name that corresponds to this provider.
PropertyNames - Class in org.apache.beam.runners.dataflow.util
Constant property names used by the SDK in CloudWorkflow specifications.
PropertyNames() - Constructor for class org.apache.beam.runners.dataflow.util.PropertyNames
 
ProtobufCoderProviderRegistrar - Class in org.apache.beam.sdk.extensions.protobuf
A CoderProviderRegistrar for standard types used with Google Protobuf.
ProtobufCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.extensions.protobuf.ProtobufCoderProviderRegistrar
 
ProtoCoder<T extends Message> - Class in org.apache.beam.sdk.extensions.protobuf
A Coder using Google Protocol Buffers binary format.
ProtoCoder(Class<T>, Set<Class<?>>) - Constructor for class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
Private constructor.
ProtoDomain - Class in org.apache.beam.sdk.extensions.protobuf
ProtoDomain is a container class for Protobuf descriptors.
ProtoDynamicMessageSchema<T> - Class in org.apache.beam.sdk.extensions.protobuf
 
ProtoMessageSchema - Class in org.apache.beam.sdk.extensions.protobuf
 
ProtoMessageSchema() - Constructor for class org.apache.beam.sdk.extensions.protobuf.ProtoMessageSchema
 
ProtoSchemaLogicalTypes - Class in org.apache.beam.sdk.extensions.protobuf
A set of Schema.LogicalType classes to represent protocol buffer types.
ProtoSchemaLogicalTypes() - Constructor for class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes
 
ProtoSchemaLogicalTypes.DurationConvert - Class in org.apache.beam.sdk.extensions.protobuf
 
ProtoSchemaLogicalTypes.Fixed32 - Class in org.apache.beam.sdk.extensions.protobuf
A Fixed32 type.
ProtoSchemaLogicalTypes.Fixed64 - Class in org.apache.beam.sdk.extensions.protobuf
A Fixed64 type.
ProtoSchemaLogicalTypes.SFixed32 - Class in org.apache.beam.sdk.extensions.protobuf
A SFixed32 type.
ProtoSchemaLogicalTypes.SFixed64 - Class in org.apache.beam.sdk.extensions.protobuf
An SFixed64 type.
ProtoSchemaLogicalTypes.SInt32 - Class in org.apache.beam.sdk.extensions.protobuf
A SInt32 type.
ProtoSchemaLogicalTypes.SInt64 - Class in org.apache.beam.sdk.extensions.protobuf
A SIn64 type.
ProtoSchemaLogicalTypes.TimestampConvert - Class in org.apache.beam.sdk.extensions.protobuf
 
ProtoSchemaLogicalTypes.UInt32 - Class in org.apache.beam.sdk.extensions.protobuf
A UInt32 type.
ProtoSchemaLogicalTypes.UInt64 - Class in org.apache.beam.sdk.extensions.protobuf
A UIn64 type.
Provider(PipelineOptions) - Constructor for class org.apache.beam.runners.fnexecution.environment.DockerEnvironmentFactory.Provider
 
Provider(PipelineOptions) - Constructor for class org.apache.beam.runners.fnexecution.environment.EmbeddedEnvironmentFactory.Provider
 
Provider() - Constructor for class org.apache.beam.runners.fnexecution.environment.ExternalEnvironmentFactory.Provider
 
Provider(PipelineOptions) - Constructor for class org.apache.beam.runners.fnexecution.environment.ProcessEnvironmentFactory.Provider
 
Provider(InstructionRequestHandler) - Constructor for class org.apache.beam.runners.fnexecution.environment.StaticRemoteEnvironmentFactory.Provider
 
PTransform<InputT extends PInput,OutputT extends POutput> - Class in org.apache.beam.sdk.transforms
A PTransform<InputT, OutputT> is an operation that takes an InputT (some subtype of PInput) and produces an OutputT (some subtype of POutput).
PTransform() - Constructor for class org.apache.beam.sdk.transforms.PTransform
 
PTransform(String) - Constructor for class org.apache.beam.sdk.transforms.PTransform
 
publish(PubsubClient.TopicPath, List<PubsubClient.OutgoingMessage>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
Publish outgoingMessages to Pubsub topic.
publish(PubsubClient.TopicPath, List<PubsubClient.OutgoingMessage>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
 
publish(PubsubClient.TopicPath, List<PubsubClient.OutgoingMessage>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
 
publish(PubsubClient.TopicPath, List<PubsubClient.OutgoingMessage>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
publish(List<PubsubMessage>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
Publish messages to TestPubsub.topicPath().
PublishResponseCoder - Class in org.apache.beam.sdk.io.aws2.sns
Custom Coder for handling publish result.
PublishResultCoders - Class in org.apache.beam.sdk.io.aws.sns
Coders for SNS PublishResult.
PUBSUB_ID_ATTRIBUTE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
PUBSUB_SERIALIZED_ATTRIBUTES_FN - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
PUBSUB_SUBSCRIPTION - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
PUBSUB_SUBSCRIPTION_OVERRIDE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
PUBSUB_TIMESTAMP_ATTRIBUTE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
PUBSUB_TOPIC - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
PUBSUB_TOPIC_OVERRIDE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
PubsubClient - Class in org.apache.beam.sdk.io.gcp.pubsub
An (abstract) helper class for talking to Pubsub via an underlying transport.
PubsubClient() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
 
PubsubClient.OutgoingMessage - Class in org.apache.beam.sdk.io.gcp.pubsub
A message to be sent to Pubsub.
PubsubClient.ProjectPath - Class in org.apache.beam.sdk.io.gcp.pubsub
Path representing a cloud project id.
PubsubClient.PubsubClientFactory - Interface in org.apache.beam.sdk.io.gcp.pubsub
Factory for creating clients.
PubsubClient.SubscriptionPath - Class in org.apache.beam.sdk.io.gcp.pubsub
Path representing a Pubsub subscription.
PubsubClient.TopicPath - Class in org.apache.beam.sdk.io.gcp.pubsub
Path representing a Pubsub topic.
PubsubCoderProviderRegistrar - Class in org.apache.beam.sdk.io.gcp.pubsub
A CoderProviderRegistrar for standard types used with PubsubIO.
PubsubCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubCoderProviderRegistrar
 
PubsubGrpcClient - Class in org.apache.beam.sdk.io.gcp.pubsub
A helper class for talking to Pubsub via grpc.
PubsubIO - Class in org.apache.beam.sdk.io.gcp.pubsub
Read and Write PTransforms for Cloud Pub/Sub streams.
PubsubIO.PubsubSubscription - Class in org.apache.beam.sdk.io.gcp.pubsub
Class representing a Cloud Pub/Sub Subscription.
PubsubIO.PubsubTopic - Class in org.apache.beam.sdk.io.gcp.pubsub
Class representing a Cloud Pub/Sub Topic.
PubsubIO.Read<T> - Class in org.apache.beam.sdk.io.gcp.pubsub
Implementation of read methods.
PubsubIO.Write<T> - Class in org.apache.beam.sdk.io.gcp.pubsub
Implementation of write methods.
PubsubIO.Write.PubsubBoundedWriter - Class in org.apache.beam.sdk.io.gcp.pubsub
Writer to Pubsub which batches messages from bounded collections.
PubsubIOTableConfiguration() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonTableProvider.PubsubIOTableConfiguration
 
PubsubJsonClient - Class in org.apache.beam.sdk.io.gcp.pubsub
A Pubsub client using JSON transport.
PubsubJsonTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.pubsub
TableProvider for PubsubIOJsonTable which wraps PubsubIO for consumption by Beam SQL.
PubsubJsonTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonTableProvider
 
PubsubJsonTableProvider.PubsubIOTableConfiguration - Class in org.apache.beam.sdk.extensions.sql.meta.provider.pubsub
 
PubsubMessage - Class in org.apache.beam.sdk.io.gcp.pubsub
Class representing a Pub/Sub message.
PubsubMessage(byte[], Map<String, String>) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
 
PubsubMessage(byte[], Map<String, String>, String) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
 
PubsubMessagePayloadOnlyCoder - Class in org.apache.beam.sdk.io.gcp.pubsub
A coder for PubsubMessage treating the raw bytes being decoded as the message's payload.
PubsubMessagePayloadOnlyCoder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessagePayloadOnlyCoder
 
PubsubMessageToRow - Class in org.apache.beam.sdk.extensions.sql.meta.provider.pubsub
Read side converter for PubsubMessage with JSON payload.
PubsubMessageToRow() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubMessageToRow
 
PubsubMessageWithAttributesAndMessageIdCoder - Class in org.apache.beam.sdk.io.gcp.pubsub
A coder for PubsubMessage including attributes and the message id from the PubSub server.
PubsubMessageWithAttributesAndMessageIdCoder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesAndMessageIdCoder
 
PubsubMessageWithAttributesCoder - Class in org.apache.beam.sdk.io.gcp.pubsub
A coder for PubsubMessage including attributes.
PubsubMessageWithAttributesCoder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithAttributesCoder
 
PubsubMessageWithMessageIdCoder - Class in org.apache.beam.sdk.io.gcp.pubsub
A coder for PubsubMessage treating the raw bytes being decoded as the message's payload, with the message id from the PubSub server.
PubsubMessageWithMessageIdCoder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessageWithMessageIdCoder
 
PubsubOptions - Interface in org.apache.beam.sdk.io.gcp.pubsub
Properties that can be set when using Google Cloud Pub/Sub with the Apache Beam SDK.
PubsubTestClient - Class in org.apache.beam.sdk.io.gcp.pubsub
A (partial) implementation of PubsubClient for use by unit tests.
PubsubTestClient() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
PubsubTestClient.PubsubTestClientFactory - Interface in org.apache.beam.sdk.io.gcp.pubsub
Closing the factory will validate all expected messages were processed.
PubsubUnboundedSink - Class in org.apache.beam.sdk.io.gcp.pubsub
A PTransform which streams messages to Pubsub.
PubsubUnboundedSink(PubsubClient.PubsubClientFactory, ValueProvider<PubsubClient.TopicPath>, String, String, int) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
 
PubsubUnboundedSink(PubsubClient.PubsubClientFactory, ValueProvider<PubsubClient.TopicPath>, String, String, int, int, int) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink
 
PubsubUnboundedSource - Class in org.apache.beam.sdk.io.gcp.pubsub
Users should use PubsubIO#read instead.
PubsubUnboundedSource(PubsubClient.PubsubClientFactory, ValueProvider<PubsubClient.ProjectPath>, ValueProvider<PubsubClient.TopicPath>, ValueProvider<PubsubClient.SubscriptionPath>, String, String, boolean) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
Construct an unbounded source to consume from the Pubsub subscription.
PubsubUnboundedSource(Clock, PubsubClient.PubsubClientFactory, ValueProvider<PubsubClient.ProjectPath>, ValueProvider<PubsubClient.TopicPath>, ValueProvider<PubsubClient.SubscriptionPath>, String, String, boolean) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
Construct an unbounded source to consume from the Pubsub subscription.
PubsubUnboundedSource(PubsubClient.PubsubClientFactory, ValueProvider<PubsubClient.ProjectPath>, ValueProvider<PubsubClient.TopicPath>, ValueProvider<PubsubClient.SubscriptionPath>, String, String, boolean, boolean) - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSource
Construct an unbounded source to consume from the Pubsub subscription.
pull(long, PubsubClient.SubscriptionPath, int, boolean) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
Request the next batch of up to batchSize messages from subscription.
pull(long, PubsubClient.SubscriptionPath, int, boolean) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubGrpcClient
 
pull(long, PubsubClient.SubscriptionPath, int, boolean) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubJsonClient
 
pull(long, PubsubClient.SubscriptionPath, int, boolean) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubTestClient
 
pull() - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
Pull up to 100 messages from TestPubsub.subscriptionPath().
pull(int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
Pull up to maxBatchSize messages from TestPubsub.subscriptionPath().
PUSH_DOWN_OPTION - Static variable in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
 
put(String, InstructionRequestHandler) - Method in interface org.apache.beam.runners.fnexecution.control.ControlClientPool.Sink
Puts an InstructionRequestHandler into a client pool.
put(K2, V2) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
put(K, V) - Method in interface org.apache.beam.sdk.state.MapState
Associates the specified value with the specified key in this state.
putAll(Map<? extends K2, ? extends V2>) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
putArtifact(StreamObserver<ArtifactApi.PutArtifactResponse>) - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactStagingService
 
putDataset(PValue, Dataset<WindowedValue<T>>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
putDatasetWildcard(PValue, Dataset<WindowedValue<?>>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
TODO: All these 3 methods (putDataset*) are temporary and they are used only for generics type checking.
putIfAbsent(K, V) - Method in interface org.apache.beam.sdk.state.MapState
A deferred read-followed-by-write.
PValue - Interface in org.apache.beam.sdk.values
For internal use.
PValueBase - Class in org.apache.beam.sdk.values
For internal use.
PValueBase(Pipeline) - Constructor for class org.apache.beam.sdk.values.PValueBase
 
PValueBase() - Constructor for class org.apache.beam.sdk.values.PValueBase
No-arg constructor to allow subclasses to implement Serializable.

Q

qualifiedComponent() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
qualifiedComponent() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifyComponentContext
 
QualifiedComponentContext(ParserRuleContext, int) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifiedComponentContext
 
Qualifier() - Constructor for class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier
 
qualifierList() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierListContext
 
qualifierList(int) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.ArrayQualifierListContext
 
qualifierList() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierListContext
 
qualifierList(int) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.MapQualifierListContext
 
qualifierList() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifiedComponentContext
 
qualifierList() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
QualifierListContext(ParserRuleContext, int) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifierListContext
 
QualifierListContext() - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifierListContext
 
QualifyComponentContext(FieldSpecifierNotationParser.DotExpressionComponentContext) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.QualifyComponentContext
 
query(String) - Static method in class org.apache.beam.sdk.extensions.sql.SqlTransform
Returns a SqlTransform representing an equivalent execution plan.
queryMetrics(MetricsFilter) - Method in class org.apache.beam.runners.jet.metrics.JetMetricResults
 
queryMetrics(MetricsFilter) - Method in class org.apache.beam.runners.portability.PortableMetrics
 
queryMetrics(MetricsFilter) - Method in class org.apache.beam.sdk.metrics.MetricResults
Query for all metric values that match a given filter.
QueryParameters() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters
 
QueryPlanner - Interface in org.apache.beam.sdk.extensions.sql.impl
An interface that planners should implement to convert sql statement to BeamRelNode or SqlNode.
QueryPlanner.QueryParameters - Class in org.apache.beam.sdk.extensions.sql.impl
 
QueryPlanner.QueryParameters.Kind - Enum in org.apache.beam.sdk.extensions.sql.impl
 
queryResultHasChecksum(String) - Static method in class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher
 
QueryStatementConverter - Class in org.apache.beam.sdk.extensions.sql.zetasql.translation
Converts a resolved Zeta SQL query represented by a tree to corresponding Calcite representation.
QueryTrait - Class in org.apache.beam.sdk.extensions.sql.zetasql
QueryTrait.
QueryTrait() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
 
queryUnflattened(String, String, boolean) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
Performs a query without flattening results.
queryWithRetries(String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
 
queryWithRetries(String, String, boolean) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
 
queryWithRetriesUsingStandardSql(String, String) - Method in class org.apache.beam.sdk.io.gcp.testing.BigqueryClient
 
quoteIdentifier(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamBigQuerySqlDialect
 

R

RandomAccessData - Class in org.apache.beam.runners.dataflow.util
An elastic-sized byte array which allows you to manipulate it as a stream, or access it directly.
RandomAccessData() - Constructor for class org.apache.beam.runners.dataflow.util.RandomAccessData
Constructs a RandomAccessData with a default buffer size.
RandomAccessData(byte[]) - Constructor for class org.apache.beam.runners.dataflow.util.RandomAccessData
Constructs a RandomAccessData with the initial buffer.
RandomAccessData(int) - Constructor for class org.apache.beam.runners.dataflow.util.RandomAccessData
Constructs a RandomAccessData with the given buffer size.
RandomAccessData.RandomAccessDataCoder - Class in org.apache.beam.runners.dataflow.util
A Coder which encodes the valid parts of this stream.
RandomAccessData.UnsignedLexicographicalComparator - Class in org.apache.beam.runners.dataflow.util
A Comparator that compares two byte arrays lexicographically.
RandomAccessDataCoder() - Constructor for class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
 
RangeTracker<PositionT> - Interface in org.apache.beam.sdk.io.range
A RangeTracker is a thread-safe helper object for implementing dynamic work rebalancing in position-based BoundedSource.BoundedReader subclasses.
RateLimitPolicy - Interface in org.apache.beam.sdk.io.kinesis
 
RateLimitPolicyFactory - Interface in org.apache.beam.sdk.io.kinesis
Implement this interface to create a RateLimitPolicy.
RateLimitPolicyFactory.DelayIntervalRateLimiter - Class in org.apache.beam.sdk.io.kinesis
 
RawUnionValue(String, Object) - Constructor for class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils.RawUnionValue
Constructs a partial union from the given union tag and value.
RawUnionValue - Class in org.apache.beam.sdk.transforms.join
This corresponds to an integer union tag and value.
RawUnionValue(int, Object) - Constructor for class org.apache.beam.sdk.transforms.join.RawUnionValue
Constructs a partial union from the given union tag and value.
read() - Method in class org.apache.beam.runners.gearpump.translators.io.GearpumpSource
 
read(JavaStreamingContext, SerializablePipelineOptions, UnboundedSource<T, CheckpointMarkT>, String) - Static method in class org.apache.beam.runners.spark.io.SparkUnboundedSource
 
read(T) - Method in interface org.apache.beam.sdk.fn.stream.DataStreams.OutputChunkConsumer
 
read() - Static method in class org.apache.beam.sdk.io.amqp.AmqpIO
 
Read() - Constructor for class org.apache.beam.sdk.io.amqp.AmqpIO.Read
 
read(Class<T>) - Static method in class org.apache.beam.sdk.io.AvroIO
Reads records of the given type from an Avro file (or multiple Avro files matching a pattern).
Read() - Constructor for class org.apache.beam.sdk.io.AvroIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO
 
Read() - Constructor for class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.aws.sqs.SqsIO
 
Read() - Constructor for class org.apache.beam.sdk.io.aws.sqs.SqsIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO
 
Read() - Constructor for class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO
 
Read() - Constructor for class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.cassandra.CassandraIO
Provide a CassandraIO.Read PTransform to read data from a Cassandra database.
Read() - Constructor for class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO
 
Read() - Constructor for class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
read(SerializableFunction<SchemaAndRecord, T>) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
Reads from a BigQuery table or query and returns a PCollection with one element per each row of the table or query result, parsed from the BigQuery AVRO format using the specified function.
read() - Static method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO
Creates an uninitialized BigtableIO.Read.
Read() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
 
read() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1
Returns an empty DatastoreV1.Read builder.
Read() - Constructor for class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
 
Read() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO
Creates an uninitialized instance of SpannerIO.Read.
Read() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
Creates an uninitialized HadoopFormatIO.Read.
Read() - Constructor for class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.hbase.HBaseIO
Creates an uninitialized HBaseIO.Read.
read() - Static method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO
Read data from Hive.
Read() - Constructor for class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
Read data from a JDBC datasource.
Read() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.jms.JmsIO
 
Read() - Constructor for class org.apache.beam.sdk.io.jms.JmsIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.kafka.KafkaIO
Creates an uninitialized KafkaIO.Read PTransform.
Read() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.kinesis.KinesisIO
Returns a new KinesisIO.Read transform for reading from Kinesis.
Read() - Constructor for class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO
Read data from GridFS.
Read() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbIO
Read data from MongoDB.
Read() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.mqtt.MqttIO
 
Read() - Constructor for class org.apache.beam.sdk.io.mqtt.MqttIO.Read
 
read(Schema) - Static method in class org.apache.beam.sdk.io.parquet.ParquetIO
Reads GenericRecord from a Parquet file (or multiple Parquet files matching the pattern).
Read() - Constructor for class org.apache.beam.sdk.io.parquet.ParquetIO.Read
 
Read - Class in org.apache.beam.sdk.io
A PTransform for reading from a Source.
Read() - Constructor for class org.apache.beam.sdk.io.Read
 
read() - Static method in class org.apache.beam.sdk.io.redis.RedisIO
Read data from a Redis server.
Read() - Constructor for class org.apache.beam.sdk.io.redis.RedisIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.solr.SolrIO
 
Read() - Constructor for class org.apache.beam.sdk.io.solr.SolrIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.TextIO
A PTransform that reads from one or more text files and returns a bounded PCollection containing one element for each line of the input files.
Read() - Constructor for class org.apache.beam.sdk.io.TextIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.TFRecordIO
A PTransform that reads from a TFRecord file (or multiple TFRecord files matching a pattern) and returns a PCollection containing the decoding of each of the records of the TFRecord file(s) as a byte array.
Read() - Constructor for class org.apache.beam.sdk.io.TFRecordIO.Read
 
read() - Static method in class org.apache.beam.sdk.io.xml.XmlIO
Reads XML files as a PCollection of a given type mapped via JAXB.
Read() - Constructor for class org.apache.beam.sdk.io.xml.XmlIO.Read
 
read() - Method in interface org.apache.beam.sdk.state.BagState
 
read() - Method in interface org.apache.beam.sdk.state.CombiningState
 
read() - Method in interface org.apache.beam.sdk.state.ReadableState
Read the current value, blocking until it is available.
Read.Bounded<T> - Class in org.apache.beam.sdk.io
PTransform that reads from a BoundedSource.
Read.Builder - Class in org.apache.beam.sdk.io
Helper class for building Read transforms.
Read.Unbounded<T> - Class in org.apache.beam.sdk.io
PTransform that reads from a UnboundedSource.
ReadableFileCoder - Class in org.apache.beam.sdk.io
ReadableFileCoder() - Constructor for class org.apache.beam.sdk.io.ReadableFileCoder
 
ReadableState<T> - Interface in org.apache.beam.sdk.state
A State that can be read via ReadableState.read().
ReadableStates - Class in org.apache.beam.sdk.state
For internal use only; no backwards-compatibility guarantees.
ReadableStates() - Constructor for class org.apache.beam.sdk.state.ReadableStates
 
readAll(Class<T>) - Static method in class org.apache.beam.sdk.io.AvroIO
Deprecated.
You can achieve The functionality of AvroIO.readAll(java.lang.Class<T>) using FileIO matching plus AvroIO.readFiles(Class). This is the preferred method to make composition explicit. AvroIO.ReadAll will not receive upgrades and will be removed in a future version of Beam.
ReadAll() - Constructor for class org.apache.beam.sdk.io.AvroIO.ReadAll
Deprecated.
 
readAll() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO
A PTransform that works like SpannerIO.read(), but executes read operations coming from a PCollection.
ReadAll() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
 
readAll() - Static method in class org.apache.beam.sdk.io.hbase.HBaseIO
A PTransform that works like HBaseIO.read(), but executes read operations coming from a PCollection of HBaseIO.Read.
readAll() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
Like JdbcIO.read(), but executes multiple instances of the query substituting each element of a PCollection as query parameters.
ReadAll() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
 
readAll() - Static method in class org.apache.beam.sdk.io.redis.RedisIO
Like RedisIO.read() but executes multiple instances of the Redis query substituting each element of a PCollection as key pattern.
ReadAll() - Constructor for class org.apache.beam.sdk.io.redis.RedisIO.ReadAll
 
readAll() - Static method in class org.apache.beam.sdk.io.TextIO
Deprecated.
You can achieve The functionality of TextIO.readAll() using FileIO matching plus TextIO.readFiles(). This is the preferred method to make composition explicit. TextIO.ReadAll will not receive upgrades and will be removed in a future version of Beam.
ReadAll() - Constructor for class org.apache.beam.sdk.io.TextIO.ReadAll
Deprecated.
 
readAllGenericRecords(Schema) - Static method in class org.apache.beam.sdk.io.AvroIO
Deprecated.
You can achieve The functionality of AvroIO.readAllGenericRecords(Schema) using FileIO matching plus AvroIO.readFilesGenericRecords(Schema). This is the preferred method to make composition explicit. AvroIO.ReadAll will not receive upgrades and will be removed in a future version of Beam.
readAllGenericRecords(String) - Static method in class org.apache.beam.sdk.io.AvroIO
Deprecated.
You can achieve The functionality of AvroIO.readAllGenericRecords(String) using FileIO matching plus AvroIO.readFilesGenericRecords(String). This is the preferred method to make composition explicit. AvroIO.ReadAll will not receive upgrades and will be removed in a future version of Beam.
ReadAllViaFileBasedSource<T> - Class in org.apache.beam.sdk.io
Reads each file in the input PCollection of FileIO.ReadableFile using given parameters for splitting files into offset ranges and for creating a FileBasedSource for a file.
ReadAllViaFileBasedSource(long, SerializableFunction<String, ? extends FileBasedSource<T>>, Coder<T>) - Constructor for class org.apache.beam.sdk.io.ReadAllViaFileBasedSource
 
readAvroGenericRecords(Schema) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
Returns a PTransform that continuously reads binary encoded Avro messages into the Avro GenericRecord type.
readAvros(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
Returns A PTransform that continuously reads binary encoded Avro messages of the given type from a Google Cloud Pub/Sub stream.
readAvrosWithBeamSchema(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
Returns a PTransform that continuously reads binary encoded Avro messages of the specific type.
ReadBoundedTranslator<T> - Class in org.apache.beam.runners.gearpump.translators
Read.Bounded is translated to Gearpump source function and BoundedSource is wrapped into Gearpump DataSource.
ReadBoundedTranslator() - Constructor for class org.apache.beam.runners.gearpump.translators.ReadBoundedTranslator
 
ReadBuilder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.ReadBuilder
 
readBytes() - Static method in class org.apache.beam.sdk.io.kafka.KafkaIO
A specific instance of uninitialized KafkaIO.read() where key and values are bytes.
readDecompressed(ReadableByteChannel) - Method in enum org.apache.beam.sdk.io.Compression
 
Reader() - Constructor for class org.apache.beam.sdk.io.Source.Reader
 
ReaderInvocationUtil<OutputT,ReaderT extends Source.Reader<OutputT>> - Class in org.apache.beam.runners.flink.metrics
Util for invoking Source.Reader methods that might require a MetricsContainerImpl to be active.
ReaderInvocationUtil(String, PipelineOptions, FlinkMetricContainer) - Constructor for class org.apache.beam.runners.flink.metrics.ReaderInvocationUtil
 
readExternal(ObjectInput) - Method in class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
 
readFiles(Class<T>) - Static method in class org.apache.beam.sdk.io.AvroIO
ReadFiles() - Constructor for class org.apache.beam.sdk.io.AvroIO.ReadFiles
 
readFiles(Schema) - Static method in class org.apache.beam.sdk.io.parquet.ParquetIO
Like ParquetIO.read(Schema), but reads each file in a PCollection of FileIO.ReadableFile, which allows more flexible usage.
ReadFiles() - Constructor for class org.apache.beam.sdk.io.parquet.ParquetIO.ReadFiles
 
readFiles() - Static method in class org.apache.beam.sdk.io.TextIO
Like TextIO.read(), but reads each file in a PCollection of FileIO.ReadableFile, returned by FileIO.readMatches().
ReadFiles() - Constructor for class org.apache.beam.sdk.io.TextIO.ReadFiles
 
readFiles() - Static method in class org.apache.beam.sdk.io.TFRecordIO
Like TFRecordIO.read(), but reads each file in a PCollection of FileIO.ReadableFile, returned by FileIO.readMatches().
ReadFiles() - Constructor for class org.apache.beam.sdk.io.TFRecordIO.ReadFiles
 
readFiles(Class<T>) - Static method in class org.apache.beam.sdk.io.thrift.ThriftIO
Reads each file in a PCollection of FileIO.ReadableFile, which allows more flexible usage.
ReadFiles() - Constructor for class org.apache.beam.sdk.io.thrift.ThriftIO.ReadFiles
 
readFiles() - Static method in class org.apache.beam.sdk.io.xml.XmlIO
Like XmlIO.read(), but reads each file in a PCollection of FileIO.ReadableFile, which allows more flexible usage via different configuration options of FileIO.match() and FileIO.readMatches() that are not explicitly provided for XmlIO.read().
ReadFiles() - Constructor for class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
 
readFilesGenericRecords(Schema) - Static method in class org.apache.beam.sdk.io.AvroIO
readFilesGenericRecords(String) - Static method in class org.apache.beam.sdk.io.AvroIO
readFrom(InputStream, int, int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
Reads length bytes from the specified input stream writing them into the backing data store starting at offset.
readFromPort(BeamFnApi.RemoteGrpcPort, String) - Static method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortRead
 
readFromSource(BoundedSource<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
Reads all elements from the given BoundedSource.
readFromSplitsOfSource(BoundedSource<T>, long, PipelineOptions) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
 
readFromStartedReader(Source.Reader<T>) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
Reads all elements from the given started Source.Reader.
readFromUnstartedReader(Source.Reader<T>) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
Reads all elements from the given unstarted Source.Reader.
readFullyAsBytes() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
Returns the full contents of the file as bytes.
readFullyAsUTF8String() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
Returns the full contents of the file as a String decoded as UTF-8.
readGenericRecords(Schema) - Static method in class org.apache.beam.sdk.io.AvroIO
Reads Avro file(s) containing records of the specified schema.
readGenericRecords(String) - Static method in class org.apache.beam.sdk.io.AvroIO
Reads Avro file(s) containing records of the specified schema.
readLater() - Method in interface org.apache.beam.sdk.state.BagState
 
readLater() - Method in interface org.apache.beam.sdk.state.CombiningState
 
readLater() - Method in interface org.apache.beam.sdk.state.GroupingState
 
readLater() - Method in interface org.apache.beam.sdk.state.ReadableState
Indicate that the value will be read later.
readLater() - Method in interface org.apache.beam.sdk.state.SetState
 
readLater() - Method in interface org.apache.beam.sdk.state.ValueState
 
readLater() - Method in interface org.apache.beam.sdk.state.WatermarkHoldState
 
readMatches() - Static method in class org.apache.beam.sdk.io.FileIO
Converts each result of FileIO.match() or FileIO.matchAll() to a FileIO.ReadableFile which can be used to read the contents of each file, optionally decompressing it.
ReadMatches() - Constructor for class org.apache.beam.sdk.io.FileIO.ReadMatches
 
readMessage() - Static method in class org.apache.beam.sdk.io.jms.JmsIO
 
readMessages() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
Returns A PTransform that continuously reads from a Google Cloud Pub/Sub stream.
readMessagesWithAttributes() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
Returns A PTransform that continuously reads from a Google Cloud Pub/Sub stream.
readMessagesWithAttributesAndMessageId() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
Returns A PTransform that continuously reads from a Google Cloud Pub/Sub stream.
readMessagesWithCoderAndParseFn(Coder<T>, SimpleFunction<PubsubMessage, T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
Returns A PTransform that continuously reads from a Google Cloud Pub/Sub stream, mapping each PubsubMessage into type T using the supplied parse function and coder.
readMessagesWithMessageId() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
Returns A PTransform that continuously reads from a Google Cloud Pub/Sub stream.
readNextBlock() - Method in class org.apache.beam.sdk.io.AvroSource.AvroReader
 
readNextBlock() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
Read the next block from the input.
readNextRecord() - Method in class org.apache.beam.sdk.io.BlockBasedSource.Block
Reads the next record from the block and returns true iff one exists.
readNextRecord() - Method in class org.apache.beam.sdk.io.BlockBasedSource.BlockBasedReader
Reads the next record from the current block if possible.
readNextRecord() - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
Reads the next record via the delegate reader.
readNextRecord() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
readNItemsFromStartedReader(Source.Reader<T>, int) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
Read elements from a Source.Reader that has already had Source.Reader#start called on it, until n elements are read.
readNItemsFromUnstartedReader(Source.Reader<T>, int) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
Read elements from a Source.Reader until n elements are read.
readOnly(String, Map<String, BeamSqlTable>) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
This method creates BeamSqlEnv using empty Pipeline Options.
ReadOnlyTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider
A ReadOnlyTableProvider provides in-memory read only set of BeamSqlTable BeamSqlTables.
ReadOnlyTableProvider(String, Map<String, BeamSqlTable>) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.ReadOnlyTableProvider
 
ReadOperation - Class in org.apache.beam.sdk.io.gcp.spanner
Encapsulates a spanner read operation.
ReadOperation() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
readProtos(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
Returns A PTransform that continuously reads binary encoded protobuf messages of the given type from a Google Cloud Pub/Sub stream.
readRemainingFromReader(Source.Reader<T>, boolean) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
Read all remaining elements from a Source.Reader.
readRows(Storage.ReadRowsRequest) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StorageClient
Read rows in the context of a specific read stream.
readRows() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
Read Beam Rows from a JDBC data source.
ReadRows() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
 
readStrings() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
Returns A PTransform that continuously reads UTF-8 encoded strings from a Google Cloud Pub/Sub stream.
readTableRows() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
Like BigQueryIO.read(SerializableFunction) but represents each row as a TableRow.
readTableRowsWithSchema() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
Like BigQueryIO.readTableRows() but with Schema support.
ReadUnboundedTranslator<T> - Class in org.apache.beam.runners.gearpump.translators
Read.Unbounded is translated to Gearpump source function and UnboundedSource is wrapped into Gearpump DataSource.
ReadUnboundedTranslator() - Constructor for class org.apache.beam.runners.gearpump.translators.ReadUnboundedTranslator
 
readValue - Variable in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForGetter
 
readValue - Variable in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.ConvertValueForSetter
 
receive(LogicalEndpoint, Coder<T>, FnDataReceiver<T>) - Method in interface org.apache.beam.runners.fnexecution.data.FnDataService
Registers a receiver to be notified upon any incoming elements.
receive(LogicalEndpoint, Coder<T>, FnDataReceiver<T>) - Method in class org.apache.beam.runners.fnexecution.data.GrpcDataService
 
recordId() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.OutgoingMessage
If using an id attribute, the record id to associate with this record's metadata so the receiver can reject duplicates.
RedisConnectionConfiguration - Class in org.apache.beam.sdk.io.redis
RedisConnectionConfiguration describes and wraps a connectionConfiguration to Redis server or cluster.
RedisConnectionConfiguration() - Constructor for class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
 
RedisIO - Class in org.apache.beam.sdk.io.redis
An IO to manipulate Redis key/value database.
RedisIO.Read - Class in org.apache.beam.sdk.io.redis
Implementation of RedisIO.read().
RedisIO.ReadAll - Class in org.apache.beam.sdk.io.redis
Implementation of RedisIO.readAll().
RedisIO.Write - Class in org.apache.beam.sdk.io.redis
A PTransform to write to a Redis server.
RedisIO.Write.Method - Enum in org.apache.beam.sdk.io.redis
Determines the method used to insert data in Redis.
ReferenceCountingExecutableStageContextFactory - Class in org.apache.beam.runners.fnexecution.control
ExecutableStageContext.Factory which counts ExecutableStageContext reference for book keeping.
ReferenceCountingExecutableStageContextFactory.Creator - Interface in org.apache.beam.runners.fnexecution.control
Interface for creator which extends Serializable.
referencesSingleField() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Returns true if this descriptor references only a single, non-wildcard field.
ReflectUtils - Class in org.apache.beam.sdk.schemas.utils
A set of reflection helper methods.
ReflectUtils() - Constructor for class org.apache.beam.sdk.schemas.utils.ReflectUtils
 
ReflectUtils.ClassWithSchema - Class in org.apache.beam.sdk.schemas.utils
Represents a class and a schema.
refreshAll() - Method in class org.apache.beam.runners.direct.WatermarkManager
Refresh the watermarks contained within this WatermarkManager, causing all watermarks to be advanced as far as possible.
Regex - Class in org.apache.beam.sdk.transforms
PTransorms to use Regular Expressions to process elements in a PCollection.
Regex.AllMatches - Class in org.apache.beam.sdk.transforms
Regex.MatchesName<String> takes a PCollection<String> and returns a PCollection<List<String>> representing the value extracted from all the Regex groups of the input PCollection to the number of times that element occurs in the input.
Regex.Find - Class in org.apache.beam.sdk.transforms
Regex.Find<String> takes a PCollection<String> and returns a PCollection<String> representing the value extracted from the Regex groups of the input PCollection to the number of times that element occurs in the input.
Regex.FindAll - Class in org.apache.beam.sdk.transforms
Regex.Find<String> takes a PCollection<String> and returns a PCollection<List<String>> representing the value extracted from the Regex groups of the input PCollection to the number of times that element occurs in the input.
Regex.FindKV - Class in org.apache.beam.sdk.transforms
Regex.MatchesKV<KV<String, String>> takes a PCollection<String> and returns a PCollection<KV<String, String>> representing the key and value extracted from the Regex groups of the input PCollection to the number of times that element occurs in the input.
Regex.FindName - Class in org.apache.beam.sdk.transforms
Regex.Find<String> takes a PCollection<String> and returns a PCollection<String> representing the value extracted from the Regex groups of the input PCollection to the number of times that element occurs in the input.
Regex.FindNameKV - Class in org.apache.beam.sdk.transforms
Regex.MatchesKV<KV<String, String>> takes a PCollection<String> and returns a PCollection<KV<String, String>> representing the key and value extracted from the Regex groups of the input PCollection to the number of times that element occurs in the input.
Regex.Matches - Class in org.apache.beam.sdk.transforms
Regex.Matches<String> takes a PCollection<String> and returns a PCollection<String> representing the value extracted from the Regex groups of the input PCollection to the number of times that element occurs in the input.
Regex.MatchesKV - Class in org.apache.beam.sdk.transforms
Regex.MatchesKV<KV<String, String>> takes a PCollection<String> and returns a PCollection<KV<String, String>> representing the key and value extracted from the Regex groups of the input PCollection to the number of times that element occurs in the input.
Regex.MatchesName - Class in org.apache.beam.sdk.transforms
Regex.MatchesName<String> takes a PCollection<String> and returns a PCollection<String> representing the value extracted from the Regex groups of the input PCollection to the number of times that element occurs in the input.
Regex.MatchesNameKV - Class in org.apache.beam.sdk.transforms
Regex.MatchesNameKV<KV<String, String>> takes a PCollection<String> and returns a PCollection<KV<String, String>> representing the key and value extracted from the Regex groups of the input PCollection to the number of times that element occurs in the input.
Regex.ReplaceAll - Class in org.apache.beam.sdk.transforms
Regex.ReplaceAll<String> takes a PCollection<String> and returns a PCollection<String> with all Strings that matched the Regex being replaced with the replacement string.
Regex.ReplaceFirst - Class in org.apache.beam.sdk.transforms
Regex.ReplaceFirst<String> takes a PCollection<String> and returns a PCollection<String> with the first Strings that matched the Regex being replaced with the replacement string.
Regex.Split - Class in org.apache.beam.sdk.transforms
Regex.Split<String> takes a PCollection<String> and returns a PCollection<String> with the input string split into individual items in a list.
RegexMatcher - Class in org.apache.beam.sdk.testing
Hamcrest matcher to assert a string matches a pattern.
RegexMatcher(String) - Constructor for class org.apache.beam.sdk.testing.RegexMatcher
 
register(WatchService, WatchEvent.Kind<?>[], WatchEvent.Modifier...) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
register(WatchService, WatchEvent.Kind<?>...) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
register(RelOptPlanner) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamIOSinkRel
 
register(RelOptPlanner) - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
 
register(Class<? extends PipelineOptions>) - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
This registers the interface with this factory.
registerByteSizeObserver(T, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.Coder
Notifies the ElementByteSizeObserver about the byte size of the encoded value using this Coder.
registerByteSizeObserver(ReadableDuration, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.DurationCoder
 
registerByteSizeObserver(IterableT, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
 
registerByteSizeObserver(KV<K, V>, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.KvCoder
Notifies ElementByteSizeObserver about the byte size of the encoded value using this coder.
registerByteSizeObserver(Map<K, V>, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.MapCoder
 
registerByteSizeObserver(T, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.coders.NullableCoder
Overridden to short-circuit the default StructuredCoder behavior of encoding and counting the bytes.
registerByteSizeObserver(RawUnionValue, ElementByteSizeObserver) - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
Notifies ElementByteSizeObserver about the byte size of the encoded value using this coder.
registerClasses(Kryo) - Method in class org.apache.beam.runners.spark.coders.SparkRunnerKryoRegistrator
 
registerCoderForClass(Class<?>, Coder<?>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
Registers the provided Coder for the given class.
registerCoderForType(TypeDescriptor<?>, Coder<?>) - Method in class org.apache.beam.sdk.coders.CoderRegistry
Registers the provided Coder for the given type.
registerCoderProvider(CoderProvider) - Method in class org.apache.beam.sdk.coders.CoderRegistry
Registers coderProvider as a potential CoderProvider which can produce Coder instances.
registerConsumer(LogicalEndpoint, BiConsumer<ByteString, Boolean>) - Method in class org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer
 
registerForProcessBundleInstructionId(String, StateRequestHandler) - Method in class org.apache.beam.runners.fnexecution.state.GrpcStateService
 
registerForProcessBundleInstructionId(String, StateRequestHandler) - Method in interface org.apache.beam.runners.fnexecution.state.StateDelegator
Registers the supplied handler for the given process bundle instruction id for all BeamFnApi.StateRequests with a matching id.
registerJavaBean(Class<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Register a JavaBean type for automatic schema inference.
registerJavaBean(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Register a JavaBean type for automatic schema inference.
registerKnownTableNames(List<TableName>) - Method in interface org.apache.beam.sdk.extensions.sql.meta.CustomTableResolver
Register the table names as extracted from the FROM clause.
registerKnownTableNames(List<TableName>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.FullNameTableProvider
 
registerMetricsForPipelineResult() - Method in class org.apache.beam.runners.flink.metrics.FlinkMetricContainer
This should be called at the end of the Flink job and sets up an accumulator to push the metrics to the PipelineResult.
registerPOJO(Class<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Register a POJO type for automatic schema inference.
registerPOJO(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Register a POJO type for automatic schema inference.
registerProvider(TableProvider) - Method in class org.apache.beam.sdk.extensions.sql.meta.store.InMemoryMetaStore
 
registerProvider(TableProvider) - Method in interface org.apache.beam.sdk.extensions.sql.meta.store.MetaStore
Register a table provider.
registerSchemaForClass(Class<T>, Schema, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Register a schema for a specific Class type.
registerSchemaForType(TypeDescriptor<T>, Schema, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Register a schema for a specific TypeDescriptor type.
registerSchemaProvider(SchemaProvider) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Register a SchemaProvider.
registerSchemaProvider(Class<T>, SchemaProvider) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Register a SchemaProvider to be used for a specific type.
registerSchemaProvider(TypeDescriptor<T>, SchemaProvider) - Method in class org.apache.beam.sdk.schemas.SchemaRegistry
Register a SchemaProvider to be used for a specific type.
registerTables(SchemaPlus, List<List<String>>) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.TableResolution
Registers tables that will be resolved during query analysis, so table providers can eagerly pre-load metadata.
registerTransformTranslator(Class<TransformT>, TransformTranslator<? extends TransformT>) - Static method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator
Records that instances of the specified PTransform class should be translated by default by the corresponding TransformTranslator.
registerUdaf(String, Combine.CombineFn) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
register a Combine.CombineFn as UDAF function used in this query.
registerUdf(String, Class<? extends BeamSqlUdf>) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
register a UDF function used in this query.
registerUdf(String, SerializableFunction) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
Register SerializableFunction as a UDF function used in this query.
Registrar() - Constructor for class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.Registrar
 
Reify - Class in org.apache.beam.sdk.transforms
PTransforms for converting between explicit and implicit form of various Beam values.
ReifyAsIterable<T> - Class in org.apache.beam.sdk.io.gcp.bigquery
This transforms turns a side input into a singleton PCollection that can be used as the main input for another transform.
ReifyAsIterable() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.ReifyAsIterable
 
rel(int) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
 
rel(SqlNode) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl
 
rel(String, QueryPlanner.QueryParameters) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl
 
relativeErrorForPrecision(int) - Static method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct
 
relativeFileNaming(ValueProvider<String>, FileIO.Write.FileNaming) - Static method in class org.apache.beam.sdk.io.FileIO.Write
 
relativize(Path) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
releaseJobIdLock(Configuration) - Method in interface org.apache.beam.sdk.io.hadoop.format.ExternalSynchronization
Deletes lock ids bounded with given job if any exists.
releaseJobIdLock(Configuration) - Method in class org.apache.beam.sdk.io.hadoop.format.HDFSSynchronization
 
releaseSavepoint(Savepoint) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
RelMdNodeStats - Class in org.apache.beam.sdk.extensions.sql.impl.planner
This is the implementation of NodeStatsMetadata.
RelMdNodeStats() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.planner.RelMdNodeStats
 
remerge() - Static method in class org.apache.beam.sdk.transforms.windowing.Window
Creates a Window PTransform that does not change assigned windows, but will cause windows to be merged again as part of the next GroupByKey.
RemoteBundle - Interface in org.apache.beam.runners.fnexecution.control
A bundle capable of handling input data elements for a bundle descriptor by forwarding them to a remote environment for processing.
RemoteEnvironment - Interface in org.apache.beam.runners.fnexecution.environment
A handle to an available remote RunnerApi.Environment.
RemoteEnvironment.SimpleRemoteEnvironment - Class in org.apache.beam.runners.fnexecution.environment
A RemoteEnvironment which uses the default RemoteEnvironment.close() behavior.
RemoteEnvironmentOptions - Interface in org.apache.beam.sdk.options
Options that are used to control configuration of the remote environment.
RemoteEnvironmentOptions.Options - Class in org.apache.beam.sdk.options
RemoteGrpcPortRead - Class in org.apache.beam.sdk.fn.data
An execution-time only RunnerApi.PTransform which represents an SDK harness reading from a BeamFnApi.RemoteGrpcPort.
RemoteGrpcPortRead() - Constructor for class org.apache.beam.sdk.fn.data.RemoteGrpcPortRead
 
RemoteGrpcPortWrite - Class in org.apache.beam.sdk.fn.data
An execution-time only RunnerApi.PTransform which represents a write from within an SDK harness to a BeamFnApi.RemoteGrpcPort.
RemoteGrpcPortWrite() - Constructor for class org.apache.beam.sdk.fn.data.RemoteGrpcPortWrite
 
RemoteInputDestination<T> - Class in org.apache.beam.runners.fnexecution.data
A pair of Coder and BeamFnApi.Target which specifies the arguments to a FnDataService to send data to a remote harness.
RemoteInputDestination() - Constructor for class org.apache.beam.runners.fnexecution.data.RemoteInputDestination
 
RemoteOutputReceiver<T> - Class in org.apache.beam.runners.fnexecution.control
A pair of Coder and FnDataReceiver which can be registered to receive elements for a LogicalEndpoint.
RemoteOutputReceiver() - Constructor for class org.apache.beam.runners.fnexecution.control.RemoteOutputReceiver
 
remove(Collection<String>) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
 
remove() - Method in class org.apache.beam.sdk.fn.stream.DataStreams.BlockingQueueIterator
 
remove() - Method in class org.apache.beam.sdk.fn.stream.DataStreams.DataStreamDecoder
 
remove(Object) - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
remove(K) - Method in interface org.apache.beam.sdk.state.MapState
Remove the mapping for a key from this map if it is present.
remove(T) - Method in interface org.apache.beam.sdk.state.SetState
Removes the specified element from this set if it is present.
removeAllPipelineOptions() - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
removeArtifacts(String) - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactStagingService
 
removeArtifacts(String) - Method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService
 
removeArtifacts(String) - Method in class org.apache.beam.runners.fnexecution.artifact.JavaFilesystemLegacyArtifactStagingService
 
removePipelineOption(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
removePrefix() - Method in class org.apache.beam.sdk.extensions.sql.impl.TableName
Remove prefix, e.g.
removeTemporaryFiles(Collection<ResourceId>) - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
 
removeUri(String) - Method in class org.apache.beam.runners.fnexecution.artifact.AbstractLegacyArtifactStagingService
 
removeUri(String) - Method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService
 
removeUri(String) - Method in class org.apache.beam.runners.fnexecution.artifact.JavaFilesystemLegacyArtifactStagingService
 
rename(List<ResourceIdT>, List<ResourceIdT>) - Method in class org.apache.beam.sdk.io.FileSystem
Renames a List of file-like resources from one location to another.
rename(List<ResourceId>, List<ResourceId>, MoveOptions...) - Static method in class org.apache.beam.sdk.io.FileSystems
Renames a List of file-like resources from one location to another.
rename(String, String) - Method in class org.apache.beam.sdk.schemas.transforms.RenameFields.Inner
Rename a specific field.
rename(FieldAccessDescriptor, String) - Method in class org.apache.beam.sdk.schemas.transforms.RenameFields.Inner
Rename a specific field.
RenameFields - Class in org.apache.beam.sdk.schemas.transforms
A transform for renaming fields inside an existing schema.
RenameFields() - Constructor for class org.apache.beam.sdk.schemas.transforms.RenameFields
 
RenameFields.Inner<T> - Class in org.apache.beam.sdk.schemas.transforms
The class implementing the actual PTransform.
render() - Method in interface org.apache.beam.runners.spark.aggregators.NamedAggregators.State
 
render() - Method in interface org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregators.State
 
renderAll() - Method in class org.apache.beam.runners.spark.aggregators.NamedAggregators
 
renderAll() - Method in class org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregators
 
Repeatedly - Class in org.apache.beam.sdk.transforms.windowing
A Trigger that fires according to its subtrigger forever.
REPLACE - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
replace(String, String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
REPLACE_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamBuiltinMethods
 
replaceAll(List<PTransformOverride>) - Method in class org.apache.beam.sdk.Pipeline
For internal use only; no backwards-compatibility guarantees.
replaceAll(String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.ReplaceAll PTransform that checks if a portion of the line matches the Regex and replaces all matches with the replacement String.
replaceAll(Pattern, String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.ReplaceAll PTransform that checks if a portion of the line matches the Regex and replaces all matches with the replacement String.
ReplaceAll(Pattern, String) - Constructor for class org.apache.beam.sdk.transforms.Regex.ReplaceAll
 
replaceFirst(String, String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.ReplaceAll PTransform that checks if a portion of the line matches the Regex and replaces the first match with the replacement String.
replaceFirst(Pattern, String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.ReplaceAll PTransform that checks if a portion of the line matches the Regex and replaces the first match with the replacement String.
ReplaceFirst(Pattern, String) - Constructor for class org.apache.beam.sdk.transforms.Regex.ReplaceFirst
 
replaceTransforms(Pipeline) - Method in class org.apache.beam.runners.dataflow.DataflowRunner
 
replaceTransforms(Pipeline, StreamingOptions) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
 
reportElementSize(long) - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
 
reportWorkItemStatus(String, ReportWorkItemStatusRequest) - Method in class org.apache.beam.runners.dataflow.DataflowClient
Reports the status of the work item for jobId.
requestsFinalization(String) - Method in interface org.apache.beam.runners.fnexecution.control.BundleFinalizationHandler
This callback is invoked whenever an inflight bundle that is being processed requests finalization.
requestsFinalization(String) - Method in class org.apache.beam.runners.fnexecution.control.BundleFinalizationHandlers.InMemoryFinalizer
 
Requirements - Class in org.apache.beam.sdk.transforms
Describes the run-time requirements of a Contextful, such as access to side inputs.
requiresDeduping() - Method in class org.apache.beam.sdk.io.UnboundedSource
Returns whether this source requires explicit deduping.
requiresSideInputs(Collection<PCollectionView<?>>) - Static method in class org.apache.beam.sdk.transforms.Requirements
Describes the need for access to the given side inputs.
requiresSideInputs(PCollectionView<?>...) - Static method in class org.apache.beam.sdk.transforms.Requirements
reset() - Method in class org.apache.beam.runners.spark.aggregators.NamedAggregatorsAccumulator
 
reset() - Method in class org.apache.beam.runners.spark.metrics.MetricsContainerStepMapAccumulator
 
reset() - Method in class org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregatorsAccumulator
 
reset() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsContainerStepMapAccumulator
 
reset() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl
 
reset() - Method in class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
 
resetCache() - Static method in class org.apache.beam.sdk.options.PipelineOptionsFactory
Resets the set of interfaces registered with this factory to the default state.
resetForNewKey() - Method in class org.apache.beam.runners.fnexecution.state.InMemoryBagUserStateFactory
Prepares previous emitted state handlers for processing a new key.
resetLocal() - Method in class org.apache.beam.runners.flink.metrics.MetricsAccumulator
 
resetTo(int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
Resets the end of the stream to the specified position.
Reshuffle<K,V> - Class in org.apache.beam.sdk.transforms
Deprecated.
this transform's intended side effects are not portable; it will likely be removed
Reshuffle.ViaRandomKey<T> - Class in org.apache.beam.sdk.transforms
Deprecated.
Implementation of Reshuffle.viaRandomKey().
ReshuffleTrigger<W extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
Deprecated.
The intended side effect of Reshuffle is not portable; it will likely be removed
ReshuffleTrigger() - Constructor for class org.apache.beam.sdk.transforms.windowing.ReshuffleTrigger
Deprecated.
 
resolve(String, ResolveOptions) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
 
resolve(Path) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
resolve(String) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
resolve(String, ResolveOptions) - Method in interface org.apache.beam.sdk.io.fs.ResourceId
Returns a child ResourceId under this.
resolve(Schema) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Resolve the FieldAccessDescriptor against a schema.
resolveAlias(ResolvedColumn) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
 
resolveCalciteTable(SchemaPlus, List<String>) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.TableResolution
Resolves tablePath according to the given schemaPlus.
resolvedTables - Variable in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
 
ResolveOptions - Interface in org.apache.beam.sdk.io.fs
ResolveOptions.StandardResolveOptions - Enum in org.apache.beam.sdk.io.fs
Defines the standard resolve options.
resolveSibling(Path) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
resolveSibling(String) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
resolveType(Type) - Method in class org.apache.beam.sdk.values.TypeDescriptor
Returns a TypeDescriptor representing the given type, with type variables resolved according to the specialization in this type.
resourceId() - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata
 
ResourceId - Interface in org.apache.beam.sdk.io.fs
An identifier which represents a file-like resource.
ResourceIdCoder - Class in org.apache.beam.sdk.io.fs
ResourceIdCoder() - Constructor for class org.apache.beam.sdk.io.fs.ResourceIdCoder
 
ResourceIdTester - Class in org.apache.beam.sdk.io.fs
A utility to test ResourceId implementations.
responseMetadata() - Static method in class org.apache.beam.sdk.io.aws.coders.AwsCoders
Returns a new coder for ResponseMetadata.
RESTRICTION_CODER - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
Deprecated.
Uses the incorrect terminology. PropertyNames.RESTRICTION_ENCODING. Should be removed once non FnAPI SplittableDoFn expansion for Dataflow is removed.
RESTRICTION_ENCODING - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
RestrictionTracker<RestrictionT,PositionT> - Class in org.apache.beam.sdk.transforms.splittabledofn
Manages access to the restriction and keeps track of its claimed part for a splittable DoFn.
RestrictionTracker() - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker
 
RestrictionTrackers - Class in org.apache.beam.sdk.fn.splittabledofn
Support utilities for interacting with RestrictionTrackers.
RestrictionTrackers() - Constructor for class org.apache.beam.sdk.fn.splittabledofn.RestrictionTrackers
 
RestrictionTrackers.ClaimObserver<PositionT> - Interface in org.apache.beam.sdk.fn.splittabledofn
Interface allowing a runner to observe the calls to RestrictionTracker.tryClaim(PositionT).
Result() - Constructor for class org.apache.beam.sdk.schemas.transforms.CoGroup.Result
 
Result() - Constructor for class org.apache.beam.sdk.transforms.WithFailures.Result
 
resume(TimerInternals.TimerData) - Method in class org.apache.beam.runners.fnexecution.splittabledofn.SDFFeederViaStateAndTimers
Resumes from a timer and returns the current element/restriction pair (with an up-to-date value of the restriction).
resume() - Static method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
Indicates that there is more work to be done for the current element.
resumeDelay() - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
A minimum duration that should elapse between the end of this DoFn.ProcessElement call and the DoFn.ProcessElement call continuing processing of the same element.
retrievalToken() - Method in class org.apache.beam.runners.fnexecution.provisioning.JobInfo
 
retrieveFieldNames(List<ResolvedColumn>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
 
retrieveRexNode(ResolvedNodes.ResolvedProjectScan, List<RelDataTypeField>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
Extract expressions from a project scan node.
retrieveRexNodeFromOrderByScan(RelOptCluster, ResolvedNodes.ResolvedOrderByScan, List<RelDataTypeField>) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
Extract expressions from order by scan node.
RetryConfiguration() - Constructor for class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.RetryConfiguration
 
RetryConfiguration() - Constructor for class org.apache.beam.sdk.io.aws.sns.SnsIO.RetryConfiguration
 
RetryConfiguration() - Constructor for class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.RetryConfiguration
 
RetryConfiguration() - Constructor for class org.apache.beam.sdk.io.aws2.sns.SnsIO.RetryConfiguration
 
RetryConfiguration() - Constructor for class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.RetryConfiguration
 
RetryConfiguration() - Constructor for class org.apache.beam.sdk.io.solr.SolrIO.RetryConfiguration
 
RetryHttpRequestInitializer - Class in org.apache.beam.sdk.extensions.gcp.util
Implements a request initializer that adds retry handlers to all HttpRequests.
RetryHttpRequestInitializer() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer
 
RetryHttpRequestInitializer(Collection<Integer>) - Constructor for class org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer
 
RetryHttpRequestInitializer(Collection<Integer>, HttpResponseInterceptor) - Constructor for class org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer
 
retryTransientErrors() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy
Retry all failures except for known persistent errors.
REVERSE - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
reverse(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
REVERSE_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamBuiltinMethods
 
reverseBytes(byte[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
Reversed() - Constructor for class org.apache.beam.sdk.transforms.Top.Reversed
 
reverseString(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
RHS_TAG - Static variable in class org.apache.beam.sdk.schemas.transforms.Join
 
RIGHT - Static variable in class org.apache.beam.sdk.extensions.sql.impl.rule.BeamJoinPushThroughJoinRule
Instance of the rule that works on logical joins only, and pushes to the right.
right(String...) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual.Impl
 
right(Integer...) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual.Impl
 
right(FieldAccessDescriptor) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual.Impl
 
right(String...) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual
 
right(Integer...) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual
 
right(FieldAccessDescriptor) - Method in class org.apache.beam.sdk.schemas.transforms.Join.FieldsEqual
 
rightOuterJoin(String, PCollection<KV<K, V1>>, PCollection<KV<K, V2>>, V1) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
Right Outer Join of two collections of KV elements.
rightOuterJoin(PCollection<KV<K, V1>>, PCollection<KV<K, V2>>, V1) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join
Right Outer Join of two collections of KV elements.
rightOuterJoin(PCollection<RhsT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Join
Perform a right outer join.
rollback() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
rollback(Savepoint) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
root() - Static method in class org.apache.beam.sdk.transforms.display.DisplayData.Path
Path for display data registered by a top-level component.
roundRobinSubList(List<T>, int, int) - Static method in class org.apache.beam.runners.jet.Utils
Assigns the list to count sublists in a round-robin fashion.
row(Schema) - Static method in class org.apache.beam.sdk.schemas.Schema.FieldType
Create a map type for the given key and value types.
Row - Class in org.apache.beam.sdk.values
Row is an immutable tuple-like schema to represent one element in a PCollection.
Row.Builder - Class in org.apache.beam.sdk.values
Builder for Row.
Row.Equals - Class in org.apache.beam.sdk.values
 
Row.FieldValueBuilder - Class in org.apache.beam.sdk.values
Builder for Row that bases a row on another row.
RowCoder - Class in org.apache.beam.sdk.coders
A sub-class of SchemaCoder that can only encode Row instances.
RowCoderGenerator - Class in org.apache.beam.sdk.coders
A utility for automatically generating a Coder for Row objects corresponding to a specific schema.
RowCoderGenerator() - Constructor for class org.apache.beam.sdk.coders.RowCoderGenerator
 
rowFromProto(SchemaApi.Row, Schema.FieldType) - Static method in class org.apache.beam.sdk.schemas.SchemaTranslation
 
RowHelpers - Class in org.apache.beam.runners.spark.structuredstreaming.translation.helpers
Helper functions for working with Row.
RowHelpers() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.RowHelpers
 
rowReceiver(DoFn<?, ?>.WindowedContext, TupleTag<T>, SchemaCoder<T>) - Static method in class org.apache.beam.sdk.transforms.DoFnOutputReceivers
Returns a DoFn.OutputReceiver that automatically converts a Row to the user's output type and delegates to WindowedContextOutputReceiver.
rows() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
RowSelector - Interface in org.apache.beam.sdk.schemas.utils
A selector interface for extracting fields from a row.
RowSelectorContainer(Schema, FieldAccessDescriptor, boolean) - Constructor for class org.apache.beam.sdk.schemas.utils.SelectHelpers.RowSelectorContainer
 
rowToProto(Row) - Static method in class org.apache.beam.sdk.schemas.SchemaTranslation
 
RowToPubsubMessage - Class in org.apache.beam.sdk.extensions.sql.meta.provider.pubsub
A PTransform to convert Row to PubsubMessage with JSON payload.
RowWithGetters - Class in org.apache.beam.sdk.values
A Concrete subclass of Row that delegates to a set of provided FieldValueGetters.
RowWithStorage - Class in org.apache.beam.sdk.values
Concrete subclass of Row that explicitly stores all fields of the row.
rpad(String, Long) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
rpad(String, Long, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
rpad(byte[], Long) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
rpad(byte[], Long, byte[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
RTRIM - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
rtrim(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
rtrim(String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
RTRIM_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamBuiltinMethods
 
RULE_arrayQualifier - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
RULE_dotExpression - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
RULE_dotExpressionComponent - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
RULE_fieldSpecifier - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
RULE_mapQualifier - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
RULE_qualifiedComponent - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
RULE_qualifierList - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
ruleNames - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
ruleNames - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
run(Pipeline) - Method in class org.apache.beam.runners.apex.ApexRunner
 
run() - Method in class org.apache.beam.runners.apex.ApexYarnLauncher.ProcessWatcher
 
run(Pipeline) - Method in class org.apache.beam.runners.apex.TestApexRunner
 
run(Pipeline) - Method in class org.apache.beam.runners.dataflow.DataflowRunner
 
run(Pipeline) - Method in class org.apache.beam.runners.dataflow.TestDataflowRunner
 
run(Pipeline) - Method in class org.apache.beam.runners.direct.DirectRunner
 
run(RunnerApi.Pipeline, JobInfo) - Method in class org.apache.beam.runners.flink.FlinkPipelineRunner
 
run(Pipeline) - Method in class org.apache.beam.runners.flink.FlinkRunner
 
run(Pipeline) - Method in class org.apache.beam.runners.flink.TestFlinkRunner
 
run(JobApi.RunJobRequest, StreamObserver<JobApi.RunJobResponse>) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.InMemoryJobService
 
run() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver
 
run(RunnerApi.Pipeline, JobInfo) - Method in class org.apache.beam.runners.fnexecution.jobsubmission.PortablePipelineJarCreator
Does not actually run the pipeline. Instead bundles the input pipeline along with all dependencies, artifacts, etc.
run(RunnerApi.Pipeline, JobInfo) - Method in interface org.apache.beam.runners.fnexecution.jobsubmission.PortablePipelineRunner
 
run(Pipeline) - Method in class org.apache.beam.runners.gearpump.GearpumpRunner
 
run(Pipeline) - Method in class org.apache.beam.runners.gearpump.TestGearpumpRunner
 
run(Pipeline) - Method in class org.apache.beam.runners.jet.JetRunner
 
run(Pipeline) - Method in class org.apache.beam.runners.portability.PortableRunner
 
run(JobApi.RunJobRequest, StreamObserver<JobApi.RunJobResponse>) - Method in class org.apache.beam.runners.portability.testing.TestJobService
 
run(Pipeline) - Method in class org.apache.beam.runners.portability.testing.TestPortableRunner
 
run(RunnerApi.Pipeline, JobInfo) - Method in class org.apache.beam.runners.spark.SparkPipelineRunner
 
run(Pipeline) - Method in class org.apache.beam.runners.spark.SparkRunner
 
run(Pipeline) - Method in class org.apache.beam.runners.spark.SparkRunnerDebugger
 
run(Pipeline) - Method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingRunner
 
run(Pipeline) - Method in class org.apache.beam.runners.spark.TestSparkRunner
 
run() - Method in interface org.apache.beam.sdk.function.ThrowingRunnable
 
run() - Method in class org.apache.beam.sdk.Pipeline
Runs this Pipeline according to the PipelineOptions used to create the Pipeline via Pipeline.create(PipelineOptions).
run(PipelineOptions) - Method in class org.apache.beam.sdk.Pipeline
Runs this Pipeline using the given PipelineOptions, using the runner specified by the options.
run(Pipeline) - Method in class org.apache.beam.sdk.PipelineRunner
Processes the given Pipeline, potentially asynchronously, returning a runner-specific type of result.
run(PTransform<PBegin, ?>, PipelineOptions) - Method in class org.apache.beam.sdk.PipelineRunner
Creates a Pipeline out of a single PTransform step, and executes it.
run(PTransform<PBegin, ?>) - Method in class org.apache.beam.sdk.PipelineRunner
Overloaded PTransform runner that runs with the default app PipelineOptions.
run(Pipeline) - Method in class org.apache.beam.sdk.testing.CrashingRunner
 
run() - Method in class org.apache.beam.sdk.testing.TestPipeline
Runs this TestPipeline, unwrapping any AssertionError that is raised during testing.
run(PipelineOptions) - Method in class org.apache.beam.sdk.testing.TestPipeline
Like TestPipeline.run() but with the given potentially modified options.
runBeforeProcessing(PipelineOptions) - Static method in class org.apache.beam.sdk.fn.JvmInitializers
Finds all registered implementations of JvmInitializer and executes their beforeProcessing methods.
Runner() - Constructor for class org.apache.beam.runners.apex.ApexRunnerRegistrar.Runner
 
Runner() - Constructor for class org.apache.beam.runners.dataflow.DataflowPipelineRegistrar.Runner
 
Runner() - Constructor for class org.apache.beam.runners.direct.DirectRegistrar.Runner
 
Runner() - Constructor for class org.apache.beam.runners.flink.FlinkRunnerRegistrar.Runner
 
Runner() - Constructor for class org.apache.beam.runners.gearpump.GearpumpRunnerRegistrar.Runner
 
Runner() - Constructor for class org.apache.beam.runners.jet.JetRunnerRegistrar.Runner
 
Runner() - Constructor for class org.apache.beam.runners.spark.SparkRunnerRegistrar.Runner
 
runOnStartup() - Static method in class org.apache.beam.sdk.fn.JvmInitializers
Finds all registered implementations of JvmInitializer and executes their onStartup methods.
runResourceIdBattery(ResourceId) - Static method in class org.apache.beam.sdk.io.fs.ResourceIdTester
Enforces that the ResourceId implementation of baseDirectory meets the ResourceId spec.
runWindowFn(WindowFn<T, W>, List<Long>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
Runs the WindowFn over the provided input, returning a map of windows to the timestamps in those windows.
runWindowFnWithValue(WindowFn<T, W>, List<TimestampedValue<T>>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
Runs the WindowFn over the provided input, returning a map of windows to the timestamps in those windows.

S

S3ClientBuilderFactory - Interface in org.apache.beam.sdk.io.aws.options
Construct AmazonS3ClientBuilder from S3 pipeline options.
S3FileSystemRegistrar - Class in org.apache.beam.sdk.io.aws.s3
AutoService registrar for the S3FileSystem.
S3FileSystemRegistrar() - Constructor for class org.apache.beam.sdk.io.aws.s3.S3FileSystemRegistrar
 
S3Options - Interface in org.apache.beam.sdk.io.aws.options
Options used to configure Amazon Web Services S3.
S3Options.S3UploadBufferSizeBytesFactory - Class in org.apache.beam.sdk.io.aws.options
Provide the default s3 upload buffer size in bytes: 64MB if more than 512MB in RAM are available and 5MB otherwise.
S3UploadBufferSizeBytesFactory() - Constructor for class org.apache.beam.sdk.io.aws.options.S3Options.S3UploadBufferSizeBytesFactory
 
safeMicrosToMillis(long) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
 
Sample - Class in org.apache.beam.sdk.transforms
PTransforms for taking samples of the elements in a PCollection, or samples of the values associated with each key in a PCollection of KVs.
Sample() - Constructor for class org.apache.beam.sdk.transforms.Sample
 
Sample.FixedSizedSampleFn<T> - Class in org.apache.beam.sdk.transforms
CombineFn that computes a fixed-size sample of a collection of values.
SampleAllFiles() - Constructor for class org.apache.beam.sdk.io.TextRowCountEstimator.SampleAllFiles
 
satisfies(RelTrait) - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
 
satisfies(SerializableFunction<Iterable<T>, Void>) - Method in interface org.apache.beam.sdk.testing.PAssert.IterableAssert
Applies the provided checking function (presumably containing assertions) to the iterable in question.
satisfies(SerializableFunction<Iterable<T>, Void>) - Method in class org.apache.beam.sdk.testing.PAssert.PCollectionContentsAssert
 
satisfies(SerializableFunction<T, Void>) - Method in interface org.apache.beam.sdk.testing.PAssert.SingletonAssert
Applies the provided checking function (presumably containing assertions) to the value in question.
saveAsync(T) - Method in interface org.apache.beam.sdk.io.cassandra.Mapper
This method is called for each save event.
SCALAR_FIELD_NAME - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
ScalarFunctionImpl - Class in org.apache.beam.sdk.extensions.sql.impl
Beam-customized version from ScalarFunctionImpl , to address BEAM-5921.
schema() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
 
schema - Variable in class org.apache.beam.sdk.extensions.sql.meta.SchemaBaseBeamTable
 
schema(Schema) - Method in class org.apache.beam.sdk.extensions.sql.meta.Table.Builder
 
Schema - Class in org.apache.beam.sdk.schemas
Schema describes the fields in Row.
Schema(List<Schema.Field>) - Constructor for class org.apache.beam.sdk.schemas.Schema
 
Schema(List<Schema.Field>, Schema.Options) - Constructor for class org.apache.beam.sdk.schemas.Schema
 
schema - Variable in class org.apache.beam.sdk.schemas.SchemaCoder
 
Schema.Builder - Class in org.apache.beam.sdk.schemas
Builder class for building Schema objects.
Schema.EquivalenceNullablePolicy - Enum in org.apache.beam.sdk.schemas
Control whether nullable is included in equivalence check.
Schema.Field - Class in org.apache.beam.sdk.schemas
Field of a row.
Schema.Field.Builder - Class in org.apache.beam.sdk.schemas
Builder for Schema.Field.
Schema.FieldType - Class in org.apache.beam.sdk.schemas
A descriptor of a single field type.
Schema.LogicalType<InputT,BaseT> - Interface in org.apache.beam.sdk.schemas
A LogicalType allows users to define a custom schema type.
Schema.Options - Class in org.apache.beam.sdk.schemas
 
Schema.Options.Builder - Class in org.apache.beam.sdk.schemas
 
Schema.TypeName - Enum in org.apache.beam.sdk.schemas
An enumerated list of type constructors.
SchemaAndRecord - Class in org.apache.beam.sdk.io.gcp.bigquery
A wrapper for a GenericRecord and the TableSchema representing the schema of the table (or query) it was generated from.
SchemaAndRecord(GenericRecord, TableSchema) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.SchemaAndRecord
 
SchemaBaseBeamTable - Class in org.apache.beam.sdk.extensions.sql.meta
Each IO in Beam has one table schema, by extending SchemaBaseBeamTable.
SchemaBaseBeamTable(Schema) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.SchemaBaseBeamTable
 
SchemaCoder<T> - Class in org.apache.beam.sdk.schemas
SchemaCoder is used as the coder for types that have schemas registered.
SchemaCoder(Schema, TypeDescriptor<T>, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Constructor for class org.apache.beam.sdk.schemas.SchemaCoder
 
schemaCoder(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Returns an SchemaCoder instance for the provided element type.
schemaCoder(Class<T>) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Returns an SchemaCoder instance for the provided element class.
schemaCoder(Schema) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Returns an SchemaCoder instance for the Avro schema.
schemaCoder(Class<T>, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Returns an SchemaCoder instance for the provided element type using the provided Avro schema.
schemaCoder(AvroCoder<T>) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Returns an SchemaCoder instance based on the provided AvroCoder for the element type.
SchemaCoderCloudObjectTranslator - Class in org.apache.beam.runners.dataflow.util
Translator for Schema coders.
SchemaCoderCloudObjectTranslator() - Constructor for class org.apache.beam.runners.dataflow.util.SchemaCoderCloudObjectTranslator
 
SchemaCreate - Annotation Type in org.apache.beam.sdk.schemas.annotations
Can be put on a constructor or a static method, in which case that constructor or method will be used to created instance of the class by Beam's schema code.
SchemaFieldName - Annotation Type in org.apache.beam.sdk.schemas.annotations
When used on a POJO field or a JavaBean getter, the specified name is used for the generated schema field.
schemaFor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoMessageSchema
 
schemaFor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.annotations.DefaultSchema.DefaultSchemaProvider
 
schemaFor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.AutoValueSchema
 
schemaFor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.AvroRecordSchema
 
schemaFor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema
 
schemaFor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.JavaFieldSchema
 
schemaFor(TypeDescriptor<T>) - Method in interface org.apache.beam.sdk.schemas.SchemaProvider
Lookup a schema for the given type.
schemaFromClass(Class<?>, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.StaticSchemaInference
Infer a schema from a Java class.
schemaFromJavaBeanClass(Class<?>, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
Create a Schema for a Java Bean class.
schemaFromPojoClass(Class<?>, FieldValueTypeSupplier) - Static method in class org.apache.beam.sdk.schemas.utils.POJOUtils
 
schemaFromProto(SchemaApi.Schema) - Static method in class org.apache.beam.sdk.schemas.SchemaTranslation
 
SchemaHelpers - Class in org.apache.beam.runners.spark.structuredstreaming.translation
A SchemaHelpers for the Spark Batch Runner.
SchemaHelpers() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.SchemaHelpers
 
SchemaIgnore - Annotation Type in org.apache.beam.sdk.schemas.annotations
When used on a POJO field or a JavaBean getter, that field or getter is ignored from the inferred schema.
SchemaProvider - Interface in org.apache.beam.sdk.schemas
Concrete implementations of this class allow creation of schema service objects that vend a Schema for a specific type.
SchemaProviderRegistrar - Interface in org.apache.beam.sdk.schemas
SchemaProvider creators have the ability to automatically have their schemaProvider registered with this SDK by creating a ServiceLoader entry and a concrete implementation of this interface.
SchemaRegistry - Class in org.apache.beam.sdk.schemas
A SchemaRegistry allows registering Schemas for a given Java Class or a TypeDescriptor.
schemaToProto(Schema, boolean) - Static method in class org.apache.beam.sdk.schemas.SchemaTranslation
 
SchemaTranslation - Class in org.apache.beam.sdk.schemas
Utility methods for translating schemas.
SchemaTranslation() - Constructor for class org.apache.beam.sdk.schemas.SchemaTranslation
 
schemaTypeCreator(Class<?>, Schema) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoMessageSchema
 
schemaTypeCreator(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.AutoValueSchema
 
schemaTypeCreator(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.AvroRecordSchema
 
schemaTypeCreator(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
Implementing class should override to return a constructor.
schemaTypeCreator(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.JavaBeanSchema
 
schemaTypeCreator(Class<?>, Schema) - Method in class org.apache.beam.sdk.schemas.JavaFieldSchema
 
SchemaUserTypeCreator - Interface in org.apache.beam.sdk.schemas
A creator interface for user types that have schemas.
SchemaUtils - Class in org.apache.beam.sdk.schemas
A set of utility functions for schemas.
SchemaUtils() - Constructor for class org.apache.beam.sdk.schemas.SchemaUtils
 
SchemaVerification - Class in org.apache.beam.sdk.values
 
SchemaVerification() - Constructor for class org.apache.beam.sdk.values.SchemaVerification
 
SchemaZipFold<T> - Class in org.apache.beam.sdk.schemas.utils
Visitor that zips schemas, and accepts pairs of fields and their types.
SchemaZipFold() - Constructor for class org.apache.beam.sdk.schemas.utils.SchemaZipFold
 
SchemaZipFold.Context - Class in org.apache.beam.sdk.schemas.utils
Context referring to a current position in a schema.
SCHEME - Static variable in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
scopedMetricsContainer(MetricsContainer) - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
Set the MetricsContainer for the current thread.
SDFFeederViaStateAndTimers<InputT,RestrictionT> - Class in org.apache.beam.runners.fnexecution.splittabledofn
Helper class for feeding element/restricton pairs into a PTransformTranslation.SPLITTABLE_PROCESS_ELEMENTS_URN transform, implementing checkpointing only, by using state and timers for storing the last element/restriction pair, similarly to SplittableParDoViaKeyedWorkItems.ProcessFn but in a portable fashion.
SDFFeederViaStateAndTimers(StateInternals, TimerInternals, Coder<InputT>, Coder<RestrictionT>, Coder<BoundedWindow>) - Constructor for class org.apache.beam.runners.fnexecution.splittabledofn.SDFFeederViaStateAndTimers
Initializes the feeder.
SdkHarnessClient - Class in org.apache.beam.runners.fnexecution.control
A high-level client for an SDK harness.
SdkHarnessClient.BundleProcessor - Class in org.apache.beam.runners.fnexecution.control
A processor capable of creating bundles for some registered BeamFnApi.ProcessBundleDescriptor.
SdkHarnessClient.BundleProcessor.ActiveBundle - Class in org.apache.beam.runners.fnexecution.control
An active bundle for a particular BeamFnApi.ProcessBundleDescriptor.
SdkHarnessLogLevelOverrides() - Constructor for class org.apache.beam.sdk.options.SdkHarnessOptions.SdkHarnessLogLevelOverrides
 
SdkHarnessOptions - Interface in org.apache.beam.sdk.options
Options that are used to control configuration of the SDK harness.
SdkHarnessOptions.LogLevel - Enum in org.apache.beam.sdk.options
The set of log levels that can be used in the SDK harness.
SdkHarnessOptions.SdkHarnessLogLevelOverrides - Class in org.apache.beam.sdk.options
Defines a log level override for a specific class, package, or name.
sdkHttpMetadata() - Static method in class org.apache.beam.sdk.io.aws.coders.AwsCoders
Returns a new coder for SdkHttpMetadata.
sdkHttpMetadataWithoutHeaders() - Static method in class org.apache.beam.sdk.io.aws.coders.AwsCoders
Returns a new coder for SdkHttpMetadata that does not serialize the response headers.
seed(WindowedValue<KV<InputT, RestrictionT>>) - Method in class org.apache.beam.runners.fnexecution.splittabledofn.SDFFeederViaStateAndTimers
Passes the initial element/restriction pair.
seekable(BeamRelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
check if BeamRelNode implements BeamSeekableTable.
seekableInputIndex() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamJoinRel
 
seekRow(Row) - Method in interface org.apache.beam.sdk.extensions.sql.BeamSqlSeekableTable
return a list of Row with given key set.
Select - Class in org.apache.beam.sdk.schemas.transforms
A PTransform for selecting a subset of fields from a schema type.
Select() - Constructor for class org.apache.beam.sdk.schemas.transforms.Select
 
select(Row) - Method in interface org.apache.beam.sdk.schemas.utils.RowSelector
 
select(Row) - Method in class org.apache.beam.sdk.schemas.utils.SelectHelpers.RowSelectorContainer
 
Select.Fields<T> - Class in org.apache.beam.sdk.schemas.transforms
 
Select.Flattened<T> - Class in org.apache.beam.sdk.schemas.transforms
A PTransform representing a flattened schema.
SelectHelpers - Class in org.apache.beam.sdk.schemas.utils
Helper methods to select subrows out of rows.
SelectHelpers() - Constructor for class org.apache.beam.sdk.schemas.utils.SelectHelpers
 
SelectHelpers.RowSelectorContainer - Class in org.apache.beam.sdk.schemas.utils
 
send(LogicalEndpoint, Coder<T>) - Method in interface org.apache.beam.runners.fnexecution.data.FnDataService
Creates a receiver to which you can write data values and have them sent over this data plane service.
send(LogicalEndpoint, Coder<T>) - Method in class org.apache.beam.runners.fnexecution.data.GrpcDataService
 
SendMessageRequestCoder - Class in org.apache.beam.sdk.io.aws2.sqs
Custom Coder for handling SendMessageRequest for using in Write.
SendMessageRequestCoderRegistrar - Class in org.apache.beam.sdk.io.aws2.sqs
A CoderProviderRegistrar for standard types used with SqsIO.
SendMessageRequestCoderRegistrar() - Constructor for class org.apache.beam.sdk.io.aws2.sqs.SendMessageRequestCoderRegistrar
 
serde(T) - Static method in class org.apache.beam.runners.jet.Utils
Returns a deep clone of an object by serializing and deserializing it (ser-de).
SerializableBiFunction<FirstInputT,SecondInputT,OutputT> - Interface in org.apache.beam.sdk.transforms
A union of the BiFunction and Serializable interfaces.
SerializableCoder<T extends java.io.Serializable> - Class in org.apache.beam.sdk.coders
A Coder for Java classes that implement Serializable.
SerializableCoder(Class<T>, TypeDescriptor<T>) - Constructor for class org.apache.beam.sdk.coders.SerializableCoder
 
SerializableCoder.SerializableCoderProviderRegistrar - Class in org.apache.beam.sdk.coders
A CoderProviderRegistrar which registers a CoderProvider which can handle serializable types.
SerializableCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.coders.SerializableCoder.SerializableCoderProviderRegistrar
 
SerializableComparator<T> - Interface in org.apache.beam.sdk.transforms
A Comparator that is also Serializable.
SerializableConfiguration - Class in org.apache.beam.sdk.io.hadoop
A wrapper to allow Hadoop Configurations to be serialized using Java's standard serialization mechanisms.
SerializableConfiguration() - Constructor for class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
 
SerializableConfiguration(Configuration) - Constructor for class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
 
SerializableFunction<InputT,OutputT> - Interface in org.apache.beam.sdk.transforms
A function that computes an output value of type OutputT from an input value of type InputT, is Serializable, and does not allow checked exceptions to be declared.
SerializableFunctions - Class in org.apache.beam.sdk.transforms
Useful SerializableFunction overrides.
SerializableFunctions() - Constructor for class org.apache.beam.sdk.transforms.SerializableFunctions
 
SerializableMatcher<T> - Interface in org.apache.beam.sdk.testing
A Matcher that is also Serializable.
SerializableMatchers - Class in org.apache.beam.sdk.testing
Static class for building and using SerializableMatcher instances.
SerializableMatchers.SerializableSupplier<T> - Interface in org.apache.beam.sdk.testing
Supplies values of type T, and is serializable.
SerializableRexFieldAccess - Class in org.apache.beam.sdk.extensions.sql.impl.utils
SerializableRexFieldAccess.
SerializableRexFieldAccess(RexFieldAccess) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.SerializableRexFieldAccess
 
SerializableRexInputRef - Class in org.apache.beam.sdk.extensions.sql.impl.utils
SerializableRexInputRef.
SerializableRexInputRef(RexInputRef) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.SerializableRexInputRef
 
SerializableRexNode - Class in org.apache.beam.sdk.extensions.sql.impl.utils
SerializableRexNode.
SerializableRexNode() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.SerializableRexNode
 
SerializableRexNode.Builder - Class in org.apache.beam.sdk.extensions.sql.impl.utils
SerializableRexNode.Builder.
SerializableSplit() - Constructor for class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.SerializableSplit
 
SerializableSplit(InputSplit) - Constructor for class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.SerializableSplit
 
serialize(String, Instant) - Method in class org.apache.beam.sdk.io.kafka.serialization.InstantSerializer
 
serialize(ValueProvider<?>, JsonGenerator, SerializerProvider) - Method in class org.apache.beam.sdk.options.ValueProvider.Serializer
 
SERIALIZED_FN - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
SERIALIZED_TEST_STREAM - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
Serializer() - Constructor for class org.apache.beam.sdk.options.ValueProvider.Serializer
 
serializeTimers(Collection<TimerInternals.TimerData>, TimerInternals.TimerDataCoderV2) - Static method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
 
serialVersionUID - Static variable in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
 
serialVersionUID - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
 
serialVersionUID - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoDomain
 
serialVersionUID - Static variable in class org.apache.beam.sdk.extensions.protobuf.ProtoDynamicMessageSchema
 
ServerConfiguration() - Constructor for class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver.ServerConfiguration
 
serverDirect() - Static method in class org.apache.beam.sdk.fn.stream.OutboundObserverFactory
Like OutboundObserverFactory.clientDirect() but for server-side RPCs.
ServerFactory - Class in org.apache.beam.runners.fnexecution
A gRPC server factory.
ServerFactory() - Constructor for class org.apache.beam.runners.fnexecution.ServerFactory
 
ServerFactory.InetSocketAddressServerFactory - Class in org.apache.beam.runners.fnexecution
Creates a gRPC Server using the default server factory.
ServerFactory.UrlFactory - Interface in org.apache.beam.runners.fnexecution
Factory that constructs client-accessible URLs from a local server address and port.
ServerInfo() - Constructor for class org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory.ServerInfo
 
Sessions - Class in org.apache.beam.sdk.transforms.windowing
A WindowFn that windows values into sessions separated by periods with no input for at least the duration specified by Sessions.getGapDuration().
set(long) - Method in class org.apache.beam.runners.jet.metrics.GaugeImpl
 
set(long) - Method in interface org.apache.beam.sdk.metrics.Gauge
Set current value for this gauge.
set(ObjectT, ValueT) - Method in interface org.apache.beam.sdk.schemas.FieldValueSetter
Sets the specified field on object to value.
set() - Static method in class org.apache.beam.sdk.state.StateSpecs
Create a StateSpec for a SetState, optimized for checking membership.
set(Coder<T>) - Static method in class org.apache.beam.sdk.state.StateSpecs
Identical to StateSpecs.set(), but with an element coder explicitly supplied.
set(Instant) - Method in interface org.apache.beam.sdk.state.Timer
Sets or resets the time in the timer's TimeDomain at which it should fire.
set(String, Instant) - Method in interface org.apache.beam.sdk.state.TimerMap
 
set(long...) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
setAllowNonRestoredState(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setApiRootUrl(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
 
setApplicationName(String) - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
 
setApplicationName(String) - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
 
setAppName(String) - Method in interface org.apache.beam.sdk.options.ApplicationNameOptions
 
setAutoBalanceWriteFilesShardingEnabled(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setAutoCommit(boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
setAutoscalingAlgorithm(DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
 
setAutoWatermarkInterval(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setAwsCredentialsProvider(AWSCredentialsProvider) - Method in interface org.apache.beam.sdk.io.aws.options.AwsOptions
 
setAwsCredentialsProvider(AwsCredentialsProvider) - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
 
setAwsRegion(String) - Method in interface org.apache.beam.sdk.io.aws.options.AwsOptions
 
setAwsServiceEndpoint(String) - Method in interface org.apache.beam.sdk.io.aws.options.AwsOptions
 
setBatchIntervalMillis(Long) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
setBlockOnRun(boolean) - Method in interface org.apache.beam.runners.direct.DirectOptions
 
setBundleSize(Long) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
setCacheDisabled(boolean) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
setCatalog(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
setCheckpointDir(String) - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
 
setCheckpointDurationMillis(Long) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
setCheckpointingInterval(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setCheckpointingMode(String) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setCheckpointTimeoutMillis(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setClientConfiguration(ClientConfiguration) - Method in interface org.apache.beam.sdk.io.aws.options.AwsOptions
 
setClientContext(ClientContext) - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
 
setClientInfo(String, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
setClientInfo(Properties) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
setCloningBehavior(DoFnTester.CloningBehavior) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
setClusterName(String) - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
 
setCodeJarPathname(String) - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
 
SetCoder<T> - Class in org.apache.beam.sdk.coders
A SetCoder encodes any Set using the format of IterableLikeCoder.
SetCoder(Coder<T>) - Constructor for class org.apache.beam.sdk.coders.SetCoder
 
setCoder(Coder<T>) - Method in class org.apache.beam.sdk.values.PCollection
Sets the Coder used by this PCollection to encode and decode the values stored in it.
setCoderRegistry(CoderRegistry) - Method in class org.apache.beam.sdk.Pipeline
Deprecated.
this should never be used - every Pipeline has a registry throughout its lifetime.
setCompression(Compression) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
 
setConfigFile(String) - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
 
setConsumerConfig(Iterable<KV<String, String>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
 
setCountryOfResidence(String) - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
 
setCreateFromSnapshot(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
 
setCredentialFactoryClass(Class<? extends CredentialFactory>) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
 
setCurrentContainer(MetricsContainer) - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
Set the MetricsContainer for the current thread.
setCurrentSchema(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
Set the current (default) schema.
setCurrentTransform(TransformHierarchy.Node, Pipeline) - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
 
setCurrentTransform(AppliedPTransform<?, ?, ?>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
setCustomerId(int) - Method in class org.apache.beam.sdk.extensions.sql.example.model.Order
 
setCustomErrors(CustomHttpErrors) - Method in class org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer
 
setDataCatalogEndpoint(String) - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogPipelineOptions
 
setDataflowClient(Dataflow) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
 
setDataflowEndpoint(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
 
setDataflowJobFile(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
 
setDataflowKmsKey(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
 
setDataflowWorkerJar(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
 
setDebuggee(Debuggee) - Method in interface org.apache.beam.runners.dataflow.options.CloudDebuggerOptions
 
setDefaultEnvironmentConfig(String) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
setDefaultEnvironmentType(String) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
setDefaultPipelineOptions(PipelineOptions) - Static method in class org.apache.beam.sdk.io.FileSystems
Sets the default configuration in workers.
setDefaultSdkHarnessLogLevel(SdkHarnessOptions.LogLevel) - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
 
setDefaultWorkerLogLevel(DataflowWorkerLoggingOptions.Level) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
Deprecated.
 
setDelimiters(byte[]) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
 
setDescription(String) - Method in class org.apache.beam.sdk.schemas.Schema.Field.Builder
 
setDirectoryTreatment(FileIO.ReadMatches.DirectoryTreatment) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
 
setDisableMetrics(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setDiskSizeGb(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
 
setDumpHeapOnOOM(boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
 
setElementsPerPeriod(Long) - Method in class org.apache.beam.sdk.io.GenerateSequence.External.ExternalConfiguration
 
setElementType(FieldValueTypeInformation) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
 
setEmbeddedExecution(boolean) - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
 
setEmbeddedExecutionDebugMode(boolean) - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
 
setEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
 
setEnableCloudDebugger(boolean) - Method in interface org.apache.beam.runners.dataflow.options.CloudDebuggerOptions
 
setEnableSparkMetricSinks(Boolean) - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
 
setEnableStreamingEngine(boolean) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
 
setEncodingPositions(Map<String, Integer>) - Method in class org.apache.beam.sdk.schemas.Schema
Sets the encoding positions for this schema.
setEndpoint(String) - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
 
setEnforceEncodability(boolean) - Method in interface org.apache.beam.runners.direct.DirectOptions
 
setEnforceImmutability(boolean) - Method in interface org.apache.beam.runners.direct.DirectOptions
 
setEnvironmentCacheMillis(int) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
setEnvironmentExpirationMillis(int) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
setExecutionModeForBatch(String) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setExecutionRetryDelay(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setExecutorService(ExecutorService) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
 
setExpectedAssertions(Integer) - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
 
setExperiments(List<String>) - Method in interface org.apache.beam.sdk.options.ExperimentalOptions
 
setExternalizedCheckpointsEnabled(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setFailOnCheckpointingErrors(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setField(Field) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
 
setFieldId(Integer) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Builder
 
setFieldName(String) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Builder
 
setFieldRename(String) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Builder
 
setFilePattern(String) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
 
setFilesToStage(List<String>) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
 
setFilesToStage(List<String>) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setFilesToStage(List<String>) - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
 
setFilesToStage(List<String>) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
setFileSystem(FileSystem) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
setFlexRSGoal(DataflowPipelineOptions.FlexResourceSchedulingGoal) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
 
setFlinkMaster(String) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setForceStreaming(boolean) - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
 
setGcpCredential(Credentials) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
 
setGcpTempLocation(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
 
setGcsEndpoint(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
 
setGcsPerformanceMetrics(Boolean) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
 
setGcsUploadBufferSizeBytes(Integer) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
 
setGcsUploadBufferSizeBytes(Integer) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsCreateOptions.Builder
 
setGcsUtil(GcsUtil) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
 
setGoogleApiTrace(GoogleApiDebugOptions.GoogleApiTracer) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GoogleApiDebugOptions
 
setHdfsConfiguration(List<Configuration>) - Method in interface org.apache.beam.sdk.io.hdfs.HadoopFileSystemOptions
 
setHoldability(int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
setHooks(DataflowRunnerHooks) - Method in class org.apache.beam.runners.dataflow.DataflowRunner
Sets callbacks to invoke during execution see DataflowRunnerHooks.
setHTTPWriteTimeout(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
 
setId(int) - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
 
setId(int) - Method in class org.apache.beam.sdk.extensions.sql.example.model.Order
 
setIdLabel(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.Configuration
 
setIdLabel(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite.Configuration
 
setInputFile(String) - Method in interface org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.WordCountOptions
 
setInsertBundleParallelism(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
 
setIsBoundedInternal(PCollection.IsBounded) - Method in class org.apache.beam.sdk.values.PCollection
For internal use only; no backwards-compatibility guarantees.
setIsReadSeekEfficient(boolean) - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
 
setJetDefaultParallelism(Integer) - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
 
setJetLocalMode(Integer) - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
 
setJetProcessorsCooperative(Boolean) - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
 
setJetServers(String) - Method in interface org.apache.beam.runners.jet.JetPipelineOptions
 
setJobEndpoint(String) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
setJobId(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
 
setJobName(String) - Method in interface org.apache.beam.sdk.options.PipelineOptions
 
setJobServerConfig(String...) - Method in interface org.apache.beam.runners.portability.testing.TestPortablePipelineOptions
 
setJobServerDriver(Class<JobServerDriver>) - Method in interface org.apache.beam.runners.portability.testing.TestPortablePipelineOptions
 
setKeyDeserializer(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
 
setKeySerializer(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write.External.Configuration
 
setLabels(Map<String, String>) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
 
setLastModifiedMillis(long) - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
 
setLatencyTrackingInterval(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setListeners(List<JavaStreamingListener>) - Method in interface org.apache.beam.runners.spark.SparkContextOptions
 
setLoadBalanceBundles(boolean) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
setMapKeyType(FieldValueTypeInformation) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
 
setMapValueType(FieldValueTypeInformation) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
 
setMaxBundleSize(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setMaxBundleTimeMills(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setMaxConditionCost(double) - Method in interface org.apache.beam.runners.dataflow.options.CloudDebuggerOptions
 
setMaxNumWorkers(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
 
setMaxParallelism(Integer) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setMaxReadTime(Long) - Method in class org.apache.beam.sdk.io.GenerateSequence.External.ExternalConfiguration
 
setMaxRecordsPerBatch(Long) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
setMaxStreamingBatchSize(Long) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
 
setMaxStreamingRowsToBatch(Long) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
 
setMemoryMB(int) - Method in class org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options
Sets the size of the memory buffer in megabytes.
setMethod(Method) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
 
setMetricsGraphiteHost(String) - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
 
setMetricsGraphitePort(Integer) - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
 
setMetricsHttpSinkUrl(String) - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
 
setMetricsPushPeriod(Long) - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
 
setMetricsSink(Class<? extends MetricsSink>) - Method in interface org.apache.beam.sdk.metrics.MetricsOptions
 
setMetricsSupported(boolean) - Static method in class org.apache.beam.sdk.metrics.MetricsEnvironment
Called by the run to indicate whether metrics reporting is supported.
setMimeType(String) - Method in class org.apache.beam.sdk.io.fs.CreateOptions.Builder
 
setMinCpuPlatform(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
 
setMinPauseBetweenCheckpoints(Long) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setMinReadTimeMillis(Long) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
setName(String) - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
 
setName(String) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
 
setName(String) - Method in class org.apache.beam.sdk.schemas.Schema.Field.Builder
 
setName(String) - Method in class org.apache.beam.sdk.values.PCollection
Sets the name of this PCollection.
setName(String) - Method in class org.apache.beam.sdk.values.PValueBase
Sets the name of this PValueBase.
setNetwork(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
 
setNetworkTimeout(Executor, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
setNullable(boolean) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
 
setNumberOfExecutionRetries(Integer) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setNumberOfWorkerHarnessThreads(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
 
setNumConcurrentCheckpoints(int) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setNumFailuresExpected(int) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
 
setNumSampledBytesPerFile(long) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
 
setNumStreamingKeys(Integer) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
 
setNumWorkers(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
 
setObjectReuse(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setOnCreateMatcher(SerializableMatcher<PipelineResult>) - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
 
setOneOfTypes(Map<String, FieldValueTypeInformation>) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
 
setOnSuccessMatcher(SerializableMatcher<PipelineResult>) - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
 
SetOperatorFilteringDoFn(String, String, BeamSetOperatorRelBase.OpType, boolean) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.transform.BeamSetOperatorsTransforms.SetOperatorFilteringDoFn
 
setOption(String, Row) - Method in class org.apache.beam.sdk.schemas.Schema.Options.Builder
 
setOption(String, Schema.FieldType, Object) - Method in class org.apache.beam.sdk.schemas.Schema.Options.Builder
 
setOption(String, Schema.FieldType, Object) - Static method in class org.apache.beam.sdk.schemas.Schema.Options
 
setOption(String, Row) - Static method in class org.apache.beam.sdk.schemas.Schema.Options
 
setOptions(Schema.Options) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
Returns a copy of the Field with isNullable set.
setOptions(Schema.Options.Builder) - Method in class org.apache.beam.sdk.schemas.Schema.Builder
 
setOptions(Schema.Options) - Method in class org.apache.beam.sdk.schemas.Schema.Field.Builder
 
setOptions(Schema.Options.Builder) - Method in class org.apache.beam.sdk.schemas.Schema.Field.Builder
 
setOptionsId(long) - Method in interface org.apache.beam.sdk.options.PipelineOptions
 
setOutput(String) - Method in interface org.apache.beam.runners.spark.structuredstreaming.examples.WordCount.WordCountOptions
 
setOutputExecutablePath(String) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
setOutputFilePrefix(String) - Method in interface org.apache.beam.sdk.extensions.sql.example.BeamSqlDataCatalogExample.DCExamplePipelineOptions
 
setOutputStream(PValue, JavaStream<OutputT>) - Method in class org.apache.beam.runners.gearpump.translators.TranslationContext
 
setOverrideWindmillBinary(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
 
setParallelism(Integer) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setParallelism(int) - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
 
setParameters(T, PreparedStatement) - Method in interface org.apache.beam.sdk.io.jdbc.JdbcIO.PreparedStatementSetter
 
setParameters(PreparedStatement) - Method in interface org.apache.beam.sdk.io.jdbc.JdbcIO.StatementPreparator
 
setParDoFusionEnabled(boolean) - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
 
setPathValidator(PathValidator) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
 
setPathValidatorClass(Class<? extends PathValidator>) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcsOptions
 
setPeriod(Long) - Method in class org.apache.beam.sdk.io.GenerateSequence.External.ExternalConfiguration
 
setPipelineOption(String, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
setPipelineOptions(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
 
setPipelineOptions(PipelineOptions) - Method in class org.apache.beam.sdk.extensions.sql.impl.JdbcConnection
 
setPipelineOptionsMap(Map<String, String>) - Method in class org.apache.beam.sdk.extensions.sql.impl.JdbcConnection
Only called from the BeamCalciteSchema.
setPipelineUrl(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
 
setPlannerName(String) - Method in interface org.apache.beam.sdk.extensions.sql.impl.BeamSqlPipelineOptions
 
setProducerConfig(Iterable<KV<String, String>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write.External.Configuration
 
setProfilingAgentConfiguration(DataflowProfilingOptions.DataflowProfilingAgentConfiguration) - Method in interface org.apache.beam.runners.dataflow.options.DataflowProfilingOptions
 
setProject(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
 
setProject(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
 
setProvidedSparkContext(JavaSparkContext) - Method in interface org.apache.beam.runners.spark.SparkContextOptions
 
setProviderRuntimeValues(ValueProvider<Map<String, Object>>) - Method in interface org.apache.beam.sdk.testing.TestPipeline.TestValueProviderOptions
 
setProxyConfiguration(ProxyConfiguration) - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
 
setPubsubRootUrl(String) - Method in interface org.apache.beam.sdk.io.gcp.pubsub.PubsubOptions
 
setQualifiers(List<FieldAccessDescriptor.FieldDescriptor.Qualifier>) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Builder
 
setQueryPlannerClassName(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
 
setQueryString(String) - Method in interface org.apache.beam.sdk.extensions.sql.example.BeamSqlDataCatalogExample.DCExamplePipelineOptions
 
setRawType(Class) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
 
setReadOnly(boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
setReadTimePercentage(Double) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
setRegion(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
 
setRegion(String) - Method in interface org.apache.beam.sdk.io.aws2.options.AwsOptions
 
setRelative() - Method in interface org.apache.beam.sdk.state.Timer
Sets the timer relative to the current time, according to any offset and alignment specified.
setRemote(Boolean) - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
 
setResourceId(ResourceId) - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
 
setRetainDockerContainers(boolean) - Method in interface org.apache.beam.sdk.options.ManualDockerEnvironmentOptions
 
setRetainExternalizedCheckpointsOnCancellation(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setRetrievalServiceType(PortablePipelineOptions.RetrievalServiceType) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
setRowSchema(Schema) - Method in class org.apache.beam.sdk.values.PCollection
Sets a schema on this PCollection.
setRuleSets(RuleSet[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv.BeamSqlEnvBuilder
Set the ruleSet used for query optimizer.
setRunMillis(long) - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
 
setRunner(Class<? extends PipelineRunner<?>>) - Method in interface org.apache.beam.sdk.options.PipelineOptions
 
setRunnerDeterminedSharding(boolean) - Method in interface org.apache.beam.runners.direct.DirectTestOptions
 
sets(TypeDescriptor<T>) - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for Set.
setS3ClientFactoryClass(Class<? extends S3ClientBuilderFactory>) - Method in interface org.apache.beam.sdk.io.aws.options.S3Options
 
setS3StorageClass(String) - Method in interface org.apache.beam.sdk.io.aws.options.S3Options
 
setS3ThreadPoolSize(int) - Method in interface org.apache.beam.sdk.io.aws.options.S3Options
 
setS3UploadBufferSizeBytes(Integer) - Method in interface org.apache.beam.sdk.io.aws.options.S3Options
 
setSamplingStrategy(TextRowCountEstimator.SamplingStrategy) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.Builder
 
setSaveHeapDumpsToGcsPath(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
 
setSavepoint() - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
setSavepoint(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
setSavepointPath(String) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setSaveProfilesToGcs(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowProfilingOptions
 
setSchema(String) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
setSchema(Schema, TypeDescriptor<T>, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Method in class org.apache.beam.sdk.values.PCollection
Sets a Schema on this PCollection.
setSdkHarnessLogLevelOverrides(SdkHarnessOptions.SdkHarnessLogLevelOverrides) - Method in interface org.apache.beam.sdk.options.SdkHarnessOptions
 
setSdkWorkerParallelism(int) - Method in interface org.apache.beam.sdk.options.PortablePipelineOptions
 
setSemiPersistDir(String) - Method in interface org.apache.beam.sdk.options.RemoteEnvironmentOptions
 
setSerializers(Map<String, String>) - Method in interface org.apache.beam.runners.gearpump.GearpumpPipelineOptions
 
setServiceAccount(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
 
setShutdownSourcesOnFinalWatermark(Boolean) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setSideInput(PCollectionView<T>, BoundedWindow, T) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
setSideInputDataset(PCollectionView<ViewT>, Dataset<WindowedValue<ElemT>>) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
setSideInputs(Map<PCollectionView<?>, Map<BoundedWindow, ?>>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
setSizeBytes(long) - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata.Builder
 
setSorterType(ExternalSorter.Options.SorterType) - Method in class org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options
Sets the sorter type.
setSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction.Builder
 
setSparkMaster(String) - Method in interface org.apache.beam.runners.spark.SparkCommonPipelineOptions
 
setSSEAlgorithm(String) - Method in interface org.apache.beam.sdk.io.aws.options.S3Options
 
setSSEAwsKeyManagementParams(SSEAwsKeyManagementParams) - Method in interface org.apache.beam.sdk.io.aws.options.S3Options
 
setSSECustomerKey(SSECustomerKey) - Method in interface org.apache.beam.sdk.io.aws.options.S3Options
 
setStableUniqueNames(PipelineOptions.CheckEnabled) - Method in interface org.apache.beam.sdk.options.PipelineOptions
 
setStager(Stager) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
 
setStagerClass(Class<? extends Stager>) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
 
setStagingLocation(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
 
setStart(Long) - Method in class org.apache.beam.sdk.io.GenerateSequence.External.ExternalConfiguration
 
SetState<T> - Interface in org.apache.beam.sdk.state
A ReadableState cell containing a set of elements.
setStateBackendFactory(Class<? extends FlinkStateBackendFactory>) - Method in interface org.apache.beam.runners.flink.FlinkPipelineOptions
 
setStatistics(BeamTableStatistics) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestUnboundedTable
 
setStop(Long) - Method in class org.apache.beam.sdk.io.GenerateSequence.External.ExternalConfiguration
 
setStopPipelineWatermark(Long) - Method in interface org.apache.beam.runners.spark.TestSparkPipelineOptions
 
setStorageClient(Storage) - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
 
setStorageLevel(String) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
setStreaming(boolean) - Method in interface org.apache.beam.sdk.options.StreamingOptions
 
setSubnetwork(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
 
setSubscription(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.Configuration
 
setTargetDataset(String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.TestBigQueryOptions
 
setTargetParallelism(int) - Method in interface org.apache.beam.runners.direct.DirectOptions
 
setTempDatasetId(String) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions
 
setTemplateLocation(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
 
setTempLocation(String) - Method in class org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options
Sets the path to a temporary location where the sorter writes intermediate files.
setTempLocation(String) - Method in interface org.apache.beam.sdk.options.PipelineOptions
 
setTempRoot(String) - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
 
SetterTypeSupplier() - Constructor for class org.apache.beam.sdk.schemas.JavaBeanSchema.SetterTypeSupplier
 
setTestMode(boolean) - Method in interface org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineOptions
 
setTestTimeoutSeconds(Long) - Method in interface org.apache.beam.sdk.testing.TestPipelineOptions
 
setTimer(TimerInternals.TimerData) - Method in class org.apache.beam.runners.direct.WatermarkManager.TimerUpdate.TimerUpdateBuilder
Adds the provided timer to the collection of set timers, removing it from deleted timers if it has previously been deleted.
setTimer(TimerInternals.TimerData) - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
 
setTimer(StateNamespace, String, String, Instant, Instant, TimeDomain) - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
 
setTimer(Instant, TimeDomain) - Method in interface org.apache.beam.sdk.state.Timers
Sets a timer to fire when the event time watermark, the current processing time, or the synchronized processing time watermark surpasses a given timestamp.
setTimestampAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.Configuration
 
setTimestampAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite.Configuration
 
setTimestampBound(TimestampBound) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction.Builder
 
setTopic(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.Configuration
 
setTopic(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite.Configuration
 
setTopic(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write.External.Configuration
 
setTopics(Iterable<String>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
 
setTransactionIsolation(int) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
setTransformNameMapping(Map<String, String>) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
 
setTruncateTimestamps(boolean) - Method in interface org.apache.beam.sdk.extensions.sql.meta.provider.datacatalog.DataCatalogPipelineOptions
 
setTruncateTimestamps(BigQueryUtils.ConversionOptions.TruncateTimestamps) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.ConversionOptions.Builder
 
setTupleTracingEnabled(boolean) - Method in interface org.apache.beam.runners.apex.ApexPipelineOptions
 
setType(TypeDescriptor) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation.Builder
 
setType(Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Field.Builder
 
setTypeDescriptor(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.values.PCollection
setTypeMap(Map<String, Class<?>>) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
setup() - Method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides.BatchStatefulDoFn
 
setup() - Method in class org.apache.beam.runners.gearpump.translators.functions.DoFnFunction
 
setUp() - Method in interface org.apache.beam.sdk.extensions.sql.BeamSqlSeekableTable
prepare the instance.
setUp() - Static method in class org.apache.beam.sdk.io.gcp.testing.FakeDatasetService
 
setUp() - Static method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
 
setUpdate(boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineOptions
 
setUsePublicIps(Boolean) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
 
setUserAgent(String) - Method in interface org.apache.beam.sdk.options.PipelineOptions
 
setUsesProvidedSparkContext(boolean) - Method in interface org.apache.beam.runners.spark.SparkPipelineOptions
 
setUUID(UUID) - Method in class org.apache.beam.sdk.schemas.Schema
Set this schema's UUID.
setValueDeserializer(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External.Configuration
 
setValueSerializer(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write.External.Configuration
 
setVerifyRowValues(Boolean) - Method in interface org.apache.beam.sdk.extensions.sql.impl.BeamSqlPipelineOptions
 
setWatermark(Instant) - Method in interface org.apache.beam.sdk.transforms.splittabledofn.ManualWatermarkEstimator
Sets a timestamp before or at the timestamps of all future elements produced by the associated DoFn.
setWatermark(Instant) - Method in class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.Manual
 
setWindmillServiceEndpoint(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
 
setWindmillServicePort(int) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions
 
setWindowedWrites(boolean) - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
Indicates that the operation will be performing windowed writes.
setWindowingStrategyInternal(WindowingStrategy<?, ?>) - Method in class org.apache.beam.sdk.values.PCollection
For internal use only; no backwards-compatibility guarantees.
setWithAttributes(Boolean) - Method in class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead.Configuration
 
setWorkerCacheMb(Integer) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
 
setWorkerDiskType(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
 
setWorkerHarnessContainerImage(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
 
setWorkerId(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerHarnessOptions
 
setWorkerLogLevelOverrides(DataflowWorkerLoggingOptions.WorkerLogLevelOverrides) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
Deprecated.
 
setWorkerMachineType(String) - Method in interface org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions
 
setWorkerRegion(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
 
setWorkerSystemErrMessageLevel(DataflowWorkerLoggingOptions.Level) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
Deprecated.
 
setWorkerSystemOutMessageLevel(DataflowWorkerLoggingOptions.Level) - Method in interface org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions
Deprecated.
 
setWorkerZone(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
 
setWriteTimeout(int) - Method in class org.apache.beam.sdk.extensions.gcp.util.RetryHttpRequestInitializer
 
setZone(String) - Method in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
ShardedKey<K> - Class in org.apache.beam.sdk.values
A key and a shard number.
ShardedKeyCoder<KeyT> - Class in org.apache.beam.sdk.coders
A Coder for ShardedKey, using a wrapped key Coder.
ShardedKeyCoder(Coder<KeyT>) - Constructor for class org.apache.beam.sdk.coders.ShardedKeyCoder
 
ShardingFunction<UserT,DestinationT> - Interface in org.apache.beam.sdk.io
Function for assigning ShardedKeys to input elements for sharded WriteFiles.
ShardNameTemplate - Class in org.apache.beam.sdk.io
Standard shard naming templates.
ShardNameTemplate() - Constructor for class org.apache.beam.sdk.io.ShardNameTemplate
 
shorts() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for Short.
shouldConvertRaggedUnionTypesToVarying() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamRelDataTypeSystem
 
shouldDefer(TransformHierarchy.Node) - Method in class org.apache.beam.runners.spark.SparkRunner.Evaluator
 
shouldResume() - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
If false, the DoFn promises that there is no more work remaining for the current element, so the runner should not resume the DoFn.ProcessElement call.
shouldRetry(InsertRetryPolicy.Context) - Method in class org.apache.beam.sdk.io.gcp.bigquery.InsertRetryPolicy
Return true if this failure should be retried.
sideInput(PCollectionView<SideInputT>) - Method in class org.apache.beam.sdk.io.FileBasedSink.DynamicDestinations
Returns the value of a given side input.
sideInput(PCollectionView<SideInputT>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations
Returns the value of a given side input.
sideInput(PCollectionView<T>) - Method in interface org.apache.beam.sdk.state.StateContext
Returns the value of the side input for the corresponding state window.
sideInput(PCollectionView<T>) - Method in class org.apache.beam.sdk.transforms.CombineWithContext.Context
Returns the value of the side input for the window corresponding to the main input's window in which values are being combined.
sideInput(PCollectionView<T>) - Method in class org.apache.beam.sdk.transforms.Contextful.Fn.Context
Accesses the given side input.
sideInput(PCollectionView<T>) - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContext
Returns the value of the side input.
SideInputBroadcast - Class in org.apache.beam.runners.spark.structuredstreaming.translation.helpers
Broadcast helper for side inputs.
SideInputBroadcast() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.SideInputBroadcast
 
SideInputBroadcast<T> - Class in org.apache.beam.runners.spark.util
Broadcast helper for side inputs.
sideInputId() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.SideInputSpec
 
sideInputJoin(PCollection<Row>, PCollection<Row>, FieldAccessDescriptor, FieldAccessDescriptor) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSideInputJoinRel
 
SideInputSpec() - Constructor for class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.SideInputSpec
 
signalStart() - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal
Outputs a message that the pipeline has started.
signalSuccessWhen(Coder<T>, SerializableFunction<T, String>, SerializableFunction<Set<T>, Boolean>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal
Outputs a success message when successPredicate is evaluated to true.
signalSuccessWhen(Coder<T>, SerializableFunction<Set<T>, Boolean>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal
SimpleCombineFn(SerializableFunction<Iterable<V>, V>) - Constructor for class org.apache.beam.sdk.transforms.Combine.SimpleCombineFn
Deprecated.
 
SimpleFunction<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
A SerializableFunction which is not a functional interface.
SimpleFunction() - Constructor for class org.apache.beam.sdk.transforms.SimpleFunction
 
SimpleFunction(SerializableFunction<InputT, OutputT>) - Constructor for class org.apache.beam.sdk.transforms.SimpleFunction
 
SimpleIdentifierContext(FieldSpecifierNotationParser.DotExpressionComponentContext) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.SimpleIdentifierContext
 
SimpleRemoteEnvironment() - Constructor for class org.apache.beam.runners.fnexecution.environment.RemoteEnvironment.SimpleRemoteEnvironment
 
SingleEnvironmentInstanceJobBundleFactory - Class in org.apache.beam.runners.fnexecution.control
Deprecated.
replace with a DefaultJobBundleFactory when appropriate if the EnvironmentFactory is a DockerEnvironmentFactory, or create an InProcessJobBundleFactory and inline the creation of the environment if appropriate.
singleOutputOverrideFactory(DataflowPipelineOptions) - Static method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides
Returns a PTransformOverrideFactory that replaces a single-output ParDo with a composite transform specialized for the DataflowRunner.
singletonView(PCollection<KV<Void, T>>, PCollectionViews.TypeDescriptorSupplier<T>, WindowingStrategy<?, W>, boolean, T, Coder<T>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
Returns a PCollectionView<T> capable of processing elements windowed using the provided WindowingStrategy.
sinh(Double) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinTrigonometricFunctions
SINH(X)
sink(Class<ElementT>) - Static method in class org.apache.beam.sdk.io.AvroIO
A AvroIO.Sink for use with FileIO.write() and FileIO.writeDynamic(), writing elements of the given generated class, like AvroIO.write(Class).
sink(Schema) - Static method in class org.apache.beam.sdk.io.AvroIO
A AvroIO.Sink for use with FileIO.write() and FileIO.writeDynamic(), writing elements with a given (common) schema, like AvroIO.writeGenericRecords(Schema).
sink(String) - Static method in class org.apache.beam.sdk.io.AvroIO
A AvroIO.Sink for use with FileIO.write() and FileIO.writeDynamic(), writing elements with a given (common) schema, like AvroIO.writeGenericRecords(String).
Sink() - Constructor for class org.apache.beam.sdk.io.AvroIO.Sink
 
sink - Variable in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
The Sink that this WriteOperation will write to.
sink(Schema) - Static method in class org.apache.beam.sdk.io.parquet.ParquetIO
Creates a ParquetIO.Sink that, for use with FileIO.write().
Sink() - Constructor for class org.apache.beam.sdk.io.parquet.ParquetIO.Sink
 
sink() - Static method in class org.apache.beam.sdk.io.TextIO
Creates a TextIO.Sink that writes newline-delimited strings in UTF-8, for use with FileIO.write().
Sink() - Constructor for class org.apache.beam.sdk.io.TextIO.Sink
 
sink() - Static method in class org.apache.beam.sdk.io.TFRecordIO
Sink() - Constructor for class org.apache.beam.sdk.io.TFRecordIO.Sink
 
sink(TProtocolFactory) - Static method in class org.apache.beam.sdk.io.thrift.ThriftIO
Sink() - Constructor for class org.apache.beam.sdk.io.thrift.ThriftIO.Sink
 
sink(Class<T>) - Static method in class org.apache.beam.sdk.io.xml.XmlIO
Outputs records as XML-formatted elements using JAXB.
Sink() - Constructor for class org.apache.beam.sdk.io.xml.XmlIO.Sink
 
SinkMetrics - Class in org.apache.beam.sdk.metrics
Standard Sink Metrics.
SinkMetrics() - Constructor for class org.apache.beam.sdk.metrics.SinkMetrics
 
sinkViaGenericRecords(Schema, AvroIO.RecordFormatter<ElementT>) - Static method in class org.apache.beam.sdk.io.AvroIO
Deprecated.
RecordFormatter will be removed in future versions.
size() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
Returns the number of bytes in the backing array that are valid.
size() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedList
 
size() - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedRow
 
size() - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
 
size() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
size() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
Returns the number of columns for this schema.
size() - Method in class org.apache.beam.sdk.values.PCollectionList
Returns the number of PCollections in this PCollectionList.
size() - Method in class org.apache.beam.sdk.values.TupleTagList
Returns the number of TupleTags in this TupleTagList.
sizeBytes() - Method in class org.apache.beam.sdk.io.fs.MatchResult.Metadata
 
Sizes - Class in org.apache.beam.sdk.transforms.splittabledofn
Definitions and convenience methods for reporting sizes for SplittableDoFns.
Sizes() - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.Sizes
 
Sizes.HasSize - Interface in org.apache.beam.sdk.transforms.splittabledofn
RestrictionTrackers which can provide a size should implement this interface.
Sketch() - Constructor for class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.Sketch
 
SketchFrequencies - Class in org.apache.beam.sdk.extensions.sketching
PTransforms to compute the estimate frequency of each element in a stream.
SketchFrequencies() - Constructor for class org.apache.beam.sdk.extensions.sketching.SketchFrequencies
 
SketchFrequencies.CountMinSketchFn<InputT> - Class in org.apache.beam.sdk.extensions.sketching
Implements the Combine.CombineFn of SketchFrequencies transforms.
SketchFrequencies.GlobalSketch<InputT> - Class in org.apache.beam.sdk.extensions.sketching
Implementation of SketchFrequencies.globally().
SketchFrequencies.PerKeySketch<K,V> - Class in org.apache.beam.sdk.extensions.sketching
Implementation of SketchFrequencies.perKey().
SketchFrequencies.Sketch<T> - Class in org.apache.beam.sdk.extensions.sketching
Wrap StreamLib's Count-Min Sketch to support counting all user types by hashing the encoded user type using the supplied deterministic coder.
skipAssignWindows(Window.Assign<T>, TranslationContext) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.WindowingHelpers
Checks if the window transformation should be applied or skipped.
skipInvalidRows() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Insert all valid rows of a request, even if invalid rows exist.
Slf4jLogWriter - Class in org.apache.beam.runners.fnexecution.logging
A LogWriter which uses an SLF4J Logger as the underlying log backend.
SlidingWindows - Class in org.apache.beam.sdk.transforms.windowing
A WindowFn that windows values into possibly overlapping fixed-size timestamp-based windows.
SMALL_INT - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
smallest(int) - Static method in class org.apache.beam.sdk.transforms.Top
Returns a PTransform that takes an input PCollection<T> and returns a PCollection<List<T>> with a single element containing the smallest count elements of the input PCollection<T>, in increasing order, sorted according to their natural order.
Smallest() - Constructor for class org.apache.beam.sdk.transforms.Top.Smallest
Deprecated.
 
smallestDoublesFn(int) - Static method in class org.apache.beam.sdk.transforms.Top
Returns a Top.TopCombineFn that aggregates the smallest count double values.
smallestFn(int) - Static method in class org.apache.beam.sdk.transforms.Top
Returns a Top.TopCombineFn that aggregates the smallest count values.
smallestIntsFn(int) - Static method in class org.apache.beam.sdk.transforms.Top
Returns a Top.TopCombineFn that aggregates the smallest count int values.
smallestLongsFn(int) - Static method in class org.apache.beam.sdk.transforms.Top
Returns a Top.TopCombineFn that aggregates the smallest count long values.
smallestPerKey(int) - Static method in class org.apache.beam.sdk.transforms.Top
Returns a PTransform that takes an input PCollection<KV<K, V>> and returns a PCollection<KV<K, List<V>>> that contains an output element mapping each distinct key in the input PCollection to the smallest count values associated with that key in the input PCollection<KV<K, V>>, in increasing order, sorted according to their natural order.
SnappyCoder<T> - Class in org.apache.beam.sdk.coders
Wraps an existing coder with Snappy compression.
snapshot(SchemaVersion) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteSchema
 
SnsClientProvider - Interface in org.apache.beam.sdk.io.aws2.sns
Provides instances of DynamoDB clients.
SnsCoderProviderRegistrar - Class in org.apache.beam.sdk.io.aws.sns
A CoderProviderRegistrar for standard types used with SnsIO.
SnsCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.aws.sns.SnsCoderProviderRegistrar
 
SnsCoderProviderRegistrar - Class in org.apache.beam.sdk.io.aws2.sns
A CoderProviderRegistrar for standard types used with SnsIO.
SnsCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.aws2.sns.SnsCoderProviderRegistrar
 
SnsIO - Class in org.apache.beam.sdk.io.aws.sns
PTransforms for writing to SNS.
SnsIO() - Constructor for class org.apache.beam.sdk.io.aws.sns.SnsIO
 
SnsIO - Class in org.apache.beam.sdk.io.aws2.sns
PTransforms for writing to SNS.
SnsIO() - Constructor for class org.apache.beam.sdk.io.aws2.sns.SnsIO
 
SnsIO.RetryConfiguration - Class in org.apache.beam.sdk.io.aws.sns
A POJO encapsulating a configuration for retry behavior when issuing requests to SNS.
SnsIO.RetryConfiguration - Class in org.apache.beam.sdk.io.aws2.sns
A POJO encapsulating a configuration for retry behavior when issuing requests to SNS.
SnsIO.Write - Class in org.apache.beam.sdk.io.aws.sns
Implementation of SnsIO.write().
SnsIO.Write<T> - Class in org.apache.beam.sdk.io.aws2.sns
Implementation of SnsIO.write().
SocketAddressFactory - Class in org.apache.beam.sdk.fn.channel
Creates a SocketAddress based upon a supplied string.
SocketAddressFactory() - Constructor for class org.apache.beam.sdk.fn.channel.SocketAddressFactory
 
SolrIO - Class in org.apache.beam.sdk.io.solr
Transforms for reading and writing data from/to Solr.
SolrIO.ConnectionConfiguration - Class in org.apache.beam.sdk.io.solr
A POJO describing a connection configuration to Solr.
SolrIO.Read - Class in org.apache.beam.sdk.io.solr
A PTransform reading data from Solr.
SolrIO.RetryConfiguration - Class in org.apache.beam.sdk.io.solr
A POJO encapsulating a configuration for retry behavior when issuing requests to Solr.
SolrIO.Write - Class in org.apache.beam.sdk.io.solr
A PTransform writing data to Solr.
sort() - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter
 
SORT_VALUES - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
sortBySchema(List<FieldValueTypeInformation>, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.StaticSchemaInference
 
SortValues<PrimaryKeyT,SecondaryKeyT,ValueT> - Class in org.apache.beam.sdk.extensions.sorter
SortValues<PrimaryKeyT, SecondaryKeyT, ValueT> takes a PCollection<KV<PrimaryKeyT, Iterable<KV<SecondaryKeyT, ValueT>>>> with elements consisting of a primary key and iterables over <secondary key, value> pairs, and returns a PCollection<KV<PrimaryKeyT, Iterable<KV<SecondaryKeyT, ValueT>>> of the same elements but with values sorted by a secondary key.
SOURCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.CalciteQueryPlanner.NonCumulativeCostImpl
 
SOURCE - Static variable in class org.apache.beam.sdk.extensions.sql.impl.planner.RelMdNodeStats
 
Source<T> - Class in org.apache.beam.sdk.io
Base class for defining input formats and creating a Source for reading the input.
Source() - Constructor for class org.apache.beam.sdk.io.Source
 
Source.Reader<T> - Class in org.apache.beam.sdk.io
The interface that readers of custom input sources must implement.
SOURCE_DOES_NOT_NEED_SPLITTING - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
SOURCE_ESTIMATED_SIZE_BYTES - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
SOURCE_IS_INFINITE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
SOURCE_METADATA - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
SOURCE_SPEC - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
SOURCE_STEP_INPUT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
SourceMetrics - Class in org.apache.beam.sdk.metrics
Standard Source Metrics.
SourceMetrics() - Constructor for class org.apache.beam.sdk.metrics.SourceMetrics
 
sourceName() - Method in class org.apache.beam.runners.spark.metrics.AggregatorMetricSource
 
sourceName() - Method in class org.apache.beam.runners.spark.metrics.CompositeSource
 
sourceName() - Method in class org.apache.beam.runners.spark.metrics.SparkBeamMetricSource
 
sourceName() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.AggregatorMetricSource
 
sourceName() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.CompositeSource
 
sourceName() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.SparkBeamMetricSource
 
SourceRDD - Class in org.apache.beam.runners.spark.io
Classes implementing Beam Source RDDs.
SourceRDD() - Constructor for class org.apache.beam.runners.spark.io.SourceRDD
 
SourceRDD.Bounded<T> - Class in org.apache.beam.runners.spark.io
A SourceRDD.Bounded reads input from a BoundedSource and creates a Spark RDD.
SourceRDD.Unbounded<T,CheckpointMarkT extends UnboundedSource.CheckpointMark> - Class in org.apache.beam.runners.spark.io
A SourceRDD.Unbounded is the implementation of a micro-batch in a SourceDStream.
SourceTestUtils - Class in org.apache.beam.sdk.testing
Helper functions and test harnesses for checking correctness of Source implementations.
SourceTestUtils() - Constructor for class org.apache.beam.sdk.testing.SourceTestUtils
 
SourceTestUtils.ExpectedSplitOutcome - Enum in org.apache.beam.sdk.testing
span(IntervalWindow) - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
Returns the minimal window that includes both this window and the given window.
SpannerConfig - Class in org.apache.beam.sdk.io.gcp.spanner
Configuration for a Cloud Spanner client.
SpannerConfig() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
SpannerConfig.Builder - Class in org.apache.beam.sdk.io.gcp.spanner
Builder for SpannerConfig.
SpannerIO - Class in org.apache.beam.sdk.io.gcp.spanner
Experimental Transforms for reading from and writing to Google Cloud Spanner.
SpannerIO.CreateTransaction - Class in org.apache.beam.sdk.io.gcp.spanner
A PTransform that create a transaction.
SpannerIO.CreateTransaction.Builder - Class in org.apache.beam.sdk.io.gcp.spanner
SpannerIO.FailureMode - Enum in org.apache.beam.sdk.io.gcp.spanner
A failure handling strategy.
SpannerIO.Read - Class in org.apache.beam.sdk.io.gcp.spanner
Implementation of SpannerIO.read().
SpannerIO.ReadAll - Class in org.apache.beam.sdk.io.gcp.spanner
Implementation of SpannerIO.readAll().
SpannerIO.Write - Class in org.apache.beam.sdk.io.gcp.spanner
A PTransform that writes Mutation objects to Google Cloud Spanner.
SpannerIO.WriteGrouped - Class in org.apache.beam.sdk.io.gcp.spanner
Same as SpannerIO.Write but supports grouped mutations.
SpannerWriteResult - Class in org.apache.beam.sdk.io.gcp.spanner
The results of a SpannerIO.write() transform.
SpannerWriteResult(Pipeline, PCollection<Void>, PCollection<MutationGroup>, TupleTag<MutationGroup>) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerWriteResult
 
SparkBeamMetricSource - Class in org.apache.beam.runners.spark.metrics
A Spark Source that is tailored to expose a SparkBeamMetric, wrapping an underlying MetricResults instance.
SparkBeamMetricSource(String) - Constructor for class org.apache.beam.runners.spark.metrics.SparkBeamMetricSource
 
SparkBeamMetricSource - Class in org.apache.beam.runners.spark.structuredstreaming.metrics
A Spark Source that is tailored to expose a SparkBeamMetric, wrapping an underlying MetricResults instance.
SparkBeamMetricSource(String) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.metrics.SparkBeamMetricSource
 
SparkCommonPipelineOptions - Interface in org.apache.beam.runners.spark
Spark runner PipelineOptions handles Spark execution-related configurations, such as the master address, and other user-related knobs.
SparkCommonPipelineOptions.TmpCheckpointDirFactory - Class in org.apache.beam.runners.spark
Returns the default checkpoint directory of /tmp/${job.name}.
SparkCompat - Class in org.apache.beam.runners.spark.util
A set of functions to provide API compatibility between Spark 2 and Spark 3.
SparkCompat() - Constructor for class org.apache.beam.runners.spark.util.SparkCompat
 
SparkContextOptions - Interface in org.apache.beam.runners.spark
A custom PipelineOptions to work with properties related to JavaSparkContext.
SparkContextOptions.EmptyListenersList - Class in org.apache.beam.runners.spark
Returns an empty list, to avoid handling null.
SparkGroupAlsoByWindowViaWindowSet - Class in org.apache.beam.runners.spark.stateful
An implementation of GroupByKeyViaGroupByKeyOnly.GroupAlsoByWindow logic for grouping by windows and controlling trigger firings and pane accumulation.
SparkGroupAlsoByWindowViaWindowSet() - Constructor for class org.apache.beam.runners.spark.stateful.SparkGroupAlsoByWindowViaWindowSet
 
SparkGroupAlsoByWindowViaWindowSet.StateAndTimers - Class in org.apache.beam.runners.spark.stateful
State and Timers wrapper.
SparkJobInvoker - Class in org.apache.beam.runners.spark
Creates a job invocation to manage the Spark runner's execution of a portable pipeline.
SparkJobServerDriver - Class in org.apache.beam.runners.spark
Driver program that starts a job server for the Spark runner.
SparkJobServerDriver.SparkServerConfiguration - Class in org.apache.beam.runners.spark
Spark runner-specific Configuration for the jobServer.
SparkNativePipelineVisitor - Class in org.apache.beam.runners.spark
Pipeline visitor for translating a Beam pipeline into equivalent Spark operations.
SparkPipelineOptions - Interface in org.apache.beam.runners.spark
Spark runner PipelineOptions handles Spark execution-related configurations, such as the master address, batch-interval, and other user-related knobs.
SparkPipelineResult - Class in org.apache.beam.runners.spark
Represents a Spark pipeline execution result.
SparkPipelineRunner - Class in org.apache.beam.runners.spark
Runs a portable pipeline on Apache Spark.
SparkPipelineRunner(SparkPipelineOptions) - Constructor for class org.apache.beam.runners.spark.SparkPipelineRunner
 
SparkRunner - Class in org.apache.beam.runners.spark
The SparkRunner translate operations defined on a pipeline to a representation executable by Spark, and then submitting the job to Spark to be executed.
SparkRunner.Evaluator - Class in org.apache.beam.runners.spark
Evaluator on the pipeline.
SparkRunnerDebugger - Class in org.apache.beam.runners.spark
Pipeline runner which translates a Beam pipeline into equivalent Spark operations, without running them.
SparkRunnerDebugger.DebugSparkPipelineResult - Class in org.apache.beam.runners.spark
PipelineResult of running a Pipeline using SparkRunnerDebugger Use SparkRunnerDebugger.DebugSparkPipelineResult.getDebugString() to get a String representation of the Pipeline translated into Spark native operations.
SparkRunnerKryoRegistrator - Class in org.apache.beam.runners.spark.coders
Custom KryoRegistrators for Beam's Spark runner needs and registering used class in spark translation for better serialization performance.
SparkRunnerKryoRegistrator() - Constructor for class org.apache.beam.runners.spark.coders.SparkRunnerKryoRegistrator
 
SparkRunnerRegistrar - Class in org.apache.beam.runners.spark
Contains the PipelineRunnerRegistrar and PipelineOptionsRegistrar for the SparkRunner.
SparkRunnerRegistrar.Options - Class in org.apache.beam.runners.spark
Registers the SparkPipelineOptions.
SparkRunnerRegistrar.Runner - Class in org.apache.beam.runners.spark
Registers the SparkRunner.
SparkServerConfiguration() - Constructor for class org.apache.beam.runners.spark.SparkJobServerDriver.SparkServerConfiguration
 
SparkSideInputReader - Class in org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions
A SideInputReader for the Spark Batch Runner.
SparkSideInputReader(Map<PCollectionView<?>, WindowingStrategy<?, ?>>, SideInputBroadcast) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.SparkSideInputReader
 
SparkSideInputReader - Class in org.apache.beam.runners.spark.util
A SideInputReader for the SparkRunner.
SparkSideInputReader(Map<TupleTag<?>, KV<WindowingStrategy<?, ?>, SideInputBroadcast<?>>>) - Constructor for class org.apache.beam.runners.spark.util.SparkSideInputReader
 
SparkStructuredStreamingPipelineOptions - Interface in org.apache.beam.runners.spark.structuredstreaming
Spark runner PipelineOptions handles Spark execution-related configurations, such as the master address, and other user-related knobs.
SparkStructuredStreamingPipelineResult - Class in org.apache.beam.runners.spark.structuredstreaming
Represents a Spark pipeline execution result.
SparkStructuredStreamingRunner - Class in org.apache.beam.runners.spark.structuredstreaming
SparkStructuredStreamingRunner is based on spark structured streaming framework and is no more based on RDD/DStream API.
SparkTimerInternals - Class in org.apache.beam.runners.spark.stateful
An implementation of TimerInternals for the SparkRunner.
SparkUnboundedSource - Class in org.apache.beam.runners.spark.io
A "composite" InputDStream implementation for UnboundedSources.
SparkUnboundedSource() - Constructor for class org.apache.beam.runners.spark.io.SparkUnboundedSource
 
SparkUnboundedSource.Metadata - Class in org.apache.beam.runners.spark.io
A metadata holder for an input stream partition.
SparkWatermarks(Instant, Instant, Instant) - Constructor for class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.SparkWatermarks
 
split(BeamFnApi.ProcessBundleSplitResponse) - Method in interface org.apache.beam.runners.fnexecution.control.BundleSplitHandler
 
split(double) - Method in interface org.apache.beam.runners.fnexecution.control.RemoteBundle
Ask the remote bundle to split its current processing based upon its knowledge of remaining work.
split(double) - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient.BundleProcessor.ActiveBundle
 
split(List<BeamFnApi.BundleApplication>, List<BeamFnApi.DelayedBundleApplication>) - Method in class org.apache.beam.runners.fnexecution.splittabledofn.SDFFeederViaStateAndTimers
Signals that a split happened.
split(int, PipelineOptions) - Method in class org.apache.beam.runners.gearpump.translators.io.ValuesSource
 
split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.BoundedSource
Splits the source into bundles of approximately desiredBundleSizeBytes.
split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
 
split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSource
 
split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource
 
split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
 
split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read.BoundedGridFSSource
 
split(long, PipelineOptions) - Method in class org.apache.beam.sdk.io.OffsetBasedSource
 
split(int) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
Returns a list of up to numSplits + 1 ByteKeys in ascending order, where the keys have been interpolated to form roughly equal sub-ranges of this ByteKeyRange, assuming a uniform distribution of keys within this range.
split(long, long) - Method in class org.apache.beam.sdk.io.range.OffsetRange
 
split(int, PipelineOptions) - Method in class org.apache.beam.sdk.io.UnboundedSource
Returns a list of UnboundedSource objects representing the instances of this source that should be used when executing the workflow.
split(String) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.Split PTransform that splits a string on the regular expression and then outputs each item.
split(Pattern) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.Split PTransform that splits a string on the regular expression and then outputs each item.
split(String, boolean) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.Split PTransform that splits a string on the regular expression and then outputs each item.
split(Pattern, boolean) - Static method in class org.apache.beam.sdk.transforms.Regex
Returns a Regex.Split PTransform that splits a string on the regular expression and then outputs each item.
Split(Pattern, boolean) - Constructor for class org.apache.beam.sdk.transforms.Regex.Split
 
SPLIT_POINTS_UNKNOWN - Static variable in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
A constant to use as the return value for BoundedSource.BoundedReader.getSplitPointsConsumed() or BoundedSource.BoundedReader.getSplitPointsRemaining() when the exact value is unknown.
splitAtFraction(double) - Method in class org.apache.beam.sdk.io.BoundedSource.BoundedReader
Tells the reader to narrow the range of the input it's going to read and give up the remainder, so that the new range would contain approximately the given fraction of the amount of data in the current range.
splitAtFraction(double) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource.BigQueryStorageStreamReader
 
splitAtFraction(double) - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
 
splitReadStream(Storage.SplitReadStreamRequest) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.StorageClient
 
SplitResult<RestrictionT> - Class in org.apache.beam.sdk.transforms.splittabledofn
A representation of a split result.
SplitResult() - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.SplitResult
 
SqlCaseWithValueOperatorRewriter - Class in org.apache.beam.sdk.extensions.sql.zetasql
Rewrites $case_with_value calls as $case_no_value calls.
SqlCaseWithValueOperatorRewriter() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.SqlCaseWithValueOperatorRewriter
 
SqlCheckConstraint - Class in org.apache.beam.sdk.extensions.sql.impl.parser
Parse tree for UNIQUE, PRIMARY KEY constraints.
SqlCoalesceOperatorRewriter - Class in org.apache.beam.sdk.extensions.sql.zetasql
Rewrites COALESCE calls as CASE ($case_no_value) calls.
SqlCoalesceOperatorRewriter() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.SqlCoalesceOperatorRewriter
 
SqlColumnDeclaration - Class in org.apache.beam.sdk.extensions.sql.impl.parser
Parse tree for column.
SqlConversionException - Exception in org.apache.beam.sdk.extensions.sql.impl
Exception thrown when BeamSQL cannot convert sql to BeamRelNode.
SqlConversionException(Throwable) - Constructor for exception org.apache.beam.sdk.extensions.sql.impl.SqlConversionException
 
SqlConversionException(String) - Constructor for exception org.apache.beam.sdk.extensions.sql.impl.SqlConversionException
 
SqlConversionException(String, Throwable) - Constructor for exception org.apache.beam.sdk.extensions.sql.impl.SqlConversionException
 
SqlCreateExternalTable - Class in org.apache.beam.sdk.extensions.sql.impl.parser
Parse tree for CREATE EXTERNAL TABLE statement.
SqlCreateExternalTable(SqlParserPos, boolean, boolean, SqlIdentifier, List<Schema.Field>, SqlNode, SqlNode, SqlNode, SqlNode) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.parser.SqlCreateExternalTable
Creates a SqlCreateExternalTable.
SqlDdlNodes - Class in org.apache.beam.sdk.extensions.sql.impl.parser
Utilities concerning SqlNode for DDL.
SqlDropTable - Class in org.apache.beam.sdk.extensions.sql.impl.parser
Parse tree for DROP TABLE statement.
SqlExtractTimestampOperatorRewriter - Class in org.apache.beam.sdk.extensions.sql.zetasql
Rewrites EXTRACT calls by swapping first two arguments to fit for calcite SqlExtractOperator.
SqlExtractTimestampOperatorRewriter() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.SqlExtractTimestampOperatorRewriter
 
SqlIfNullOperatorRewriter - Class in org.apache.beam.sdk.extensions.sql.zetasql
Rewrites IFNULL calls as CASE ($case_no_value) calls.
SqlIfNullOperatorRewriter() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.SqlIfNullOperatorRewriter
 
SqlNullIfOperatorRewriter - Class in org.apache.beam.sdk.extensions.sql.zetasql
Rewrites NULLIF calls as CASE ($case_no_value) calls.
SqlNullIfOperatorRewriter() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.SqlNullIfOperatorRewriter
 
SqlOperatorRewriter - Interface in org.apache.beam.sdk.extensions.sql.zetasql
Interface for rewriting calls a specific ZetaSQL operator.
SqlOperators - Class in org.apache.beam.sdk.extensions.sql.zetasql
A separate SqlOperators table for those functions that do not exist or not compatible with Calcite.
SqlOperators() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
SqlSetOptionBeam - Class in org.apache.beam.sdk.extensions.sql.impl.parser
SQL parse tree node to represent SET and RESET statements.
SqlSetOptionBeam(SqlParserPos, String, SqlIdentifier, SqlNode) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.parser.SqlSetOptionBeam
 
SqlStdOperatorMappingTable - Class in org.apache.beam.sdk.extensions.sql.zetasql
SqlStdOperatorMappingTable.
SqlStdOperatorMappingTable() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.SqlStdOperatorMappingTable
 
SqlTransform - Class in org.apache.beam.sdk.extensions.sql
SqlTransform is the DSL interface of Beam SQL.
SqlTransform() - Constructor for class org.apache.beam.sdk.extensions.sql.SqlTransform
 
sqlTypeWithAutoCast(RelDataTypeFactory, Type) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
SQL-Java type mapping, with specified Beam rules:
1.
SqsClientProvider - Interface in org.apache.beam.sdk.io.aws2.sqs
Provides instances of Sqs clients.
SqsIO - Class in org.apache.beam.sdk.io.aws.sqs
An unbounded source for Amazon Simple Queue Service (SQS).
SqsIO - Class in org.apache.beam.sdk.io.aws2.sqs
An unbounded source for Amazon Simple Queue Service (SQS).
SqsIO.Read - Class in org.apache.beam.sdk.io.aws.sqs
A PTransform to read/receive messages from SQS.
SqsIO.Read - Class in org.apache.beam.sdk.io.aws2.sqs
A PTransform to read/receive messages from SQS.
SqsIO.Write - Class in org.apache.beam.sdk.io.aws.sqs
A PTransform to send messages to SQS.
SqsIO.Write - Class in org.apache.beam.sdk.io.aws2.sqs
A PTransform to send messages to SQS.
SqsMessage - Class in org.apache.beam.sdk.io.aws2.sqs
 
SqsMessage() - Constructor for class org.apache.beam.sdk.io.aws2.sqs.SqsMessage
 
StageBundleFactory - Interface in org.apache.beam.runners.fnexecution.control
A bundle factory scoped to a particular ExecutableStage, which has all of the resources it needs to provide new RemoteBundles.
stageDefaultFiles() - Method in class org.apache.beam.runners.dataflow.util.GcsStager
Stages DataflowPipelineWorkerPoolOptions.getFilesToStage(), which defaults to every file on the classpath unless overridden, as well as DataflowPipelineDebugOptions.getOverrideWindmillBinary() if specified.
stageDefaultFiles() - Method in interface org.apache.beam.runners.dataflow.util.Stager
Stage default files and return a list of DataflowPackage objects describing the actual location at which each file was staged.
stageFiles(List<String>) - Method in class org.apache.beam.runners.dataflow.util.GcsStager
Stages files to DataflowPipelineOptions.getStagingLocation(), suffixed with their md5 hash to avoid collisions.
stageFiles(List<String>) - Method in interface org.apache.beam.runners.dataflow.util.Stager
Stage files and return a list of packages DataflowPackage objects describing th actual location at which each file was staged.
Stager - Interface in org.apache.beam.runners.dataflow.util
Interface for staging files needed for running a Dataflow pipeline.
StagerFactory() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowPipelineDebugOptions.StagerFactory
 
stageToFile(byte[], String) - Method in class org.apache.beam.runners.dataflow.util.GcsStager
 
stageToFile(byte[], String) - Method in interface org.apache.beam.runners.dataflow.util.Stager
Stage bytes to a target file name wherever this stager stages things.
StagingLocationFactory() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowPipelineOptions.StagingLocationFactory
 
StagingSessionToken() - Constructor for class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService.StagingSessionToken
 
StandardCreateOptions() - Constructor for class org.apache.beam.sdk.io.fs.CreateOptions.StandardCreateOptions
 
start() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation
Start the job.
start() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver
 
start() - Method in class org.apache.beam.runners.portability.ExternalWorkerService
 
start() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource.Reader
 
start() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource.BigQueryStorageStreamReader
 
start() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
 
start() - Method in class org.apache.beam.sdk.io.Source.Reader
Initializes the reader and advances the reader to the first record.
start() - Method in class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
Initializes the reader and advances the reader to the first record.
start() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
Returns the start of this window, inclusive.
START_WITHS - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
startBundle() - Method in class org.apache.beam.runners.flink.metrics.DoFnRunnerWithMetricsUpdate
 
startBundle() - Method in class org.apache.beam.sdk.extensions.ml.AnnotateVideo
 
startBundle(DoFn<T, Void>.StartBundleContext) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write.PubsubBoundedWriter
 
startBundle() - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
StartBundleContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.StartBundleContext
 
startCopyJob(JobReference, JobConfigurationTableCopy) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.JobService
Start a BigQuery copy job.
startCopyJob(JobReference, JobConfigurationTableCopy) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
 
startExtractJob(JobReference, JobConfigurationExtract) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.JobService
Start a BigQuery extract job.
startExtractJob(JobReference, JobConfigurationExtract) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
 
startImpl() - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
 
startImpl() - Method in class org.apache.beam.sdk.io.OffsetBasedSource.OffsetBasedReader
Initializes the OffsetBasedSource.OffsetBasedReader and advances to the first record, returning true if there is a record available to be read.
startLoadJob(JobReference, JobConfigurationLoad) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.JobService
Start a BigQuery load job.
startLoadJob(JobReference, JobConfigurationLoad) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
 
startPipeline() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
Starts the pipeline.
startProcess(String, String, List<String>, Map<String, String>) - Method in class org.apache.beam.runners.fnexecution.environment.ProcessManager
Forks a process with the given command, arguments, and additional environment variables.
startProcess(String, String, List<String>, Map<String, String>, File) - Method in class org.apache.beam.runners.fnexecution.environment.ProcessManager
 
startQueryJob(JobReference, JobConfigurationQuery) - Method in interface org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices.JobService
Start a BigQuery query job.
startQueryJob(JobReference, JobConfigurationQuery) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeJobService
 
startReading(ReadableByteChannel) - Method in class org.apache.beam.sdk.io.AvroSource.AvroReader
 
startReading(ReadableByteChannel) - Method in class org.apache.beam.sdk.io.CompressedSource.CompressedReader
Creates a decompressing channel from the input channel and passes it to its delegate reader's FileBasedReader#startReading(ReadableByteChannel).
startReading(ReadableByteChannel) - Method in class org.apache.beam.sdk.io.FileBasedSource.FileBasedReader
Performs any initialization of the subclass of FileBasedReader that involves IO operations.
startRunnerBundle(DoFnRunner<KV<?, ?>, OutputT>) - Method in class org.apache.beam.runners.jet.processors.StatefulParDoP
 
STARTS_WITH_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamBuiltinMethods
 
startsWith(Path) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
startsWith(String) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
startsWith(String, String) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
startsWith(String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
startsWith(String) - Static method in class org.apache.beam.sdk.testing.SerializableMatchers
startWorker(BeamFnApi.StartWorkerRequest, StreamObserver<BeamFnApi.StartWorkerResponse>) - Method in class org.apache.beam.runners.portability.ExternalWorkerService
 
state(StreamObserver<BeamFnApi.StateResponse>) - Method in class org.apache.beam.runners.fnexecution.state.GrpcStateService
 
State - Interface in org.apache.beam.sdk.state
A state cell, supporting a State.clear() operation.
STATE_CACHE_SIZE - Static variable in interface org.apache.beam.sdk.options.ExperimentalOptions
 
StateBinder - Interface in org.apache.beam.sdk.state
For internal use only; no backwards-compatibility guarantees.
StateContext<W extends BoundedWindow> - Interface in org.apache.beam.sdk.state
For internal use only; no backwards-compatibility guarantees.
StateContexts - Class in org.apache.beam.sdk.state
For internal use only; no backwards-compatibility guarantees.
StateContexts() - Constructor for class org.apache.beam.sdk.state.StateContexts
 
StateDelegator - Interface in org.apache.beam.runners.fnexecution.state
The StateDelegator is able to delegate BeamFnApi.StateRequests to a set of registered handlers.
StateDelegator.Registration - Interface in org.apache.beam.runners.fnexecution.state
Allows callers to deregister from receiving further state requests.
StatefulParDoP<OutputT> - Class in org.apache.beam.runners.jet.processors
Jet Processor implementation for Beam's stateful ParDo primitive.
StatefulParDoP.Supplier<OutputT> - Class in org.apache.beam.runners.jet.processors
Jet Processor supplier that will provide instances of StatefulParDoP.
stateInternals() - Method in class org.apache.beam.runners.gearpump.translators.utils.NoOpStepContext
 
stateInternals() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.NoOpStepContext
 
StateRequestHandler - Interface in org.apache.beam.runners.fnexecution.state
Handler for StateRequests.
StateRequestHandlers - Class in org.apache.beam.runners.fnexecution.state
A set of utility methods which construct StateRequestHandlers.
StateRequestHandlers() - Constructor for class org.apache.beam.runners.fnexecution.state.StateRequestHandlers
 
StateRequestHandlers.BagUserStateHandler<K,V,W extends BoundedWindow> - Interface in org.apache.beam.runners.fnexecution.state
A handler for bag user state.
StateRequestHandlers.BagUserStateHandlerFactory<K,V,W extends BoundedWindow> - Interface in org.apache.beam.runners.fnexecution.state
A factory which constructs StateRequestHandlers.BagUserStateHandlers.
StateRequestHandlers.IterableSideInputHandler<V,W extends BoundedWindow> - Interface in org.apache.beam.runners.fnexecution.state
A handler for iterable side inputs.
StateRequestHandlers.MultimapSideInputHandler<K,V,W extends BoundedWindow> - Interface in org.apache.beam.runners.fnexecution.state
A handler for multimap side inputs.
StateRequestHandlers.SideInputHandler - Interface in org.apache.beam.runners.fnexecution.state
Marker interface that denotes some type of side input handler.
StateRequestHandlers.SideInputHandlerFactory - Interface in org.apache.beam.runners.fnexecution.state
StateSpec<StateT extends State> - Interface in org.apache.beam.sdk.state
A specification of a persistent state cell.
StateSpec.Cases<ResultT> - Interface in org.apache.beam.sdk.state
Cases for doing a "switch" on the type of StateSpec.
StateSpec.Cases.WithDefault<ResultT> - Class in org.apache.beam.sdk.state
A base class for a visitor with a default method for cases it is not interested in.
StateSpecFunctions - Class in org.apache.beam.runners.spark.stateful
A class containing StateSpec mappingFunctions.
StateSpecFunctions() - Constructor for class org.apache.beam.runners.spark.stateful.StateSpecFunctions
 
StateSpecs - Class in org.apache.beam.sdk.state
Static methods for working with StateSpecs.
StaticGrpcProvisionService - Class in org.apache.beam.runners.fnexecution.provisioning
A provision service that returns a static response to all calls.
StaticRemoteEnvironment - Class in org.apache.beam.runners.fnexecution.environment
A RemoteEnvironment that connects to Dataflow runner harness.
StaticRemoteEnvironmentFactory - Class in org.apache.beam.runners.fnexecution.environment
An EnvironmentFactory that creates StaticRemoteEnvironment used by a runner harness that would like to use an existing InstructionRequestHandler.
StaticRemoteEnvironmentFactory.Provider - Class in org.apache.beam.runners.fnexecution.environment
Provider for StaticRemoteEnvironmentFactory.
StaticSchemaInference - Class in org.apache.beam.sdk.schemas.utils
A set of utilities for inferring a Beam Schema from static Java types.
StaticSchemaInference() - Constructor for class org.apache.beam.sdk.schemas.utils.StaticSchemaInference
 
status() - Method in class org.apache.beam.sdk.io.fs.MatchResult
Status of the MatchResult.
STATUS_BACKOFF_FACTORY - Static variable in class org.apache.beam.runners.dataflow.DataflowPipelineJob
 
stepName() - Method in class org.apache.beam.sdk.metrics.MetricKey
The step name that is associated with this metric or Null if none is associated.
steps() - Method in class org.apache.beam.sdk.metrics.MetricsFilter
 
stop() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobServerDriver
 
stop() - Method in class org.apache.beam.runners.spark.SparkPipelineResult
 
stop() - Method in class org.apache.beam.runners.spark.SparkRunnerDebugger.DebugSparkPipelineResult
 
stop() - Method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineResult
 
stop() - Static method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
Indicates that there is no more work to be done for the current element.
stopProcess(String) - Method in class org.apache.beam.runners.fnexecution.environment.ProcessManager
Stops a previously started process identified by its unique id.
stopSampling(int, long) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.LimitNumberOfFiles
 
stopSampling(int, long) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.LimitNumberOfTotalBytes
 
stopSampling(int, long) - Method in class org.apache.beam.sdk.io.TextRowCountEstimator.SampleAllFiles
 
stopSampling(int, long) - Method in interface org.apache.beam.sdk.io.TextRowCountEstimator.SamplingStrategy
 
storageObject() - Method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.StorageObjectOrIOException
Returns the StorageObject.
StorageObjectOrIOException() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.GcsUtil.StorageObjectOrIOException
 
storeWindowedValueInRow(WindowedValue<T>, Coder<T>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.RowHelpers
Serialize a windowedValue to bytes using windowedValueCoder WindowedValue.FullWindowedValueCoder and stores it an InternalRow.
STREAMING_ENGINE_EXPERIMENT - Static variable in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
Experiment to turn on the Streaming Engine experiment.
StreamingInserts<DestinationT,ElementT> - Class in org.apache.beam.sdk.io.gcp.bigquery
PTransform that performs streaming BigQuery write.
StreamingInserts(BigQueryIO.Write.CreateDisposition, DynamicDestinations<?, DestinationT>, Coder<ElementT>, SerializableFunction<ElementT, TableRow>) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StreamingInserts
Constructor.
StreamingIT - Interface in org.apache.beam.sdk.testing
Deprecated.
tests which use unbounded PCollections should be in the category UsesUnboundedPCollections. Beyond that, it is up to the runner and test configuration to decide whether to run in streaming mode.
StreamingOptions - Interface in org.apache.beam.sdk.options
Options used to configure streaming.
StreamingWriteTables<ElementT> - Class in org.apache.beam.sdk.io.gcp.bigquery
This transform takes in key-value pairs of TableRow entries and the TableDestination it should be written to.
StreamingWriteTables() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.StreamingWriteTables
 
STRING - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
STRING - Static variable in class org.apache.beam.sdk.schemas.Schema.FieldType
The type of string fields.
STRING_TYPES - Static variable in enum org.apache.beam.sdk.schemas.Schema.TypeName
 
StringDelegateCoder<T> - Class in org.apache.beam.sdk.coders
A Coder that wraps a Coder<String> and encodes/decodes values via string representations.
StringDelegateCoder(Class<T>, TypeDescriptor<T>) - Constructor for class org.apache.beam.sdk.coders.StringDelegateCoder
 
StringFunctions - Class in org.apache.beam.sdk.extensions.sql.zetasql
StringFunctions.
StringFunctions() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
strings() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
The TypeDescriptor for String.
StringUtf8Coder - Class in org.apache.beam.sdk.coders
A Coder that encodes Strings in UTF-8 encoding.
stripGetterPrefix(String) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
 
StripIdsDoFn() - Constructor for class org.apache.beam.sdk.values.ValueWithRecordId.StripIdsDoFn
 
stripPartitionDecorator(String) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
Strip off any partition decorator information from a tablespec.
stripPrefix(String, String) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
 
stripSetterPrefix(String) - Static method in class org.apache.beam.sdk.schemas.utils.ReflectUtils
 
Structs - Class in org.apache.beam.runners.dataflow.util
A collection of static methods for manipulating datastructure representations transferred via the Dataflow API.
StructuralByteArray - Class in org.apache.beam.sdk.coders
A wrapper around a byte[] that uses structural, value-based equality rather than byte[]'s normal object identity.
StructuralByteArray(byte[]) - Constructor for class org.apache.beam.sdk.coders.StructuralByteArray
 
StructuralKey<K> - Class in org.apache.beam.runners.local
A (Key, Coder) pair that uses the structural value of the key (as provided by Coder.structuralValue(Object)) to perform equality and hashing.
structuralValue(byte[]) - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
Returns an object with an Object.equals() method that represents structural equality on the argument.
structuralValue(T) - Method in class org.apache.beam.sdk.coders.Coder
Returns an object with an Object.equals() method that represents structural equality on the argument.
structuralValue(T) - Method in class org.apache.beam.sdk.coders.DelegateCoder
Returns an object with an Object.equals() method that represents structural equality on the argument.
structuralValue(Iterable<T>) - Method in class org.apache.beam.sdk.coders.IterableCoder
 
structuralValue(KV<K, V>) - Method in class org.apache.beam.sdk.coders.KvCoder
 
structuralValue(List<T>) - Method in class org.apache.beam.sdk.coders.ListCoder
 
structuralValue(Map<K, V>) - Method in class org.apache.beam.sdk.coders.MapCoder
 
structuralValue(T) - Method in class org.apache.beam.sdk.coders.NullableCoder
 
structuralValue(T) - Method in class org.apache.beam.sdk.coders.SerializableCoder
The structural value of the object is the object itself.
structuralValue(T) - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
 
structuralValue(Void) - Method in class org.apache.beam.sdk.coders.VoidCoder
 
structuralValue(KafkaRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
 
structuralValue(ProducerRecord<K, V>) - Method in class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
 
structuralValue(TimestampedValue<T>) - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
 
structuralValueConsistentWithEquals(Coder<T>, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<T> and values of type T, the structural values are equal if and only if the encoded bytes are equal.
structuralValueConsistentWithEqualsInContext(Coder<T>, Coder.Context, T, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<T>, Coder.Context, and values of type T, the structural values are equal if and only if the encoded bytes are equal, in any Coder.Context.
structuralValueDecodeEncodeEqual(Coder<T>, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<T> and value of type T, the structural value is equal to the structural value yield by encoding and decoding the original value.
structuralValueDecodeEncodeEqualInContext(Coder<T>, Coder.Context, T) - Static method in class org.apache.beam.sdk.testing.CoderProperties
Verifies that for the given Coder<T>, Coder.Context, and value of type T, the structural value is equal to the structural value yield by encoding and decoding the original value, in any Coder.Context.
StructuredCoder<T> - Class in org.apache.beam.sdk.coders
An abstract base class to implement a Coder that defines equality, hashing, and printing via the class name and recursively using StructuredCoder.getComponents().
StructuredCoder() - Constructor for class org.apache.beam.sdk.coders.StructuredCoder
 
subclassGetterInterface(ByteBuddy, Type, Type) - Static method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils
 
subclassSetterInterface(ByteBuddy, Type, Type) - Static method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils
 
subpath(int, int) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
subPathMatches(String, String) - Static method in class org.apache.beam.sdk.metrics.MetricFiltering
subPathMatches(haystack, needle) returns true if needle represents a path within haystack.
subscriptionPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
Subscription path used to listen for messages on TestPubsub.topicPath().
subscriptionPathFromName(String, String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
 
subscriptionPathFromPath(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
 
SUBSTR - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
substr(String, long, long) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
SUBSTR_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamBuiltinMethods
 
SUBSTR_PARAMETER_EXCEED_INTEGER - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
subTriggers - Variable in class org.apache.beam.sdk.transforms.windowing.Trigger
 
subTriggers() - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
 
success(String, String, Metadata) - Static method in class org.apache.beam.sdk.io.tika.ParseResult
 
success(String, String) - Static method in class org.apache.beam.sdk.io.tika.ParseResult
 
success() - Static method in class org.apache.beam.sdk.testing.SuccessOrFailure
 
SUCCESS_COUNTER - Static variable in class org.apache.beam.sdk.testing.PAssert
 
SuccessOrFailure - Class in org.apache.beam.sdk.testing
Output of PAssert.
Sum - Class in org.apache.beam.sdk.transforms
PTransforms for computing the sum of the elements in a PCollection, or the sum of the values associated with each key in a PCollection of KVs.
supplier(Coder, Coder, WindowingStrategy<T, BoundedWindow>, String) - Static method in class org.apache.beam.runners.jet.processors.AssignWindowP
 
supplier(BoundedSource<T>, SerializablePipelineOptions, Coder, String) - Static method in class org.apache.beam.runners.jet.processors.BoundedSourceP
 
Supplier(Map<String, Coder>, Coder, String) - Constructor for class org.apache.beam.runners.jet.processors.FlattenP.Supplier
 
supplier(Coder, String) - Static method in class org.apache.beam.runners.jet.processors.ImpulseP
 
Supplier(String, String, DoFn<InputT, OutputT>, WindowingStrategy<?, ?>, DoFnSchemaInformation, SerializablePipelineOptions, TupleTag<OutputT>, Set<TupleTag<OutputT>>, Coder<InputT>, Map<PCollectionView<?>, Coder<?>>, Map<TupleTag<?>, Coder<?>>, Coder<InputT>, Map<TupleTag<?>, Coder<?>>, List<PCollectionView<?>>) - Constructor for class org.apache.beam.runners.jet.processors.ParDoP.Supplier
 
Supplier(String, String, DoFn<KV<?, ?>, OutputT>, WindowingStrategy<?, ?>, DoFnSchemaInformation, SerializablePipelineOptions, TupleTag<OutputT>, Set<TupleTag<OutputT>>, Coder<KV<?, ?>>, Map<PCollectionView<?>, Coder<?>>, Map<TupleTag<?>, Coder<?>>, Coder<KV<?, ?>>, Map<TupleTag<?>, Coder<?>>, List<PCollectionView<?>>) - Constructor for class org.apache.beam.runners.jet.processors.StatefulParDoP.Supplier
 
supplier(UnboundedSource<T, CmT>, SerializablePipelineOptions, Coder, String) - Static method in class org.apache.beam.runners.jet.processors.UnboundedSourceP
 
supplier(Coder, Coder, WindowingStrategy<?, ?>, String) - Static method in class org.apache.beam.runners.jet.processors.ViewP
 
supplier(SerializablePipelineOptions, WindowedValue.WindowedValueCoder<KV<K, V>>, Coder, WindowingStrategy, String) - Static method in class org.apache.beam.runners.jet.processors.WindowGroupP
 
supportsNestedAggregations() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamBigQuerySqlDialect
 
supportsProjects() - Method in class org.apache.beam.sdk.extensions.sql.meta.BaseBeamTable
 
supportsProjects() - Method in interface org.apache.beam.sdk.extensions.sql.meta.BeamSqlTable
Whether project push-down is supported by the IO API.
supportsProjects() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable
 
SynchronizedStreamObserver<V> - Class in org.apache.beam.sdk.fn.stream
A StreamObserver which provides synchronous access access to an underlying StreamObserver.

T

T__0 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
T__0 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
T__1 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
T__1 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
T__2 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
T__2 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
T__3 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
T__3 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
T__4 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
T__4 - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
table() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.hcatalog.HCatalogTable
 
Table - Class in org.apache.beam.sdk.extensions.sql.meta
Represents the metadata of a BeamSqlTable.
Table() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.Table
 
table() - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
 
Table.Builder - Class in org.apache.beam.sdk.extensions.sql.meta
Builder class for Table.
TABLE_ROW_ERROR_CONTAINER - Static variable in interface org.apache.beam.sdk.io.gcp.bigquery.ErrorContainer
 
TableAndQuery() - Constructor for class org.apache.beam.sdk.io.gcp.testing.BigqueryMatcher.TableAndQuery
 
TableDestination - Class in org.apache.beam.sdk.io.gcp.bigquery
Encapsulates a BigQuery table destination.
TableDestination(String, String) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
TableDestination(TableReference, String) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
TableDestination(TableReference, String, TimePartitioning) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
TableDestination(String, String, TimePartitioning) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
TableDestination(String, String, TimePartitioning, Clustering) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
TableDestination(String, String, String) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
TableDestination(TableReference, String, String) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
TableDestination(TableReference, String, String, String) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
TableDestination(String, String, String, String) - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
TableDestinationCoder - Class in org.apache.beam.sdk.io.gcp.bigquery
A coder for TableDestination objects.
TableDestinationCoderV2 - Class in org.apache.beam.sdk.io.gcp.bigquery
A Coder for TableDestination that includes time partitioning information.
TableDestinationCoderV2() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV2
 
TableDestinationCoderV3 - Class in org.apache.beam.sdk.io.gcp.bigquery
A Coder for TableDestination that includes time partitioning and clustering information.
TableDestinationCoderV3() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV3
 
TableName - Class in org.apache.beam.sdk.extensions.sql.impl
Represents a parsed table name that is specified in a FROM clause (and other places).
TableName() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.TableName
 
TableNameExtractionUtils - Class in org.apache.beam.sdk.extensions.sql
Helper class to extract table identifiers from the query.
TableNameExtractionUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.TableNameExtractionUtils
 
TableProvider - Interface in org.apache.beam.sdk.extensions.sql.meta.provider
A TableProvider handles the metadata CRUD of a specified kind of tables.
tableReference() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery
 
TableResolution - Class in org.apache.beam.sdk.extensions.sql.zetasql
Utility methods to resolve a table, given a top-level Calcite schema and a table path.
TableResolution() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.TableResolution
 
tableRowFromBeamRow() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
 
TableRowJsonCoder - Class in org.apache.beam.sdk.io.gcp.bigquery
A Coder that encodes BigQuery TableRow objects in their native JSON format.
tableRows(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
 
tableRowToBeamRow() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
 
tables() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
 
TableSchema - Class in org.apache.beam.sdk.io.clickhouse
A descriptor for ClickHouse table schema.
TableSchema() - Constructor for class org.apache.beam.sdk.io.clickhouse.TableSchema
 
TableSchema.Column - Class in org.apache.beam.sdk.io.clickhouse
A column in ClickHouse table.
TableSchema.ColumnType - Class in org.apache.beam.sdk.io.clickhouse
A descriptor for a column type.
TableSchema.DefaultType - Enum in org.apache.beam.sdk.io.clickhouse
An enumeration of possible kinds of default values in ClickHouse.
TableSchema.TypeName - Enum in org.apache.beam.sdk.io.clickhouse
An enumeration of possible types in ClickHouse.
tableSpec() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TestBigQuery
 
TableWithRows(long, Table) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider.TableWithRows
 
TaggedKeyedPCollection(TupleTag<V>, PCollection<KV<K, V>>) - Constructor for class org.apache.beam.sdk.transforms.join.KeyedPCollectionTuple.TaggedKeyedPCollection
 
TaggedPValue - Class in org.apache.beam.sdk.values
For internal use only; no backwards-compatibility guarantees.
TaggedPValue() - Constructor for class org.apache.beam.sdk.values.TaggedPValue
 
take(String, Duration) - Method in interface org.apache.beam.runners.fnexecution.control.ControlClientPool.Source
Retrieves the InstructionRequestHandler for the given worker id, blocking until available or the request times out.
takeOutputElements() - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
takeOutputElements(TupleTag<T>) - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
takeOutputElementsWithTimestamp() - Method in class org.apache.beam.sdk.transforms.DoFnTester
Deprecated.
Use TestPipeline with the DirectRunner.
tanh(Double) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinTrigonometricFunctions
TANH(X)
TDigestQuantiles - Class in org.apache.beam.sdk.extensions.sketching
PTransforms for getting information about quantiles in a stream.
TDigestQuantiles() - Constructor for class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles
 
TDigestQuantiles.GlobalDigest - Class in org.apache.beam.sdk.extensions.sketching
Implementation of TDigestQuantiles.globally().
TDigestQuantiles.PerKeyDigest<K> - Class in org.apache.beam.sdk.extensions.sketching
Implementation of TDigestQuantiles.perKey().
TDigestQuantiles.TDigestQuantilesFn - Class in org.apache.beam.sdk.extensions.sketching
Implements the Combine.CombineFn of TDigestQuantiles transforms.
teardown() - Method in class org.apache.beam.runners.dataflow.BatchStatefulParDoOverrides.BatchStatefulDoFn
 
teardown() - Method in class org.apache.beam.runners.gearpump.translators.functions.DoFnFunction
 
teardown() - Method in class org.apache.beam.sdk.extensions.ml.AnnotateVideo
 
tearDown() - Method in interface org.apache.beam.sdk.extensions.sql.BeamSqlSeekableTable
cleanup resources of the instance.
tempDirectory - Variable in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
Directory for temporary output files.
test(RunnerApi.PTransform) - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator.IsFlinkNativeTransform
 
test(RunnerApi.PTransform) - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator.IsFlinkNativeTransform
 
TEST_DATABASE - Static variable in class org.apache.beam.sdk.io.hcatalog.test.HCatalogIOTestUtils
 
TEST_FILTER - Static variable in class org.apache.beam.sdk.io.hcatalog.test.HCatalogIOTestUtils
 
TEST_RECORDS_COUNT - Static variable in class org.apache.beam.sdk.io.hcatalog.test.HCatalogIOTestUtils
 
TEST_TABLE - Static variable in class org.apache.beam.sdk.io.hcatalog.test.HCatalogIOTestUtils
 
TestApexRunner - Class in org.apache.beam.runners.apex
Apex PipelineRunner for testing.
TestBigQuery - Class in org.apache.beam.sdk.io.gcp.bigquery
Test rule which creates a new table with specified schema, with randomized name and exposes few APIs to work with it.
TestBigQuery.PollingAssertion - Interface in org.apache.beam.sdk.io.gcp.bigquery
Interface to implement a polling assertion.
TestBigQuery.RowsAssertion - Class in org.apache.beam.sdk.io.gcp.bigquery
Interface for creating a polling eventual assertion.
TestBigQueryOptions - Interface in org.apache.beam.sdk.io.gcp.bigquery
TestBoundedTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.test
Mocked table for bounded data sources.
TestBoundedTable(Schema) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestBoundedTable
 
testByteCount(Coder<T>, Coder.Context, T[]) - Static method in class org.apache.beam.sdk.testing.CoderProperties
A utility method that passes the given (unencoded) elements through coder's registerByteSizeObserver() and encode() methods, and confirms they are mutually consistent.
testCombineFn(Combine.CombineFn<InputT, AccumT, OutputT>, List<InputT>, OutputT) - Static method in class org.apache.beam.sdk.testing.CombineFnTester
Tests that the the Combine.CombineFn, when applied to the provided input, produces the provided output.
testCombineFn(Combine.CombineFn<InputT, AccumT, OutputT>, List<InputT>, Matcher<? super OutputT>) - Static method in class org.apache.beam.sdk.testing.CombineFnTester
 
TestDataflowPipelineOptions - Interface in org.apache.beam.runners.dataflow
A set of options used to configure the TestPipeline.
TestDataflowRunner - Class in org.apache.beam.runners.dataflow
TestDataflowRunner is a pipeline runner that wraps a DataflowRunner when running tests against the TestPipeline.
TestElementByteSizeObserver() - Constructor for class org.apache.beam.sdk.testing.CoderProperties.TestElementByteSizeObserver
 
TestExecutors - Class in org.apache.beam.sdk.fn.test
A TestRule that validates that all submitted tasks finished and were completed.
TestExecutors() - Constructor for class org.apache.beam.sdk.fn.test.TestExecutors
 
TestExecutors.TestExecutorService - Interface in org.apache.beam.sdk.fn.test
A union of the ExecutorService and TestRule interfaces.
TestFlinkRunner - Class in org.apache.beam.runners.flink
Test Flink runner.
TestGearpumpRunner - Class in org.apache.beam.runners.gearpump
Gearpump PipelineRunner for tests, which uses EmbeddedCluster.
testingPipelineOptions() - Static method in class org.apache.beam.sdk.testing.TestPipeline
Creates PipelineOptions for testing.
TestJobService - Class in org.apache.beam.runners.portability.testing
A JobService for tests.
TestJobService(Endpoints.ApiServiceDescriptor, String, String, JobApi.JobState.Enum, JobApi.MetricResults) - Constructor for class org.apache.beam.runners.portability.testing.TestJobService
 
TestPipeline - Class in org.apache.beam.sdk.testing
A creator of test pipelines that can be used inside of tests that can be configured to run locally or against a remote pipeline runner.
TestPipeline.AbandonedNodeException - Exception in org.apache.beam.sdk.testing
An exception thrown in case an abandoned PTransform is detected, that is, a PTransform that has not been run.
TestPipeline.PipelineRunMissingException - Exception in org.apache.beam.sdk.testing
An exception thrown in case a test finishes without invoking Pipeline.run().
TestPipeline.TestValueProviderOptions - Interface in org.apache.beam.sdk.testing
Implementation detail of TestPipeline.newProvider(T), do not use.
TestPipelineOptions - Interface in org.apache.beam.sdk.testing
TestPipelineOptions is a set of options for test pipelines.
TestPipelineOptions.AlwaysPassMatcher - Class in org.apache.beam.sdk.testing
Matcher which will always pass.
TestPipelineOptions.AlwaysPassMatcherFactory - Class in org.apache.beam.sdk.testing
Factory for PipelineResult matchers which always pass.
TestPortablePipelineOptions - Interface in org.apache.beam.runners.portability.testing
Options for TestPortableRunner.
TestPortablePipelineOptions.DefaultJobServerConfigFactory - Class in org.apache.beam.runners.portability.testing
Factory for default config.
TestPortablePipelineOptions.TestPortablePipelineOptionsRegistrar - Class in org.apache.beam.runners.portability.testing
TestPortablePipelineOptionsRegistrar() - Constructor for class org.apache.beam.runners.portability.testing.TestPortablePipelineOptions.TestPortablePipelineOptionsRegistrar
 
TestPortableRunner - Class in org.apache.beam.runners.portability.testing
TestPortableRunner is a pipeline runner that wraps a PortableRunner when running tests against the TestPipeline.
TestPubsub - Class in org.apache.beam.sdk.io.gcp.pubsub
Test rule which creates a new topic and subscription with randomized names and exposes the APIs to work with them.
TestPubsub.PollingAssertion - Interface in org.apache.beam.sdk.io.gcp.pubsub
 
TestPubsubOptions - Interface in org.apache.beam.sdk.io.gcp.pubsub
TestPubsubSignal - Class in org.apache.beam.sdk.io.gcp.pubsub
Test rule which observes elements of the PCollection and checks whether they match the success criteria.
TestSparkPipelineOptions - Interface in org.apache.beam.runners.spark
TestSparkPipelineOptions.DefaultStopPipelineWatermarkFactory - Class in org.apache.beam.runners.spark
A factory to provide the default watermark to stop a pipeline that reads from an unbounded source.
TestSparkRunner - Class in org.apache.beam.runners.spark
The SparkRunner translate operations defined on a pipeline to a representation executable by Spark, and then submitting the job to Spark to be executed.
TestStream<T> - Class in org.apache.beam.sdk.testing
A testing input that generates an unbounded PCollection of elements, advancing the watermark and processing time as elements are emitted.
TestStream.Builder<T> - Class in org.apache.beam.sdk.testing
An incomplete TestStream.
TestStream.ElementEvent<T> - Class in org.apache.beam.sdk.testing
A TestStream.Event that produces elements.
TestStream.Event<T> - Interface in org.apache.beam.sdk.testing
An event in a TestStream.
TestStream.EventType - Enum in org.apache.beam.sdk.testing
The types of TestStream.Event that are supported by TestStream.
TestStream.ProcessingTimeEvent<T> - Class in org.apache.beam.sdk.testing
A TestStream.Event that advances the processing time clock.
TestStream.TestStreamCoder<T> - Class in org.apache.beam.sdk.testing
Coder for TestStream.
TestStream.WatermarkEvent<T> - Class in org.apache.beam.sdk.testing
A TestStream.Event that advances the watermark.
TestStreams - Class in org.apache.beam.sdk.fn.test
Utility methods which enable testing of StreamObservers.
TestStreams() - Constructor for class org.apache.beam.sdk.fn.test.TestStreams
 
TestStreams.Builder<T> - Class in org.apache.beam.sdk.fn.test
A builder for a test CallStreamObserver that performs various callbacks.
TestTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.test
Base class for mocked table.
TestTable(Schema) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTable
 
TestTableFilter - Class in org.apache.beam.sdk.extensions.sql.meta.provider.test
 
TestTableFilter(List<RexNode>) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableFilter
 
TestTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.test
Test in-memory table provider for use in tests.
TestTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider
 
TestTableProvider.PushDownOptions - Enum in org.apache.beam.sdk.extensions.sql.meta.provider.test
 
TestTableProvider.TableWithRows - Class in org.apache.beam.sdk.extensions.sql.meta.provider.test
TableWitRows.
TestTableUtils - Class in org.apache.beam.sdk.extensions.sql.meta.provider.test
Utility functions for mock classes.
TestTableUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableUtils
 
TestUnboundedTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.test
A mocked unbounded table.
TextIO - Class in org.apache.beam.sdk.io
PTransforms for reading and writing text files.
TextIO.CompressionType - Enum in org.apache.beam.sdk.io
Deprecated.
TextIO.Read - Class in org.apache.beam.sdk.io
Implementation of TextIO.read().
TextIO.ReadAll - Class in org.apache.beam.sdk.io
Deprecated.
See TextIO.readAll() for details.
TextIO.ReadFiles - Class in org.apache.beam.sdk.io
Implementation of TextIO.readFiles().
TextIO.Sink - Class in org.apache.beam.sdk.io
Implementation of TextIO.sink().
TextIO.TypedWrite<UserT,DestinationT> - Class in org.apache.beam.sdk.io
Implementation of TextIO.write().
TextIO.Write - Class in org.apache.beam.sdk.io
This class is used as the default return value of TextIO.write().
TextJsonTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.text
TextJsonTable is a BeamSqlTable that reads text files and converts them according to the JSON format.
TextJsonTable(Schema, String, TextTableProvider.JsonToRow, TextTableProvider.RowToJson) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextJsonTable
 
TextRowCountEstimator - Class in org.apache.beam.sdk.io
This returns a row count estimation for files associated with a file pattern.
TextRowCountEstimator() - Constructor for class org.apache.beam.sdk.io.TextRowCountEstimator
 
TextRowCountEstimator.Builder - Class in org.apache.beam.sdk.io
TextRowCountEstimator.LimitNumberOfFiles - Class in org.apache.beam.sdk.io
This strategy stops sampling if we sample enough number of bytes.
TextRowCountEstimator.LimitNumberOfTotalBytes - Class in org.apache.beam.sdk.io
This strategy stops sampling when total number of sampled bytes are more than some threshold.
TextRowCountEstimator.NoEstimationException - Exception in org.apache.beam.sdk.io
An exception that will be thrown if the estimator cannot get an estimation of the number of lines.
TextRowCountEstimator.SampleAllFiles - Class in org.apache.beam.sdk.io
This strategy samples all the files.
TextRowCountEstimator.SamplingStrategy - Interface in org.apache.beam.sdk.io
Sampling Strategy shows us when should we stop reading further files.
TextTable - Class in org.apache.beam.sdk.extensions.sql.meta.provider.text
TextTable is a BeamSqlTable that reads text files and converts them according to the specified format.
TextTable(Schema, String, PTransform<PCollection<String>, PCollection<Row>>, PTransform<PCollection<Row>, PCollection<String>>) - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTable
Text table with the specified read and write transforms.
TextTableProvider - Class in org.apache.beam.sdk.extensions.sql.meta.provider.text
Text table provider.
TextTableProvider() - Constructor for class org.apache.beam.sdk.extensions.sql.meta.provider.text.TextTableProvider
 
TextTableProvider.CsvToRow - Class in org.apache.beam.sdk.extensions.sql.meta.provider.text
Read-side converter for TextTable with format 'csv'.
TextTableProvider.LinesReadConverter - Class in org.apache.beam.sdk.extensions.sql.meta.provider.text
Read-side converter for TextTable with format 'lines'.
TextTableProvider.LinesWriteConverter - Class in org.apache.beam.sdk.extensions.sql.meta.provider.text
Write-side converter for for TextTable with format 'lines'.
TextualIntegerCoder - Class in org.apache.beam.sdk.coders
A Coder that encodes Integer Integers as the ASCII bytes of their textual, decimal, representation.
TextualIntegerCoder() - Constructor for class org.apache.beam.sdk.coders.TextualIntegerCoder
 
TFRecordIO - Class in org.apache.beam.sdk.io
PTransforms for reading and writing TensorFlow TFRecord files.
TFRecordIO.CompressionType - Enum in org.apache.beam.sdk.io
Deprecated.
TFRecordIO.Read - Class in org.apache.beam.sdk.io
Implementation of TFRecordIO.read().
TFRecordIO.ReadFiles - Class in org.apache.beam.sdk.io
Implementation of TFRecordIO.readFiles().
TFRecordIO.Sink - Class in org.apache.beam.sdk.io
TFRecordIO.Write - Class in org.apache.beam.sdk.io
Implementation of TFRecordIO.write().
that(PCollection<T>) - Static method in class org.apache.beam.sdk.testing.PAssert
Constructs an PAssert.IterableAssert for the elements of the provided PCollection.
that(String, PCollection<T>) - Static method in class org.apache.beam.sdk.testing.PAssert
Constructs an PAssert.IterableAssert for the elements of the provided PCollection with the specified reason.
thatMap(PCollection<KV<K, V>>) - Static method in class org.apache.beam.sdk.testing.PAssert
Constructs a PAssert.SingletonAssert for the value of the provided PCollection, which must have at most one value per key.
thatMap(String, PCollection<KV<K, V>>) - Static method in class org.apache.beam.sdk.testing.PAssert
Constructs a PAssert.SingletonAssert for the value of the provided PCollection with the specified reason.
thatMultimap(PCollection<KV<K, V>>) - Static method in class org.apache.beam.sdk.testing.PAssert
Constructs a PAssert.SingletonAssert for the value of the provided PCollection.
thatMultimap(String, PCollection<KV<K, V>>) - Static method in class org.apache.beam.sdk.testing.PAssert
Constructs a PAssert.SingletonAssert for the value of the provided PCollection with the specified reason.
thatSingleton(PCollection<T>) - Static method in class org.apache.beam.sdk.testing.PAssert
Constructs a PAssert.SingletonAssert for the value of the provided PCollection PCollection<T>, which must be a singleton.
thatSingleton(String, PCollection<T>) - Static method in class org.apache.beam.sdk.testing.PAssert
Constructs a PAssert.SingletonAssert for the value of the provided PCollection PCollection<T> with the specified reason.
thatSingletonIterable(PCollection<? extends Iterable<T>>) - Static method in class org.apache.beam.sdk.testing.PAssert
Constructs an PAssert.IterableAssert for the value of the provided PCollection which must contain a single Iterable<T> value.
thatSingletonIterable(String, PCollection<? extends Iterable<T>>) - Static method in class org.apache.beam.sdk.testing.PAssert
Constructs an PAssert.IterableAssert for the value of the provided PCollection with the specified reason.
threadSafe(WatermarkEstimator<WatermarkEstimatorStateT>) - Static method in class org.apache.beam.sdk.fn.splittabledofn.WatermarkEstimators
Returns a thread safe WatermarkEstimator which allows getting a snapshot of the current watermark and watermark estimator state.
ThriftIO - Class in org.apache.beam.sdk.io.thrift
PTransforms for reading and writing files containing Thrift encoded data.
ThriftIO.ReadFiles<T> - Class in org.apache.beam.sdk.io.thrift
ThriftIO.Sink<T extends org.apache.thrift.TBase<?,?>> - Class in org.apache.beam.sdk.io.thrift
ThriftIO.ThriftWriter<T extends org.apache.thrift.TBase<?,?>> - Class in org.apache.beam.sdk.io.thrift
Writer to write Thrift object to OutputStream.
ThrowingBiConsumer<T1,T2> - Interface in org.apache.beam.sdk.function
A BiConsumer which can throw Exceptions.
ThrowingBiFunction<T1,T2,T3> - Interface in org.apache.beam.sdk.function
A BiFunction which can throw Exceptions.
ThrowingConsumer<ExceptionT extends java.lang.Exception,T> - Interface in org.apache.beam.sdk.function
A Consumer which can throw Exceptions.
ThrowingFunction<T1,T2> - Interface in org.apache.beam.sdk.function
A Function which can throw Exceptions.
ThrowingRunnable - Interface in org.apache.beam.sdk.function
A Runnable which can throw Exceptions.
throwNullCredentialException() - Static method in class org.apache.beam.sdk.extensions.gcp.auth.NullCredentialInitializer
 
TikaIO - Class in org.apache.beam.sdk.io.tika
Transforms for parsing arbitrary files using Apache Tika.
TikaIO() - Constructor for class org.apache.beam.sdk.io.tika.TikaIO
 
TikaIO.Parse - Class in org.apache.beam.sdk.io.tika
Implementation of TikaIO.parse().
TikaIO.ParseFiles - Class in org.apache.beam.sdk.io.tika
Implementation of TikaIO.parseFiles().
TIME - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
TIME_WITH_LOCAL_TZ - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
TimeDomain - Enum in org.apache.beam.sdk.state
TimeDomain specifies whether an operation is based on timestamps of elements or current "real-world" time as reported while processing.
timeDomain() - Method in class org.apache.beam.sdk.transforms.DoFn.OnTimerContext
Returns the time domain of the current timer.
timer(String, String, String) - Static method in class org.apache.beam.sdk.fn.data.LogicalEndpoint
 
Timer - Interface in org.apache.beam.sdk.state
A timer for a specified time domain that can be set to register the desire for further processing at particular time in its specified time domain.
timer(TimeDomain) - Static method in class org.apache.beam.sdk.state.TimerSpecs
 
timerId() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.TimerSpec
 
timerInternals() - Method in class org.apache.beam.runners.gearpump.translators.utils.NoOpStepContext
 
timerInternals() - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.batch.functions.NoOpStepContext
 
TimerMap - Interface in org.apache.beam.sdk.state
 
timerMap(TimeDomain) - Static method in class org.apache.beam.sdk.state.TimerSpecs
 
TimerReceiverFactory - Class in org.apache.beam.runners.fnexecution.control
A factory that passes timers to TimerReceiverFactory.timerDataConsumer.
TimerReceiverFactory(StageBundleFactory, BiConsumer<Timer<?>, TimerInternals.TimerData>, Coder) - Constructor for class org.apache.beam.runners.fnexecution.control.TimerReceiverFactory
 
Timers - Interface in org.apache.beam.sdk.state
Interface for interacting with time.
TimerSpec() - Constructor for class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.TimerSpec
 
TimerSpec - Interface in org.apache.beam.sdk.state
A specification for a Timer.
TimerSpecs - Class in org.apache.beam.sdk.state
Static methods for working with TimerSpecs.
TimerSpecs() - Constructor for class org.apache.beam.sdk.state.TimerSpecs
 
TimerUtils - Class in org.apache.beam.runners.spark.util
 
TimerUtils() - Constructor for class org.apache.beam.runners.spark.util.TimerUtils
 
TIMESTAMP - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
TIMESTAMP - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
timestamp(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.TimestampFunctions
 
timestamp(String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.TimestampFunctions
 
timestamp(Integer) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.TimestampFunctions
 
timestamp(Integer, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.TimestampFunctions
 
timestamp() - Method in class org.apache.beam.sdk.transforms.DoFn.OnTimerContext
Returns the timestamp of the current timer.
timestamp() - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContext
Returns the timestamp of the input element.
timestamp() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn.AssignContext
Returns the timestamp of the current element.
TIMESTAMP_ADD_FN - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
TIMESTAMP_MAX_VALUE - Static variable in class org.apache.beam.sdk.transforms.windowing.BoundedWindow
The maximum value for any Beam timestamp.
TIMESTAMP_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamBuiltinMethods
 
TIMESTAMP_MIN_VALUE - Static variable in class org.apache.beam.sdk.transforms.windowing.BoundedWindow
The minimum value for any Beam timestamp.
TIMESTAMP_OP - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
TIMESTAMP_WITH_LOCAL_TZ - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
TIMESTAMP_WITH_NULLABILITY - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
timestampColumnIndex(int) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestUnboundedTable
 
TimestampCombiner - Enum in org.apache.beam.sdk.transforms.windowing
Policies for combining timestamps that occur within a window.
TimeStampComparator() - Constructor for class org.apache.beam.runners.dataflow.util.MonitoringUtil.TimeStampComparator
 
TimestampConvert() - Constructor for class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.TimestampConvert
 
timestamped(Iterable<TimestampedValue<T>>) - Static method in class org.apache.beam.sdk.transforms.Create
Returns a new Create.TimestampedValues transform that produces a PCollection containing the elements of the provided Iterable with the specified timestamps.
timestamped(TimestampedValue<T>, TimestampedValue<T>...) - Static method in class org.apache.beam.sdk.transforms.Create
Returns a new Create.TimestampedValues transform that produces a PCollection containing the specified elements with the specified timestamps.
timestamped(Iterable<T>, Iterable<Long>) - Static method in class org.apache.beam.sdk.transforms.Create
Returns a new root transform that produces a PCollection containing the specified elements with the specified timestamps.
TimestampedValue<V> - Class in org.apache.beam.sdk.values
An immutable pair of a value and a timestamp.
TimestampedValue(V, Instant) - Constructor for class org.apache.beam.sdk.values.TimestampedValue
 
TimestampedValue.TimestampedValueCoder<T> - Class in org.apache.beam.sdk.values
TimestampFunctions - Class in org.apache.beam.sdk.extensions.sql.zetasql
TimestampFunctions.
TimestampFunctions() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.TimestampFunctions
 
timestampMsSinceEpoch() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.OutgoingMessage
Timestamp for element (ms since epoch).
TimestampObservingWatermarkEstimator<WatermarkEstimatorStateT> - Interface in org.apache.beam.sdk.transforms.splittabledofn
A WatermarkEstimator that observes the timestamps of all records output from a DoFn.
TimestampPolicy<K,V> - Class in org.apache.beam.sdk.io.kafka
A timestamp policy to assign event time for messages in a Kafka partition and watermark for it.
TimestampPolicy() - Constructor for class org.apache.beam.sdk.io.kafka.TimestampPolicy
 
TimestampPolicy.PartitionContext - Class in org.apache.beam.sdk.io.kafka
The context contains state maintained in the reader for the partition.
TimestampPolicyFactory<KeyT,ValueT> - Interface in org.apache.beam.sdk.io.kafka
An extendable factory to create a TimestampPolicy for each partition at runtime by KafkaIO reader.
TimestampPolicyFactory.LogAppendTimePolicy<K,V> - Class in org.apache.beam.sdk.io.kafka
Assigns Kafka's log append time (server side ingestion time) to each record.
TimestampPolicyFactory.ProcessingTimePolicy<K,V> - Class in org.apache.beam.sdk.io.kafka
A simple policy that uses current time for event time and watermark.
TimestampPolicyFactory.TimestampFnPolicy<K,V> - Class in org.apache.beam.sdk.io.kafka
Internal policy to support deprecated withTimestampFn API.
timestamps() - Static method in class org.apache.beam.sdk.transforms.Reify
Create a PTransform that will output all inputs wrapped in a TimestampedValue.
timestampsInValue() - Static method in class org.apache.beam.sdk.transforms.Reify
Create a PTransform that will output all input KVs with the timestamp inside the value.
TimestampTransform - Class in org.apache.beam.sdk.transforms.windowing
For internal use only; no backwards-compatibility guarantees.
TimestampTransform.AlignTo - Class in org.apache.beam.sdk.transforms.windowing
For internal use only; no backwards-compatibility guarantees.
TimestampTransform.Delay - Class in org.apache.beam.sdk.transforms.windowing
For internal use only; no backwards-compatibility guarantees.
TimestampWithLocalTzType() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.TimestampWithLocalTzType
 
TimeType() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.TimeType
 
TimeUtil - Class in org.apache.beam.runners.dataflow.util
A helper class for converting between Dataflow API and SDK time representations.
TimeWithLocalTzType() - Constructor for class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils.TimeWithLocalTzType
 
TINY_INT - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
TmpCheckpointDirFactory() - Constructor for class org.apache.beam.runners.spark.SparkCommonPipelineOptions.TmpCheckpointDirFactory
 
to(String) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Writes to file(s) with the given output prefix.
to(ResourceId) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Writes to file(s) with the given output prefix.
to(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
to(FileBasedSink.FilenamePolicy) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Writes to files named according to the given FileBasedSink.FilenamePolicy.
to(DynamicAvroDestinations<UserT, NewDestinationT, OutputT>) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Deprecated.
to(String) - Method in class org.apache.beam.sdk.io.AvroIO.Write
to(ResourceId) - Method in class org.apache.beam.sdk.io.AvroIO.Write
to(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.AvroIO.Write
to(FileBasedSink.FilenamePolicy) - Method in class org.apache.beam.sdk.io.AvroIO.Write
See TypedWrite#to(FilenamePolicy).
to(DynamicAvroDestinations<T, ?, T>) - Method in class org.apache.beam.sdk.io.AvroIO.Write
Deprecated.
to(String) - Method in class org.apache.beam.sdk.io.FileIO.Write
Specifies a common directory for all generated files.
to(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.FileIO.Write
to(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Writes to the given table, specified in the format described in BigQueryHelpers.parseTableSpec(java.lang.String).
to(TableReference) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Writes to the given table, specified as a TableReference.
to(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
to(SerializableFunction<ValueInSingleWindow<T>, TableDestination>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Writes to table specified by the specified table function.
to(DynamicDestinations<T, ?>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Writes to the table and schema specified by the DynamicDestinations object.
to(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
Publishes to the specified topic.
to(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
Like topic() but with a ValueProvider.
to(long) - Method in class org.apache.beam.sdk.io.GenerateSequence
Specifies the maximum number to generate (exclusive).
to(String) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Write
Provide name of collection while reading from Solr.
to(String) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Writes to text files with the given prefix.
to(ResourceId) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
to(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
to(FileBasedSink.FilenamePolicy) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Writes to files named according to the given FileBasedSink.FilenamePolicy.
to(FileBasedSink.DynamicDestinations<UserT, NewDestinationT, String>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Deprecated.
to(SerializableFunction<UserT, DefaultFilenamePolicy.Params>, DefaultFilenamePolicy.Params) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Deprecated.
to(String) - Method in class org.apache.beam.sdk.io.TextIO.Write
to(ResourceId) - Method in class org.apache.beam.sdk.io.TextIO.Write
to(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.TextIO.Write
to(FileBasedSink.FilenamePolicy) - Method in class org.apache.beam.sdk.io.TextIO.Write
See TypedWrite#to(FilenamePolicy).
to(FileBasedSink.DynamicDestinations<String, ?, String>) - Method in class org.apache.beam.sdk.io.TextIO.Write
Deprecated.
to(SerializableFunction<String, DefaultFilenamePolicy.Params>, DefaultFilenamePolicy.Params) - Method in class org.apache.beam.sdk.io.TextIO.Write
Deprecated.
to(String) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
Writes TFRecord file(s) with the given output prefix.
to(ResourceId) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
Writes TFRecord file(s) with a prefix given by the specified resource.
to(FileBasedSink<UserT, DestinationT, OutputT>) - Static method in class org.apache.beam.sdk.io.WriteFiles
Creates a WriteFiles transform that writes to the given FileBasedSink, letting the runner control how many different shards are produced.
to(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
Writes to files with the given path prefix.
to(Class<OutputT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Convert
Convert a PCollection<InputT> to a PCollection<OutputT>.
to(TypeDescriptor<OutputT>) - Static method in class org.apache.beam.sdk.schemas.transforms.Convert
Convert a PCollection<InputT> to a PCollection<OutputT>.
toAbsolutePath() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
toAdditionalInputs(Iterable<PCollectionView<?>>) - Static method in class org.apache.beam.sdk.values.PCollectionViews
Expands a list of PCollectionView into the form needed for PTransform.getAdditionalInputs().
toArrayRelDataType(RexBuilder, ArrayType, boolean) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.TypeUtils
 
toAvroField(Schema.Field, String) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Get Avro Field from Beam Field.
toAvroSchema(Schema, String, String) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Converts a Beam Schema into an AVRO schema.
toAvroSchema(Schema) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
 
toAvroType(String, String) - Method in class org.apache.beam.sdk.schemas.utils.AvroUtils.FixedBytesField
Convert to an AVRO type.
toBaseType(EnumerationType.Value) - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
 
toBaseType(byte[]) - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
 
toBaseType(Duration) - Method in class org.apache.beam.sdk.schemas.logicaltypes.NanosDuration
 
toBaseType(Instant) - Method in class org.apache.beam.sdk.schemas.logicaltypes.NanosInstant
 
toBaseType(OneOfType.Value) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
 
toBaseType(T) - Method in class org.apache.beam.sdk.schemas.logicaltypes.PassThroughLogicalType
 
toBaseType(InputT) - Method in interface org.apache.beam.sdk.schemas.Schema.LogicalType
 
toBeamField(Schema.Field) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Get Beam Field from avro Field.
toBeamRow(GenericRecord, Schema, BigQueryUtils.ConversionOptions) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
 
toBeamRow(Schema, TableRow) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
Tries to convert a JSON TableRow from BigQuery into a Beam Row.
toBeamRow(Schema, TableSchema, TableRow) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
Tries to parse the JSON TableRow from BigQuery.
toBeamRowStrict(GenericRecord, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Strict conversion from AVRO to Beam, strict because it doesn't do widening or narrowing during conversion.
toBeamSchema(Schema) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Converts AVRO schema to Beam row schema.
toBeamType() - Method in class org.apache.beam.sdk.schemas.utils.AvroUtils.FixedBytesField
Convert to a Beam type.
toBuilder() - Method in class org.apache.beam.sdk.extensions.sql.meta.Table
 
toBuilder() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
 
toBuilder() - Method in class org.apache.beam.sdk.io.kinesis.WatermarkParameters
 
toBuilder() - Method in class org.apache.beam.sdk.schemas.Schema.Field
 
toBuilder() - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
 
toBuilder() - Method in class org.apache.beam.sdk.transforms.DoFnSchemaInformation
 
toByteArray(T, Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
Utility method for serializing an object using the specified coder.
toByteArray(T, Coder<T>) - Static method in class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.CoderHelpers
Utility method for serializing an object using the specified coder.
toByteArrays(Iterable<T>, Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
Utility method for serializing a Iterable of values using the specified coder.
toByteArrayWithTs(T, Coder<T>, Instant) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
Utility method for serializing an object using the specified coder, appending timestamp representation.
toByteFunction(Coder<T>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
A function wrapper for converting an object to a bytearray.
toByteFunction(Coder<K>, Coder<V>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
A function wrapper for converting a key-value pair to a byte array pair.
toByteFunctionWithTs(Coder<K>, Coder<V>, Function<Tuple2<K, V>, Instant>) - Static method in class org.apache.beam.runners.spark.coders.CoderHelpers
A function wrapper for converting a key-value pair to a byte array pair, where the key in resulting ByteArray contains (key, timestamp).
toCalciteRowType(Schema, RelDataTypeFactory) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
Create an instance of RelDataType so it can be used to create a table.
toCloudDuration(ReadableDuration) - Static method in class org.apache.beam.runners.dataflow.util.TimeUtil
Converts a ReadableDuration into a Dataflow API duration string.
toCloudObject(T, SdkComponents) - Method in interface org.apache.beam.runners.dataflow.util.CloudObjectTranslator
Converts the provided object into an equivalent CloudObject.
toCloudObject(SchemaCoder, SdkComponents) - Method in class org.apache.beam.runners.dataflow.util.SchemaCoderCloudObjectTranslator
Convert to a cloud object.
toCloudTime(ReadableInstant) - Static method in class org.apache.beam.runners.dataflow.util.TimeUtil
Converts a ReadableInstant into a Dateflow API time value.
toDefaultPolicies(SerializableFunction<UserT, DefaultFilenamePolicy.Params>, DefaultFilenamePolicy.Params, SerializableFunction<UserT, OutputT>) - Static method in class org.apache.beam.sdk.io.DynamicFileDestinations
Returns a FileBasedSink.DynamicDestinations that returns instances of DefaultFilenamePolicy configured with the given DefaultFilenamePolicy.Params.
toDuration(Row) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.DurationConvert
ByteBuddy conversion for NanosDuration base type to Duration.
toEnumerable(BeamRelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter
 
toField(RelDataTypeField) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
toField(String, RelDataType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
toFieldType(SqlTypeName) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
toFieldType(RelDataType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
toFile() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
toGcpBackOff(BackOff) - Static method in class org.apache.beam.sdk.extensions.gcp.util.BackOffAdapter
Returns an adapter to convert from BackOff to BackOff.
toGenericAvroSchema(String, List<TableFieldSchema>) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
Convert a list of BigQuery TableFieldSchema to Avro Schema.
toGenericRecord(Row, Schema) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils
Convert from a Beam Row to an AVRO GenericRecord.
toHex(byte[]) - Method in class org.apache.beam.sdk.extensions.sql.impl.udf.BuiltinStringFunctions
 
toInputType(Integer) - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
 
toInputType(byte[]) - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
 
toInputType(Row) - Method in class org.apache.beam.sdk.schemas.logicaltypes.NanosDuration
 
toInputType(Row) - Method in class org.apache.beam.sdk.schemas.logicaltypes.NanosInstant
 
toInputType(Row) - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
 
toInputType(T) - Method in class org.apache.beam.sdk.schemas.logicaltypes.PassThroughLogicalType
 
toInputType(BaseT) - Method in interface org.apache.beam.sdk.schemas.Schema.LogicalType
Convert the Java type used by the base Schema.FieldType to the input type.
ToJson<T> - Class in org.apache.beam.sdk.transforms
Experimental
toJsonString(Object) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
 
tokenNames - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
Deprecated.
tokenNames - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
Deprecated.
toList(JavaStream<TranslatorUtils.RawUnionValue>) - Static method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils
 
toLong(DateTime, Schema, LogicalType) - Method in class org.apache.beam.sdk.coders.AvroCoder.JodaTimestampConversion
 
toModificationRel(RelOptCluster, RelOptTable, Prepare.CatalogReader, RelNode, TableModify.Operation, List<String>, List<RexNode>, boolean) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteTable
 
toNullableRecordField(Object[], int) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableUtils
 
Top - Class in org.apache.beam.sdk.transforms
PTransforms for finding the largest (or smallest) set of elements in a PCollection, or the largest (or smallest) set of values associated with each key in a PCollection of KVs.
Top.Largest<T extends java.lang.Comparable<? super T>> - Class in org.apache.beam.sdk.transforms
Deprecated.
use Top.Natural instead
Top.Natural<T extends java.lang.Comparable<? super T>> - Class in org.apache.beam.sdk.transforms
A Serializable Comparator that that uses the compared elements' natural ordering.
Top.Reversed<T extends java.lang.Comparable<? super T>> - Class in org.apache.beam.sdk.transforms
Serializable Comparator that that uses the reverse of the compared elements' natural ordering.
Top.Smallest<T extends java.lang.Comparable<? super T>> - Class in org.apache.beam.sdk.transforms
Deprecated.
use Top.Reversed instead
Top.TopCombineFn<T,ComparatorT extends java.util.Comparator<T> & java.io.Serializable> - Class in org.apache.beam.sdk.transforms
CombineFn for Top transforms that combines a bunch of Ts into a single count-long List<T>, using compareFn to choose the largest Ts.
toPCollection(Pipeline, BeamRelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamSqlRelUtils
 
TopCombineFn(int, ComparatorT) - Constructor for class org.apache.beam.sdk.transforms.Top.TopCombineFn
 
topicPath() - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
Topic path where events will be published to.
topicPathFromName(String, String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
 
topicPathFromPath(String) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient
 
toProto() - Method in class org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation
Convert to JobApi.JobInfo.
toProvisionInfo() - Method in class org.apache.beam.runners.fnexecution.provisioning.JobInfo
 
toPTransform() - Method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortRead
 
toPTransform() - Method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortWrite
 
toRealPath(LinkOption...) - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
toRecordField(Object[], int) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableUtils
 
toRel(RelOptTable.ToRelContext, RelOptTable) - Method in class org.apache.beam.sdk.extensions.sql.impl.BeamCalciteTable
 
toRelDataType(RelDataTypeFactory, Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
toRelDataType(RexBuilder, Type, boolean) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.TypeUtils
 
toResource(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
toResource(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.AvroIO.Write
toResource(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
toResource(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.TextIO.Write
toResource(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
toResourceName() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
toRow(Duration) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.DurationConvert
ByteBuddy conversion for Duration to NanosDuration base type.
toRow(Timestamp) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.TimestampConvert
ByteBuddy conversion for Timestamp to NanosInstant base type.
toRow(Schema) - Static method in class org.apache.beam.sdk.values.Row
Creates a Row from the list of values and Row.getSchema().
toRowFunction(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.annotations.DefaultSchema.DefaultSchemaProvider
Given a type, return a function that converts that type to a Row object If no schema exists, returns null.
toRowFunction(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.schemas.GetterBasedSchemaProvider
 
toRowFunction(TypeDescriptor<T>) - Method in interface org.apache.beam.sdk.schemas.SchemaProvider
Given a type, return a function that converts that type to a Row object If no schema exists, returns null.
toRowList(BeamRelNode) - Static method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamEnumerableConverter
 
toRows() - Static method in class org.apache.beam.sdk.schemas.transforms.Convert
Convert a PCollection<InputT> into a PCollection<Row>.
toSchema(RelDataType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
Generate Schema from RelDataType which is used to create table.
toSchema() - Static method in class org.apache.beam.sdk.schemas.Schema
Collects a stream of Schema.Fields into a Schema.
toSimpleRelDataType(ZetaSQLType.TypeKind, RexBuilder) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.TypeUtils
 
toSimpleRelDataType(ZetaSQLType.TypeKind, RexBuilder, boolean) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.TypeUtils
 
toSql(RexProgram, RexNode) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamSqlUnparseContext
 
toSqlTypeName(Schema.FieldType) - Static method in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
toState(String) - Static method in class org.apache.beam.runners.dataflow.util.MonitoringUtil
 
toString() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator
 
toString() - Method in class org.apache.beam.runners.dataflow.DataflowRunner
 
toString() - Method in class org.apache.beam.runners.dataflow.TestDataflowRunner
 
toString() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
 
toString() - Method in class org.apache.beam.runners.direct.WatermarkManager.FiredTimers
 
toString() - Method in class org.apache.beam.runners.direct.WatermarkManager.TimerUpdate
 
toString() - Method in class org.apache.beam.runners.direct.WatermarkManager.TransformWatermarks
 
toString() - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
 
toString() - Method in class org.apache.beam.runners.flink.FlinkRunner
 
toString() - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
 
toString(Metric) - Static method in class org.apache.beam.runners.flink.metrics.Metrics
 
toString() - Method in class org.apache.beam.runners.fnexecution.artifact.BeamFileSystemLegacyArtifactStagingService.StagingSessionToken
 
toString() - Method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils.RawUnionValue
 
toString() - Method in class org.apache.beam.runners.portability.PortableRunner
 
toString() - Method in class org.apache.beam.runners.spark.aggregators.NamedAggregators
 
toString() - Method in class org.apache.beam.runners.spark.stateful.SparkTimerInternals
 
toString() - Method in class org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregators
 
toString() - Method in class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.SparkWatermarks
 
toString() - Method in class org.apache.beam.sdk.coders.Coder.Context
Deprecated.
 
toString() - Method in class org.apache.beam.sdk.coders.DelegateCoder
 
toString() - Method in class org.apache.beam.sdk.coders.SerializableCoder
 
toString() - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
 
toString() - Method in class org.apache.beam.sdk.coders.StructuralByteArray
 
toString() - Method in class org.apache.beam.sdk.coders.StructuredCoder
 
toString() - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsResourceId
 
toString() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
toString() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Customer
 
toString() - Method in class org.apache.beam.sdk.extensions.sql.example.model.Order
 
toString() - Method in class org.apache.beam.sdk.extensions.sql.impl.planner.BeamCostModel
 
toString() - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
 
toString() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BigQueryFilter
 
toString() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableFilter
 
toString() - Method in class org.apache.beam.sdk.fn.data.BeamFnDataGrpcMultiplexer
 
toString() - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
 
toString() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
 
toString() - Method in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
 
toString() - Method in class org.apache.beam.sdk.io.FileBasedSource
 
toString() - Method in class org.apache.beam.sdk.io.FileIO.ReadableFile
 
toString() - Method in interface org.apache.beam.sdk.io.fs.ResourceId
Returns the string representation of this ResourceId.
toString() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryStorageStreamSource
 
toString() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
toString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
 
toString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
 
toString() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.WriteWithResults
 
toString() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
 
toString() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.ProjectPath
 
toString() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.SubscriptionPath
 
toString() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubClient.TopicPath
 
toString() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription
 
toString() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic
 
toString() - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubMessage
 
toString() - Method in class org.apache.beam.sdk.io.gcp.spanner.MutationGroup
 
toString() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark.PartitionMark
 
toString() - Method in class org.apache.beam.sdk.io.kafka.KafkaCheckpointMark
 
toString() - Method in enum org.apache.beam.sdk.io.kafka.KafkaTimestampType
 
toString() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
 
toString() - Method in class org.apache.beam.sdk.io.range.ByteKey
 
toString() - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
 
toString() - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
 
toString() - Method in class org.apache.beam.sdk.io.range.OffsetRange
 
toString() - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
 
toString() - Method in class org.apache.beam.sdk.io.tika.ParseResult
 
toString() - Method in class org.apache.beam.sdk.metrics.MetricKey
 
toString() - Method in class org.apache.beam.sdk.metrics.MetricName
 
toString() - Method in class org.apache.beam.sdk.metrics.MetricQueryResults
 
toString() - Method in class org.apache.beam.sdk.metrics.MetricResults
 
toString() - Method in class org.apache.beam.sdk.options.ValueProvider.NestedValueProvider
 
toString() - Method in class org.apache.beam.sdk.options.ValueProvider.RuntimeValueProvider
 
toString() - Method in class org.apache.beam.sdk.options.ValueProvider.StaticValueProvider
 
toString() - Method in class org.apache.beam.sdk.Pipeline
 
toString() - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
 
toString(EnumerationType.Value) - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
 
toString() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
 
toString() - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType.Value
 
toString() - Method in class org.apache.beam.sdk.schemas.logicaltypes.FixedBytes
 
toString() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType
 
toString() - Method in class org.apache.beam.sdk.schemas.logicaltypes.OneOfType.Value
 
toString() - Method in class org.apache.beam.sdk.schemas.Schema.Options
 
toString() - Method in class org.apache.beam.sdk.schemas.Schema
 
toString() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
 
toString() - Method in class org.apache.beam.sdk.schemas.transforms.Cast.Narrowing
 
toString() - Method in class org.apache.beam.sdk.schemas.transforms.Cast.Widening
 
toString() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
toString() - Method in class org.apache.beam.sdk.testing.SuccessOrFailure
 
toString() - Method in class org.apache.beam.sdk.testing.TestPipeline
 
toString() - Method in class org.apache.beam.sdk.transforms.Combine.Holder
 
toString() - Method in class org.apache.beam.sdk.transforms.Contextful
 
toString() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Identifier
 
toString() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Item
 
toString() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
 
toString() - Method in class org.apache.beam.sdk.transforms.display.DisplayData.Path
 
toString() - Method in class org.apache.beam.sdk.transforms.display.DisplayData
 
toString() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult
 
toString() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResultSchema
 
toString() - Method in class org.apache.beam.sdk.transforms.join.RawUnionValue
 
toString() - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
 
toString() - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
 
toString() - Method in class org.apache.beam.sdk.transforms.PTransform
 
toString() - Method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
 
toString() - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
 
ToString - Class in org.apache.beam.sdk.transforms
toString() - Method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
 
toString(StateT) - Method in interface org.apache.beam.sdk.transforms.Watch.Growth.TerminationCondition
Creates a human-readable representation of the given state of this condition.
toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterAll
 
toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterEach
 
toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterFirst
 
toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterPane
 
toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterProcessingTime
 
toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterSynchronizedProcessingTime
 
toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
 
toString() - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
 
toString() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
 
toString() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow
 
toString() - Method in class org.apache.beam.sdk.transforms.windowing.OrFinallyTrigger
 
toString() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo
 
toString() - Method in class org.apache.beam.sdk.transforms.windowing.Repeatedly
 
toString() - Method in class org.apache.beam.sdk.transforms.windowing.ReshuffleTrigger
Deprecated.
 
toString() - Method in class org.apache.beam.sdk.transforms.windowing.Trigger
 
toString() - Method in class org.apache.beam.sdk.values.KV
 
toString() - Method in class org.apache.beam.sdk.values.PCollectionViews.SimplePCollectionView
 
toString() - Method in class org.apache.beam.sdk.values.PValueBase
 
toString() - Method in class org.apache.beam.sdk.values.Row
 
toString() - Method in class org.apache.beam.sdk.values.ShardedKey
 
toString() - Method in class org.apache.beam.sdk.values.TimestampedValue
 
toString() - Method in class org.apache.beam.sdk.values.TupleTag
 
toString() - Method in class org.apache.beam.sdk.values.TupleTagList
 
toString() - Method in class org.apache.beam.sdk.values.TypeDescriptor
 
toString() - Method in class org.apache.beam.sdk.values.TypeParameter
 
toString() - Method in class org.apache.beam.sdk.values.ValueWithRecordId
 
toString() - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
toStringTimestamp(long) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.BeamCodegenUtils
 
toStringUTF8(byte[]) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.BeamCodegenUtils
 
toStructRelDataType(RexBuilder, StructType, boolean) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.TypeUtils
 
toTableRow() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
Convert a Beam Row to a BigQuery TableRow.
toTableRow(SerializableFunction<T, Row>) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
Convert a Beam schema type to a BigQuery TableRow.
toTableRow(Row) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
Convert a BigQuery TableRow to a Beam Row.
toTableSchema(Schema) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils
Convert a Beam Schema to a BigQuery TableSchema.
toTableSpec(TableReference) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
Returns a canonical string representation of the TableReference.
toTableSpec(TableReferenceProto.TableReference) - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryHelpers
Returns a canonical string representation of a TableReferenceProto.TableReference.
toTimestamp(Row) - Static method in class org.apache.beam.sdk.extensions.protobuf.ProtoSchemaLogicalTypes.TimestampConvert
ByteBuddy conversion for NanosInstant base type to Timestamp.
toUnsplittableSource(BoundedSource<T>) - Static method in class org.apache.beam.sdk.testing.SourceTestUtils
Returns an equivalent unsplittable BoundedSource<T>.
toUri() - Method in class org.apache.beam.sdk.extensions.gcp.util.gcsfs.GcsPath
 
Transaction - Class in org.apache.beam.sdk.io.gcp.spanner
A transaction object.
Transaction() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.Transaction
 
transactionId() - Method in class org.apache.beam.sdk.io.gcp.spanner.Transaction
 
transfer() - Method in class org.apache.beam.runners.portability.CloseableResource
Returns a new CloseableResource that owns the underlying resource and relinquishes ownership from this CloseableResource.
transform(int, RelTraitSet, RelNode) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl
 
transform(Function<T, V>) - Method in class org.apache.beam.sdk.metrics.MetricResult
 
TRANSFORM_URN - Static variable in class org.apache.beam.runners.spark.io.ConsoleIO.Write.Unbound
 
TRANSFORM_URN - Static variable in class org.apache.beam.runners.spark.io.CreateStream
 
transformContainer(Iterable<FromT>, Function<FromT, DestT>) - Static method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils
 
TransformExecutor - Interface in org.apache.beam.runners.direct
A Runnable that will execute a PTransform on some bundle of input.
transformId() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.BagUserStateSpec
 
transformId() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.SideInputSpec
 
transformId() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.TimerSpec
 
TransformingMap(Map<K1, V1>, Function<K1, K2>, Function<V1, V2>) - Constructor for class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
transformStepNames - Variable in class org.apache.beam.runners.dataflow.DataflowPipelineJob
 
transformTo(RelNode, Map<RelNode, RelNode>) - Method in class org.apache.beam.sdk.extensions.sql.impl.rule.JoinRelOptRuleCall
 
TransformTranslator<TransformT extends PTransform> - Interface in org.apache.beam.runners.dataflow
A TransformTranslator knows how to translate a particular subclass of PTransform for the Cloud Dataflow service.
TransformTranslator<T extends PTransform> - Interface in org.apache.beam.runners.gearpump.translators
Translates PTransform to Gearpump functions.
TransformTranslator<TransformT extends PTransform> - Interface in org.apache.beam.runners.spark.structuredstreaming.translation
Supports translation between a Beam transform, and Spark's operations on Datasets.
TransformTranslator.StepTranslationContext - Interface in org.apache.beam.runners.dataflow
The interface for a TransformTranslator to build a Dataflow step.
TransformTranslator.TranslationContext - Interface in org.apache.beam.runners.dataflow
The interface provided to registered callbacks for interacting with the DataflowRunner, including reading and writing the values of PCollections and side inputs.
translate(Pipeline, ApexPipelineOptions) - Static method in class org.apache.beam.runners.apex.TestApexRunner
 
translate(Pipeline, DataflowRunner, List<DataflowPackage>) - Method in class org.apache.beam.runners.dataflow.DataflowPipelineTranslator
Translates a Pipeline into a JobSpecification.
translate(AppliedPTransform<?, ?, PrimitiveParDoSingleFactory.ParDoSingle<?, ?>>, SdkComponents) - Method in class org.apache.beam.runners.dataflow.PrimitiveParDoSingleFactory.PayloadTranslator
 
translate(TransformT, TransformTranslator.TranslationContext) - Method in interface org.apache.beam.runners.dataflow.TransformTranslator
 
translate(FlinkBatchPortablePipelineTranslator.BatchTranslationContext, RunnerApi.Pipeline) - Method in class org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator
 
translate(T, RunnerApi.Pipeline) - Method in interface org.apache.beam.runners.flink.FlinkPortablePipelineTranslator
Translates the given pipeline.
translate(FlinkStreamingPortablePipelineTranslator.StreamingTranslationContext, RunnerApi.Pipeline) - Method in class org.apache.beam.runners.flink.FlinkStreamingPortablePipelineTranslator
 
translate(CreateStreamingGearpumpView.CreateGearpumpPCollectionView<ElemT, ViewT>, TranslationContext) - Method in class org.apache.beam.runners.gearpump.translators.CreateGearpumpPCollectionViewTranslator
 
translate(Flatten.PCollections<T>, TranslationContext) - Method in class org.apache.beam.runners.gearpump.translators.FlattenPCollectionsTranslator
 
translate(Pipeline) - Method in class org.apache.beam.runners.gearpump.translators.GearpumpPipelineTranslator
 
translate(GroupByKey<K, V>, TranslationContext) - Method in class org.apache.beam.runners.gearpump.translators.GroupByKeyTranslator
 
translate(ParDo.MultiOutput<InputT, OutputT>, TranslationContext) - Method in class org.apache.beam.runners.gearpump.translators.ParDoMultiOutputTranslator
 
translate(Read.Bounded<T>, TranslationContext) - Method in class org.apache.beam.runners.gearpump.translators.ReadBoundedTranslator
 
translate(Read.Unbounded<T>, TranslationContext) - Method in class org.apache.beam.runners.gearpump.translators.ReadUnboundedTranslator
 
translate(T, TranslationContext) - Method in interface org.apache.beam.runners.gearpump.translators.TransformTranslator
 
translate(Window.Assign<T>, TranslationContext) - Method in class org.apache.beam.runners.gearpump.translators.WindowAssignTranslator
 
translate(TransformHierarchy.Node, TransformT) - Method in class org.apache.beam.runners.spark.SparkRunner.Evaluator
Determine if this Node belongs to a Bounded branch of the pipeline, or Unbounded, and translate with the proper translator.
translate(Pipeline) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
Translates the pipeline by passing this class as a visitor.
translateOnly - Variable in class org.apache.beam.runners.apex.ApexRunner
 
translateTransform(TransformT, TranslationContext) - Method in interface org.apache.beam.runners.spark.structuredstreaming.translation.TransformTranslator
Base class for translators of PTransform.
TranslationContext - Class in org.apache.beam.runners.gearpump.translators
Maintains context data for TransformTranslators.
TranslationContext(JavaStreamApp, GearpumpPipelineOptions) - Constructor for class org.apache.beam.runners.gearpump.translators.TranslationContext
 
translationContext - Variable in class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
 
TranslationContext - Class in org.apache.beam.runners.spark.structuredstreaming.translation
Base class that gives a context for PTransform translation: keeping track of the datasets, the SparkSession, the current transform being translated.
TranslationContext(SparkStructuredStreamingPipelineOptions) - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.TranslationContext
 
translator - Variable in class org.apache.beam.runners.spark.SparkRunner.Evaluator
 
TranslatorUtils - Class in org.apache.beam.runners.gearpump.translators.utils
Utility methods for translators.
TranslatorUtils() - Constructor for class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils
 
TranslatorUtils.FromRawUnionValue<OutputT> - Class in org.apache.beam.runners.gearpump.translators.utils
Converts @link{RawUnionValue} to @link{WindowedValue}.
TranslatorUtils.RawUnionValue - Class in org.apache.beam.runners.gearpump.translators.utils
This is copied from org.apache.beam.sdk.transforms.join.RawUnionValue.
Transport - Class in org.apache.beam.sdk.extensions.gcp.util
Helpers for cloud communication.
Transport() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.Transport
 
traverseTopologically(Pipeline.PipelineVisitor) - Method in class org.apache.beam.sdk.Pipeline
For internal use only; no backwards-compatibility guarantees.
Trigger - Class in org.apache.beam.sdk.transforms.windowing
Triggers control when the elements for a specific key and window are output.
Trigger(List<Trigger>) - Constructor for class org.apache.beam.sdk.transforms.windowing.Trigger
 
Trigger() - Constructor for class org.apache.beam.sdk.transforms.windowing.Trigger
 
Trigger.OnceTrigger - Class in org.apache.beam.sdk.transforms.windowing
For internal use only; no backwards-compatibility guarantees.
triggering(Trigger) - Method in class org.apache.beam.sdk.transforms.windowing.Window
Sets a non-default trigger for this Window PTransform.
TRIM - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
trim(String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
trim(String, String) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.StringFunctions
 
TRIM_METHOD - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.BeamBuiltinMethods
 
trivial() - Static method in class org.apache.beam.sdk.fn.stream.OutboundObserverFactory
Creates an OutboundObserverFactory that simply delegates to the base factory, with no flow control or synchronization.
trueLiteral() - Method in class org.apache.beam.sdk.extensions.sql.zetasql.translation.ExpressionConverter
 
tryAcquireJobLock(Configuration) - Method in interface org.apache.beam.sdk.io.hadoop.format.ExternalSynchronization
Tries to acquire lock for given job.
tryAcquireJobLock(Configuration) - Method in class org.apache.beam.sdk.io.hadoop.format.HDFSSynchronization
 
tryClaim(ByteKey) - Method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
Attempts to claim the given key.
tryClaim(Long) - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
Attempts to claim the given offset.
tryClaim(PositionT) - Method in class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker
Attempts to claim the block of work in the current restriction identified by the given position.
tryProcess(int, Object) - Method in class org.apache.beam.runners.jet.processors.AssignWindowP
 
tryProcess(int, Object) - Method in class org.apache.beam.runners.jet.processors.FlattenP
 
tryProcess(int, Object) - Method in class org.apache.beam.runners.jet.processors.ViewP
 
tryProcess() - Method in class org.apache.beam.runners.jet.processors.WindowGroupP
 
tryProcess(int, Object) - Method in class org.apache.beam.runners.jet.processors.WindowGroupP
 
tryProcessWatermark(Watermark) - Method in class org.apache.beam.runners.jet.processors.StatefulParDoP
 
tryProcessWatermark(Watermark) - Method in class org.apache.beam.runners.jet.processors.WindowGroupP
 
tryReturnRecordAt(boolean, ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
 
tryReturnRecordAt(boolean, Long) - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
 
tryReturnRecordAt(boolean, long) - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
 
tryReturnRecordAt(boolean, PositionT) - Method in interface org.apache.beam.sdk.io.range.RangeTracker
Atomically determines whether a record at the given position can be returned and updates internal state.
trySplit(double) - Method in class org.apache.beam.sdk.transforms.splittabledofn.ByteKeyRangeTracker
 
trySplit(double) - Method in class org.apache.beam.sdk.transforms.splittabledofn.OffsetRangeTracker
 
trySplit(double) - Method in class org.apache.beam.sdk.transforms.splittabledofn.RestrictionTracker
Splits current restriction based on fractionOfRemainder.
trySplitAtPosition(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKeyRangeTracker
 
trySplitAtPosition(Long) - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
 
trySplitAtPosition(long) - Method in class org.apache.beam.sdk.io.range.OffsetRangeTracker
 
trySplitAtPosition(PositionT) - Method in interface org.apache.beam.sdk.io.range.RangeTracker
Atomically splits the current range [RangeTracker.getStartPosition(), RangeTracker.getStopPosition()) into a "primary" part [RangeTracker.getStartPosition(), splitPosition) and a "residual" part [splitPosition, RangeTracker.getStopPosition()), assuming the current last-consumed position is within [RangeTracker.getStartPosition(), splitPosition) (i.e., splitPosition has not been consumed yet).
TUPLE_TAGS - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
TupleTag<V> - Class in org.apache.beam.sdk.values
A TupleTag is a typed tag to use as the key of a heterogeneously typed tuple, like PCollectionTuple.
TupleTag() - Constructor for class org.apache.beam.sdk.values.TupleTag
Constructs a new TupleTag, with a fresh unique id.
TupleTag(String) - Constructor for class org.apache.beam.sdk.values.TupleTag
Constructs a new TupleTag with the given id.
TupleTagList - Class in org.apache.beam.sdk.values
A TupleTagList is an immutable list of heterogeneously typed TupleTags.
type - Variable in class org.apache.beam.runners.dataflow.util.OutputReference
 
type(String) - Method in class org.apache.beam.sdk.extensions.sql.meta.Table.Builder
 
TypeConversion() - Constructor for class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TypeConversion
 
TypeDescriptor<T> - Class in org.apache.beam.sdk.values
A description of a Java type, including actual generic parameters where possible.
TypeDescriptor() - Constructor for class org.apache.beam.sdk.values.TypeDescriptor
Creates a TypeDescriptor representing the type parameter T.
TypeDescriptor(Object) - Constructor for class org.apache.beam.sdk.values.TypeDescriptor
Creates a TypeDescriptor representing the type parameter T, which should resolve to a concrete type in the context of the class clazz.
TypeDescriptor(Class<?>) - Constructor for class org.apache.beam.sdk.values.TypeDescriptor
Creates a TypeDescriptor representing the type parameter T, which should resolve to a concrete type in the context of the class clazz.
TypeDescriptors - Class in org.apache.beam.sdk.values
A utility class for creating TypeDescriptor objects for different types, such as Java primitive types, containers and KVs of other TypeDescriptor objects, and extracting type variables of parameterized types (e.g.
TypeDescriptors() - Constructor for class org.apache.beam.sdk.values.TypeDescriptors
 
TypeDescriptors.TypeVariableExtractor<InputT,OutputT> - Interface in org.apache.beam.sdk.values
TypedRead() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
 
TypedWrite() - Constructor for class org.apache.beam.sdk.io.AvroIO.TypedWrite
 
TypedWrite() - Constructor for class org.apache.beam.sdk.io.TextIO.TypedWrite
 
typeName() - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
TypeParameter<T> - Class in org.apache.beam.sdk.values
TypeParameter() - Constructor for class org.apache.beam.sdk.values.TypeParameter
 
typesEqual(Schema.Field) - Method in class org.apache.beam.sdk.schemas.Schema.Field
Returns true if two fields are equal, ignoring name and description.
typesEqual(Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
Returns true if two FieldTypes are equal.
typesEqual(Schema) - Method in class org.apache.beam.sdk.schemas.Schema
Returns true if two schemas are equal ignoring field names and descriptions.
TypeUtils - Class in org.apache.beam.sdk.extensions.sql.zetasql
Utility to convert types from Calcite Schema types.
TypeUtils() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.TypeUtils
 

U

UdafImpl<InputT,AccumT,OutputT> - Class in org.apache.beam.sdk.extensions.sql.impl
Implement AggregateFunction to take a Combine.CombineFn as UDAF.
UDF_METHOD - Static variable in interface org.apache.beam.sdk.extensions.sql.BeamSqlUdf
 
UdfImplReflectiveFunctionBase - Class in org.apache.beam.sdk.extensions.sql.impl
Beam-customized version from ReflectiveFunctionBase, to address BEAM-5921.
UdfImplReflectiveFunctionBase(Method) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.UdfImplReflectiveFunctionBase
UdfImplReflectiveFunctionBase constructor.
UdfImplReflectiveFunctionBase.ParameterListBuilder - Class in org.apache.beam.sdk.extensions.sql.impl
Helps build lists of FunctionParameter.
UdfUdafProvider - Interface in org.apache.beam.sdk.extensions.sql.meta.provider
Provider for UDF and UDAF.
UINT16 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
UINT32 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
UINT64 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
UINT8 - Static variable in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
Unbounded(SparkContext, SerializablePipelineOptions, MicrobatchSource<T, CheckpointMarkT>, int) - Constructor for class org.apache.beam.runners.spark.io.SourceRDD.Unbounded
 
unbounded() - Static method in class org.apache.beam.sdk.io.CountingSource
Deprecated.
use GenerateSequence instead
UNBOUNDED_UNKNOWN - Static variable in class org.apache.beam.sdk.extensions.sql.impl.BeamTableStatistics
 
UnboundedReader() - Constructor for class org.apache.beam.sdk.io.UnboundedSource.UnboundedReader
 
UnboundedSource<OutputT,CheckpointMarkT extends UnboundedSource.CheckpointMark> - Class in org.apache.beam.sdk.io
A Source that reads an unbounded amount of input and, because of that, supports some additional operations such as checkpointing, watermarks, and record ids.
UnboundedSource() - Constructor for class org.apache.beam.sdk.io.UnboundedSource
 
UnboundedSource.CheckpointMark - Interface in org.apache.beam.sdk.io
A marker representing the progress and state of an UnboundedSource.UnboundedReader.
UnboundedSource.CheckpointMark.NoopCheckpointMark - Class in org.apache.beam.sdk.io
A checkpoint mark that does nothing when finalized.
UnboundedSource.UnboundedReader<OutputT> - Class in org.apache.beam.sdk.io
A Reader that reads an unbounded amount of input.
UnboundedSourceP<T,CmT extends UnboundedSource.CheckpointMark> - Class in org.apache.beam.runners.jet.processors
Jet Processor implementation for reading from an unbounded Beam source.
UnboundedSourceWrapper<OutputT,CheckpointMarkT extends UnboundedSource.CheckpointMark> - Class in org.apache.beam.runners.gearpump.translators.io
wrapper over UnboundedSource for Gearpump DataSource API.
UnboundedSourceWrapper(UnboundedSource<OutputT, CheckpointMarkT>, PipelineOptions) - Constructor for class org.apache.beam.runners.gearpump.translators.io.UnboundedSourceWrapper
 
unboundedWithTimestampFn(SerializableFunction<Long, Instant>) - Static method in class org.apache.beam.sdk.io.CountingSource
unboxedType - Variable in class org.apache.beam.sdk.schemas.utils.ConvertHelpers.ConvertedSchemaInformation
 
union(Iterable<FieldAccessDescriptor>) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
 
union(Contextful...) - Static method in class org.apache.beam.sdk.transforms.Requirements
 
UnionCoder - Class in org.apache.beam.sdk.transforms.join
A UnionCoder encodes RawUnionValues.
UNKNOWN - Static variable in class org.apache.beam.sdk.extensions.sql.impl.planner.NodeStats
Returns an instance with all values set to INFINITY.
unknown() - Static method in class org.apache.beam.sdk.io.fs.MatchResult
unparse(SqlWriter, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlCheckConstraint
 
unparse(SqlWriter, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlColumnDeclaration
 
unparse(SqlWriter, int, int) - Method in class org.apache.beam.sdk.extensions.sql.impl.parser.SqlCreateExternalTable
 
unparseCall(SqlWriter, SqlCall, int, int) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamBigQuerySqlDialect
 
unparseOffsetFetch(SqlWriter, SqlNode, SqlNode) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamBigQuerySqlDialect
 
unparseSqlIntervalLiteral(SqlWriter, SqlIntervalLiteral, int, int) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamBigQuerySqlDialect
BigQuery interval syntax: INTERVAL int64 time_unit.
unparseSqlIntervalQualifier(SqlWriter, SqlIntervalQualifier, RelDataTypeSystem) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.bigquery.BeamBigQuerySqlDialect
 
unpersist() - Method in class org.apache.beam.runners.spark.util.SideInputBroadcast
 
UNSIGNED_LEXICOGRAPHICAL_COMPARATOR - Static variable in class org.apache.beam.runners.dataflow.util.RandomAccessData
 
unsupported() - Static method in interface org.apache.beam.runners.fnexecution.control.BundleSplitHandler
Returns a bundle split handler that throws on any split response.
unsupported() - Static method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandler
 
unsupported() - Static method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.BagUserStateHandlerFactory
Throws a UnsupportedOperationException on the first access.
unsupported() - Static method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.SideInputHandlerFactory
Throws a UnsupportedOperationException on the first access.
unwindowedFilename(int, int, FileBasedSink.OutputFileHints) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy
 
unwindowedFilename(int, int, FileBasedSink.OutputFileHints) - Method in class org.apache.beam.sdk.io.FileBasedSink.FilenamePolicy
When a sink has not requested windowed or triggered output, this method will be invoked to return the file resource to be created given the base output directory and a FileBasedSink.OutputFileHints containing information about the file, including a suggested (e.g.
unwrap(Class<T>) - Method in class org.apache.beam.sdk.extensions.sql.impl.CalciteConnectionWrapper
 
update(long) - Method in class org.apache.beam.runners.jet.metrics.DistributionImpl
 
update(long, long, long, long) - Method in class org.apache.beam.runners.jet.metrics.DistributionImpl
 
update(InputT) - Method in interface org.apache.beam.runners.spark.aggregators.NamedAggregators.State
 
update(InputT) - Method in interface org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregators.State
 
update(KinesisRecord) - Method in interface org.apache.beam.sdk.io.kinesis.WatermarkPolicy
 
update(KinesisRecord) - Method in class org.apache.beam.sdk.io.kinesis.WatermarkPolicyFactory.ArrivalTimeWatermarkPolicy
 
update(KinesisRecord) - Method in class org.apache.beam.sdk.io.kinesis.WatermarkPolicyFactory.CustomWatermarkPolicy
 
update(KinesisRecord) - Method in class org.apache.beam.sdk.io.kinesis.WatermarkPolicyFactory.ProcessingTimeWatermarkPolicy
 
update(long) - Method in interface org.apache.beam.sdk.metrics.Distribution
Add an observation to this distribution.
update(long, long, long, long) - Method in interface org.apache.beam.sdk.metrics.Distribution
 
updateCacheCandidates(Pipeline, SparkPipelineTranslator, EvaluationContext) - Static method in class org.apache.beam.runners.spark.SparkRunner
Evaluator that update/populate the cache candidates.
updateConsumerProperties(Map<String, Object>) - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.kafka.BeamKafkaTable
 
updateConsumerProperties(Map<String, Object>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Deprecated.
as of version 2.13. Use KafkaIO.Read.withConsumerConfigUpdates(Map) instead
updateJob(String, Job) - Method in class org.apache.beam.runners.dataflow.DataflowClient
Updates the Dataflow Job with the given jobId.
updateMetrics(String, List<MetricsApi.MonitoringInfo>) - Method in class org.apache.beam.runners.flink.metrics.FlinkMetricContainer
Update this container with metrics from the passed MetricsApi.MonitoringInfos, and send updates along to Flink's internal metrics framework.
updateProducerProperties(Map<String, Object>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
Deprecated.
as of version 2.13. Use KafkaIO.Write.withProducerConfigUpdates(Map) instead.
updateProducerProperties(Map<String, Object>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
Deprecated.
updateSerializedOptions(String, Map<String, String>) - Static method in class org.apache.beam.sdk.options.ValueProviders
Deprecated.
updateWatermarks(Bundle<?, ? extends CollectionT>, WatermarkManager.TimerUpdate, ExecutableT, Bundle<?, ? extends CollectionT>, Iterable<? extends Bundle<?, ? extends CollectionT>>, Instant) - Method in class org.apache.beam.runners.direct.WatermarkManager
Updates the watermarks of a executable with one or more inputs.
updateWindowingStrategy(WindowingStrategy<?, ?>) - Method in class org.apache.beam.sdk.transforms.GroupByKey
 
UploadIdResponseInterceptor - Class in org.apache.beam.sdk.extensions.gcp.util
Implements a response intercepter that logs the upload id if the upload id header exists and it is the first request (does not have upload_id parameter in the request).
UploadIdResponseInterceptor() - Constructor for class org.apache.beam.sdk.extensions.gcp.util.UploadIdResponseInterceptor
 
upTo(long) - Static method in class org.apache.beam.sdk.io.CountingSource
Deprecated.
use GenerateSequence instead
URN - Static variable in class org.apache.beam.sdk.fn.data.RemoteGrpcPortRead
 
URN - Static variable in class org.apache.beam.sdk.fn.data.RemoteGrpcPortWrite
 
URN - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.ExternalRead
 
URN - Static variable in class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite
 
URN - Static variable in class org.apache.beam.sdk.io.GenerateSequence.External
 
URN - Static variable in class org.apache.beam.sdk.io.kafka.KafkaIO.Read.External
 
URN - Static variable in class org.apache.beam.sdk.io.kafka.KafkaIO.Write.External
 
USE_INDEXED_FORMAT - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
useAbstractConvertersForConversion(RelTraitSet, RelTraitSet) - Method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
 
useAvroLogicalTypes() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Enables interpreting logical types into their corresponding types (ie.
useBeamSchema() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
If true, then the BigQuery schema will be inferred from the input schema.
useDlq() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonTableProvider.PubsubIOTableConfiguration
 
useDlq() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubMessageToRow
 
useFlatSchema() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubMessageToRow
 
USER_FN - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
USER_NAME - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
UserAgentFactory() - Constructor for class org.apache.beam.sdk.options.PipelineOptions.UserAgentFactory
 
userStateId() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.BagUserStateSpec
 
USES_KEYED_STATE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
UsesAttemptedMetrics - Class in org.apache.beam.sdk.testing
Category tag for validation tests which utilize Metrics.
UsesAttemptedMetrics() - Constructor for class org.apache.beam.sdk.testing.UsesAttemptedMetrics
 
UsesBoundedSplittableParDo - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilize splittable ParDo with a DoFn.BoundedPerElement DoFn.
UsesBundleFinalizer - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which use DoFn.BundleFinalizer.
UsesCommittedMetrics - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilize Metrics.
UsesCounterMetrics - Class in org.apache.beam.sdk.testing
Category tag for validation tests which utilize Counter.
UsesCounterMetrics() - Constructor for class org.apache.beam.sdk.testing.UsesCounterMetrics
 
UsesCrossLanguageTransforms - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which use cross-language transforms.
UsesCustomWindowMerging - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilize custom window merging.
UsesDistributionMetrics - Class in org.apache.beam.sdk.testing
Category tag for validation tests which utilize Distribution.
UsesDistributionMetrics() - Constructor for class org.apache.beam.sdk.testing.UsesDistributionMetrics
 
UsesFailureMessage - Interface in org.apache.beam.sdk.testing
Category tag for tests which validate that currect failure message is provided by failed pipeline.
UsesGaugeMetrics - Class in org.apache.beam.sdk.testing
Category tag for validation tests which utilize Gauge.
UsesGaugeMetrics() - Constructor for class org.apache.beam.sdk.testing.UsesGaugeMetrics
 
UsesImpulse - Class in org.apache.beam.sdk.testing
Category for tests that use Impulse transformations.
UsesImpulse() - Constructor for class org.apache.beam.sdk.testing.UsesImpulse
 
UsesKms - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilize --tempRoot from TestPipelineOptions and and expect a default KMS key enable for the bucket specified.
UsesMapState - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilize MapState.
UsesMetricsPusher - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilize the metrics pusher feature.
UsesParDoLifecycle - Interface in org.apache.beam.sdk.testing
Category tag for the ParDoLifecycleTest for exclusion (BEAM-3241).
UsesRequiresTimeSortedInput - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilizeDoFn.RequiresTimeSortedInput in stateful ParDo.
UsesSchema - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilize schemas.
UsesSetState - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilize SetState.
UsesSideInputs - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which use sideinputs.
UsesSideInputsWithDifferentCoders - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which use multiple side inputs with different coders.
UsesSplittableParDoWithWindowedSideInputs - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilize splittable ParDo and use windowed side inputs.
UsesStatefulParDo - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilize stateful ParDo.
UsesStrictTimerOrdering - Interface in org.apache.beam.sdk.testing
Category for tests that enforce strict event-time ordering of fired timers, even in situations where multiple tests mutually set one another and watermark hops arbitrarily far to the future.
UsesSystemMetrics - Interface in org.apache.beam.sdk.testing
Category tag for tests that use System metrics.
UsesTestStream - Interface in org.apache.beam.sdk.testing
Category tag for tests that use TestStream, which is not a part of the Beam model but a special feature currently only implemented by the direct runner and the Flink Runner (streaming).
UsesTestStreamWithMultipleStages - Interface in org.apache.beam.sdk.testing
Subcategory for UsesTestStream tests which use TestStream # across multiple stages.
UsesTestStreamWithOutputTimestamp - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which use outputTimestamp.
UsesTestStreamWithProcessingTime - Interface in org.apache.beam.sdk.testing
Subcategory for UsesTestStream tests which use the processing time feature of TestStream.
UsesTimerMap - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which use timerMap.
UsesTimersInParDo - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilize timers in ParDo.
UsesUnboundedPCollections - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilize at least one unbounded PCollection.
UsesUnboundedSplittableParDo - Interface in org.apache.beam.sdk.testing
Category tag for validation tests which utilize splittable ParDo with a DoFn.UnboundedPerElement DoFn.
useTimestampAttribute() - Method in class org.apache.beam.sdk.extensions.sql.meta.provider.pubsub.PubsubJsonTableProvider.PubsubIOTableConfiguration
 
using(String...) - Method in class org.apache.beam.sdk.schemas.transforms.Join.Impl
Perform a natural join between the PCollections.
using(Integer...) - Method in class org.apache.beam.sdk.schemas.transforms.Join.Impl
Perform a natural join between the PCollections.
using(FieldAccessDescriptor) - Method in class org.apache.beam.sdk.schemas.transforms.Join.Impl
Perform a natural join between the PCollections.
usingFnApiClient(InstructionRequestHandler, FnDataService) - Static method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient
Creates a client for a particular SDK harness.
usingStandardSql() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
Enables BigQuery's Standard SQL dialect when reading from a query.
usingStandardSql() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
Utils - Class in org.apache.beam.runners.jet
Various common methods used by the Jet based runner.
Utils() - Constructor for class org.apache.beam.runners.jet.Utils
 
Utils.ByteArrayKey - Class in org.apache.beam.runners.jet
A wrapper of byte[] that can be used as a hash-map key.

V

v1() - Static method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreIO
Returns a DatastoreV1 that provides an API for accessing Cloud Datastore through v1 version of Datastore Client library.
validate() - Method in class org.apache.beam.runners.spark.io.MicrobatchSource
 
validate(SqlNode) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl
 
validate() - Method in class org.apache.beam.sdk.io.AvroSource
 
validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
 
validate() - Method in class org.apache.beam.sdk.io.CompressedSource
Validates that the delegate source is a valid source and that the channel factory is not null.
validate() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource
 
validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.FileBasedSink
 
validate() - Method in class org.apache.beam.sdk.io.FileBasedSource
 
validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
 
validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
 
validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
 
validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
 
validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.WriteWithResults
 
validate() - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
validate() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.HadoopInputFormatBoundedSource
 
validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write
 
validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
 
validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
 
validate(T) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
 
validate() - Method in class org.apache.beam.sdk.io.OffsetBasedSource
 
validate() - Method in class org.apache.beam.sdk.io.Source
Checks that this source is valid, before it can be used in a pipeline.
validate(PipelineOptions) - Method in class org.apache.beam.sdk.io.WriteFiles
 
validate(Class<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.options.PipelineOptionsValidator
Validates that the passed PipelineOptions conforms to all the validation criteria from the passed in interface.
validate(PipelineOptions) - Method in class org.apache.beam.sdk.transforms.PTransform
Called before running the Pipeline to verify this transform is fully and correctly specified.
VALIDATE_TIME_INTERVAL - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
VALIDATE_TIMESTAMP - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlOperators
 
validateAndGetType(SqlNode) - Method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl
 
validateCli(Class<T>, PipelineOptions) - Static method in class org.apache.beam.sdk.options.PipelineOptionsValidator
Validates that the passed PipelineOptions from command line interface (CLI) conforms to all the validation criteria from the passed in interface.
validateGetOutputTimestamp(WindowFn<T, W>, long) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
Assigns the given timestamp to windows using the specified windowFn, and verifies that result of windowFn.getOutputTime for later windows (as defined by maxTimestamp won't prevent the watermark from passing the end of earlier windows.
validateGetOutputTimestamps(WindowFn<T, W>, TimestampCombiner, List<List<Long>>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
Verifies that later-ending merged windows from any of the timestamps hold up output of earlier-ending windows, using the provided WindowFn and TimestampCombiner.
validateGetOutputTimestampsWithValue(WindowFn<T, W>, TimestampCombiner, List<List<TimestampedValue<T>>>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
Verifies that later-ending merged windows from any of the timestampValues hold up output of earlier-ending windows, using the provided WindowFn and TimestampCombiner.
validateGetOutputTimestampWithValue(WindowFn<T, W>, TimestampedValue<T>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
Assigns the given timestampedValue to windows using the specified windowFn, and verifies that result of windowFn.getOutputTime for later windows (as defined by maxTimestamp won't prevent the watermark from passing the end of earlier windows.
validateInputFilePatternSupported(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
Validates the the input GCS path is accessible and that the path is well formed.
validateInputFilePatternSupported(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
 
validateInputFilePatternSupported(String) - Method in interface org.apache.beam.sdk.extensions.gcp.storage.PathValidator
Validate that a file pattern is conforming.
validateJavaBean(List<FieldValueTypeInformation>, List<FieldValueTypeInformation>) - Static method in class org.apache.beam.sdk.schemas.utils.JavaBeanUtils
 
validateNonInterferingOutputTimes(WindowFn<T, W>, long) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
Assigns the given timestamp to windows using the specified windowFn, and verifies that result of windowFn.getOutputTimestamp for each window is within the proper bound.
validateNonInterferingOutputTimesWithValue(WindowFn<T, W>, TimestampedValue<T>) - Static method in class org.apache.beam.sdk.testing.WindowFnTestUtils
Assigns the given timestampedValue to windows using the specified windowFn, and verifies that result of windowFn.getOutputTimestamp for each window is within the proper bound.
validateOutputFilePrefixSupported(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
Validates the the output GCS path is accessible and that the path is well formed.
validateOutputFilePrefixSupported(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
 
validateOutputFilePrefixSupported(String) - Method in interface org.apache.beam.sdk.extensions.gcp.storage.PathValidator
Validate that an output file prefix is conforming.
validateOutputResourceSupported(ResourceId) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
 
validateOutputResourceSupported(ResourceId) - Method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
 
validateOutputResourceSupported(ResourceId) - Method in interface org.apache.beam.sdk.extensions.gcp.storage.PathValidator
Validates that an output path is conforming.
ValidatesRunner - Interface in org.apache.beam.sdk.testing
Category tag for tests which validate that a Beam runner is correctly implemented.
validateTimeInterval(Long, TimeUnit) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
This function validates that interval is compatible with ZetaSQL timestamp values range.
validateTimestamp(Long) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.DateTimeUtils
This function validates that Long representation of timestamp is compatible with ZetaSQL timestamp values range.
validateTransform() - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
Validates construction of this transform.
Validation - Annotation Type in org.apache.beam.sdk.options
Validation represents a set of annotations that can be used to annotate getter properties on PipelineOptions with information representing the validation criteria to be used when validating with the PipelineOptionsValidator.
Validation.Required - Annotation Type in org.apache.beam.sdk.options
This criteria specifies that the value must be not null.
validator() - Method in class org.apache.beam.sdk.schemas.transforms.Cast
 
VALUE - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
value() - Method in class org.apache.beam.runners.spark.aggregators.NamedAggregatorsAccumulator
 
value() - Method in class org.apache.beam.runners.spark.metrics.MetricsContainerStepMapAccumulator
 
value() - Method in class org.apache.beam.runners.spark.structuredstreaming.aggregators.NamedAggregatorsAccumulator
 
value() - Method in class org.apache.beam.runners.spark.structuredstreaming.metrics.MetricsContainerStepMapAccumulator
 
value(Object) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedList
 
value(Object) - Method in class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedMap
 
Value(int) - Constructor for class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType.Value
 
Value(EnumerationType.Value, Object) - Constructor for class org.apache.beam.sdk.schemas.logicaltypes.OneOfType.Value
 
value() - Static method in class org.apache.beam.sdk.state.StateSpecs
Create a StateSpec for a single value of type T.
value(Coder<T>) - Static method in class org.apache.beam.sdk.state.StateSpecs
Identical to StateSpecs.value(), but with a coder explicitly supplied.
valueCoder() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.BagUserStateSpec
 
valueCoder() - Method in interface org.apache.beam.runners.fnexecution.state.StateRequestHandlers.MultimapSideInputHandler
Returns the Coder to use for the elements of the resulting values iterable.
ValueInSingleWindow<T> - Class in org.apache.beam.sdk.values
An immutable tuple of value, timestamp, window, and pane.
ValueInSingleWindow() - Constructor for class org.apache.beam.sdk.values.ValueInSingleWindow
 
ValueInSingleWindow.Coder<T> - Class in org.apache.beam.sdk.values
A coder for ValueInSingleWindow.
valueOf(String) - Static method in enum org.apache.beam.runners.dataflow.options.DataflowPipelineOptions.FlexResourceSchedulingGoal
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.Level
Deprecated.
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.runners.local.ExecutionDriver.DriverState
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.annotations.Experimental.Kind
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.coders.CannotProvideCoderException.ReasonCode
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options.SorterType
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters.Kind
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamSetOperatorRelBase.OpType
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sql.meta.ProjectSupport
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider.PushDownOptions
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.cassandra.CassandraIO.MutationType
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.clickhouse.TableSchema.DefaultType
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.clickhouse.TableSchema.TypeName
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.CompressedSource.CompressionMode
Deprecated.
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.Compression
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
Deprecated.
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.FileBasedSource.Mode
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.FileIO.ReadMatches.DirectoryTreatment
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.fs.EmptyMatchTreatment
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.fs.MatchResult.Status
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.fs.MoveOptions.StandardMoveOptions
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.fs.ResolveOptions.StandardResolveOptions
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead.Method
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead.QueryPriority
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.CreateDisposition
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.Method
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.SchemaUpdateOption
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.WriteDisposition
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.ConversionOptions.TruncateTimestamps
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.gcp.spanner.SpannerIO.FailureMode
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.kafka.KafkaTimestampType
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.redis.RedisIO.Write.Method
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.TextIO.CompressionType
Deprecated.
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.TFRecordIO.CompressionType
Deprecated.
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.io.xml.XmlIO.Read.CompressionType
Deprecated.
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.options.PipelineOptions.CheckEnabled
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.options.PortablePipelineOptions.RetrievalServiceType
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.options.SdkHarnessOptions.LogLevel
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.Pipeline.PipelineVisitor.CompositeBehavior
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.PipelineResult.State
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.ListQualifier
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.MapQualifier
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier.Kind
Returns the enum constant of this type with the specified name.
valueOf(String) - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
Return an EnumerationType.Value corresponding to one of the enumeration strings.
valueOf(int) - Method in class org.apache.beam.sdk.schemas.logicaltypes.EnumerationType
Return an EnumerationType.Value corresponding to one of the enumeration integer values.
valueOf(String) - Static method in enum org.apache.beam.sdk.schemas.Schema.EquivalenceNullablePolicy
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.schemas.Schema.TypeName
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.state.TimeDomain
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.testing.SourceTestUtils.ExpectedSplitOutcome
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.testing.TestStream.EventType
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.display.DisplayData.Type
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.DoFnTester.CloningBehavior
Deprecated.
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.windowing.PaneInfo.Timing
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.windowing.Window.ClosingBehavior
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.transforms.windowing.Window.OnTimeBehavior
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.values.PCollection.IsBounded
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.beam.sdk.values.WindowingStrategy.AccumulationMode
Returns the enum constant of this type with the specified name.
ValueProvider<T> - Interface in org.apache.beam.sdk.options
A ValueProvider abstracts the notion of fetching a value that may or may not be currently available.
ValueProvider.Deserializer - Class in org.apache.beam.sdk.options
For internal use only; no backwards compatibility guarantees.
ValueProvider.NestedValueProvider<T,X> - Class in org.apache.beam.sdk.options
ValueProvider.NestedValueProvider is an implementation of ValueProvider that allows for wrapping another ValueProvider object.
ValueProvider.RuntimeValueProvider<T> - Class in org.apache.beam.sdk.options
ValueProvider.RuntimeValueProvider is an implementation of ValueProvider that allows for a value to be provided at execution time rather than at graph construction time.
ValueProvider.Serializer - Class in org.apache.beam.sdk.options
For internal use only; no backwards compatibility guarantees.
ValueProvider.StaticValueProvider<T> - Class in org.apache.beam.sdk.options
ValueProvider.StaticValueProvider is an implementation of ValueProvider that allows for a static value to be provided.
ValueProviders - Class in org.apache.beam.sdk.options
Utilities for working with the ValueProvider interface.
values() - Static method in enum org.apache.beam.runners.dataflow.options.DataflowPipelineOptions.FlexResourceSchedulingGoal
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.Level
Deprecated.
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.runners.local.ExecutionDriver.DriverState
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.annotations.Experimental.Kind
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.coders.CannotProvideCoderException.ReasonCode
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.extensions.sorter.ExternalSorter.Options.SorterType
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.extensions.sql.impl.QueryPlanner.QueryParameters.Kind
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamLogicalConvention
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.extensions.sql.impl.rel.BeamSetOperatorRelBase.OpType
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.extensions.sql.meta.ProjectSupport
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.extensions.sql.meta.provider.test.TestTableProvider.PushDownOptions
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.cassandra.CassandraIO.MutationType
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.clickhouse.TableSchema.DefaultType
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.clickhouse.TableSchema.TypeName
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.CompressedSource.CompressionMode
Deprecated.
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.Compression
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.FileBasedSink.CompressionType
Deprecated.
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.FileBasedSource.Mode
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.FileIO.ReadMatches.DirectoryTreatment
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.fs.EmptyMatchTreatment
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.fs.MatchResult.Status
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.fs.MoveOptions.StandardMoveOptions
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.fs.ResolveOptions.StandardResolveOptions
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead.Method
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead.QueryPriority
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.CreateDisposition
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.Method
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.SchemaUpdateOption
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.WriteDisposition
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.gcp.bigquery.BigQueryUtils.ConversionOptions.TruncateTimestamps
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.gcp.spanner.SpannerIO.FailureMode
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
Writes just the values to Kafka.
values() - Static method in enum org.apache.beam.sdk.io.kafka.KafkaTimestampType
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.redis.RedisIO.Write.Method
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.TextIO.CompressionType
Deprecated.
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.TFRecordIO.CompressionType
Deprecated.
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.io.xml.XmlIO.Read.CompressionType
Deprecated.
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.options.PipelineOptions.CheckEnabled
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.options.PortablePipelineOptions.RetrievalServiceType
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.options.SdkHarnessOptions.LogLevel
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.Pipeline.PipelineVisitor.CompositeBehavior
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.PipelineResult.State
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.ListQualifier
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.MapQualifier
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.schemas.FieldAccessDescriptor.FieldDescriptor.Qualifier.Kind
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.schemas.Schema.EquivalenceNullablePolicy
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.schemas.Schema.TypeName
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Method in class org.apache.beam.sdk.schemas.utils.ByteBuddyUtils.TransformingMap
 
values() - Method in interface org.apache.beam.sdk.state.MapState
Returns an Iterable over the values contained in this map.
values() - Static method in enum org.apache.beam.sdk.state.TimeDomain
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.testing.SourceTestUtils.ExpectedSplitOutcome
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.testing.TestStream.EventType
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in class org.apache.beam.sdk.transforms.Deduplicate
Returns a deduplication transform that deduplicates values for up to 10 mins within the processing time domain.
values() - Static method in enum org.apache.beam.sdk.transforms.display.DisplayData.Type
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.transforms.DoFnTester.CloningBehavior
Deprecated.
Returns an array containing the constants of this enum type, in the order they are declared.
Values<V> - Class in org.apache.beam.sdk.transforms
Values<V> takes a PCollection of KV<K, V>s and returns a PCollection<V> of the values.
values() - Static method in enum org.apache.beam.sdk.transforms.windowing.PaneInfo.Timing
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.transforms.windowing.TimestampCombiner
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.transforms.windowing.Window.ClosingBehavior
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.transforms.windowing.Window.OnTimeBehavior
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.values.PCollection.IsBounded
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.beam.sdk.values.WindowingStrategy.AccumulationMode
Returns an array containing the constants of this enum type, in the order they are declared.
ValuesSource<T> - Class in org.apache.beam.runners.gearpump.translators.io
unbounded source that reads from a Java Iterable.
ValuesSource(Iterable<T>, Coder<T>) - Constructor for class org.apache.beam.runners.gearpump.translators.io.ValuesSource
 
ValueState<T> - Interface in org.apache.beam.sdk.state
A ReadableState cell containing a single value.
ValueWithRecordId<ValueT> - Class in org.apache.beam.sdk.values
For internal use only; no backwards compatibility guarantees.
ValueWithRecordId(ValueT, byte[]) - Constructor for class org.apache.beam.sdk.values.ValueWithRecordId
 
ValueWithRecordId.StripIdsDoFn<T> - Class in org.apache.beam.sdk.values
DoFn to turn a ValueWithRecordId<T> back to the value T.
ValueWithRecordId.ValueWithRecordIdCoder<ValueT> - Class in org.apache.beam.sdk.values
A Coder for ValueWithRecordId, using a wrapped value Coder.
ValueWithRecordIdCoder(Coder<ValueT>) - Constructor for class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
 
VARBINARY - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
VARCHAR - Static variable in class org.apache.beam.sdk.extensions.sql.impl.utils.CalciteUtils
 
VarianceFn<T extends java.lang.Number> - Class in org.apache.beam.sdk.extensions.sql.impl.transform.agg
Combine.CombineFn for Variance on Number types.
VarIntCoder - Class in org.apache.beam.sdk.coders
A Coder that encodes Integers using between 1 and 5 bytes.
VarLongCoder - Class in org.apache.beam.sdk.coders
A Coder that encodes Longs using between 1 and 10 bytes.
verifyCompatibility(Schema) - Method in class org.apache.beam.sdk.schemas.transforms.Cast
 
verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
 
verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
 
verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
 
verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
 
verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
 
verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
 
verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
 
verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
 
verifyCompatibility(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
Throw IncompatibleWindowException if this WindowFn does not perform the same merging as the given $WindowFn.
verifyDeterministic() - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData.RandomAccessDataCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.AtomicCoder
Throw Coder.NonDeterministicException if the coding is not deterministic.
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.AvroCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BigDecimalCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BigEndianIntegerCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BigEndianLongCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BigEndianShortCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BigIntegerCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.BitSetCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.ByteArrayCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.ByteCoder
Throw Coder.NonDeterministicException if the coding is not deterministic.
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.Coder
Throw Coder.NonDeterministicException if the coding is not deterministic.
verifyDeterministic(Coder<?>, String, Iterable<Coder<?>>) - Static method in class org.apache.beam.sdk.coders.Coder
Verifies all of the provided coders are deterministic.
verifyDeterministic(Coder<?>, String, Coder<?>...) - Static method in class org.apache.beam.sdk.coders.Coder
Verifies all of the provided coders are deterministic.
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.CustomCoder
Throw Coder.NonDeterministicException if the coding is not deterministic.
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.DelegateCoder
Throw Coder.NonDeterministicException if the coding is not deterministic.
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.DoubleCoder
Throw Coder.NonDeterministicException if the coding is not deterministic.
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.DurationCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.FloatCoder
Throw Coder.NonDeterministicException if the coding is not deterministic.
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.InstantCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.IterableLikeCoder
Throw Coder.NonDeterministicException if the coding is not deterministic.
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.KvCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.LengthPrefixCoder
LengthPrefixCoder is deterministic if the nested Coder is.
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.ListCoder
List sizes are always known, so ListIterable may be deterministic while the general IterableLikeCoder is not.
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.MapCoder
Throw Coder.NonDeterministicException if the coding is not deterministic.
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.NullableCoder
NullableCoder is deterministic if the nested Coder is.
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.SerializableCoder
Throw Coder.NonDeterministicException if the coding is not deterministic.
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.SetCoder
Throw Coder.NonDeterministicException if the coding is not deterministic.
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.ShardedKeyCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.SnappyCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.StringDelegateCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.StringUtf8Coder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.TextualIntegerCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.VarIntCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.VarLongCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.coders.VoidCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.extensions.protobuf.ByteStringCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResultCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryInsertErrorCoder
Throw Coder.NonDeterministicException if the coding is not deterministic.
verifyDeterministic() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV2
 
verifyDeterministic() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestinationCoderV3
 
verifyDeterministic() - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableRowJsonCoder
Throw Coder.NonDeterministicException if the coding is not deterministic.
verifyDeterministic() - Method in class org.apache.beam.sdk.io.hadoop.WritableCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.io.kafka.KafkaRecordCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.io.kafka.ProducerRecordCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.schemas.SchemaCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.testing.TestStream.TestStreamCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.join.CoGbkResult.CoGbkResultCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.join.UnionCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindow.Coder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.windowing.IntervalWindow.IntervalWindowCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.transforms.windowing.PaneInfo.PaneInfoCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.values.TimestampedValue.TimestampedValueCoder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.values.ValueInSingleWindow.Coder
 
verifyDeterministic() - Method in class org.apache.beam.sdk.values.ValueWithRecordId.ValueWithRecordIdCoder
 
verifyFieldValue(Object, Schema.FieldType, String) - Static method in class org.apache.beam.sdk.values.SchemaVerification
 
verifyPAssertsSucceeded(Pipeline, PipelineResult) - Static method in class org.apache.beam.sdk.testing.TestPipeline
Verifies all {PAsserts} in the pipeline have been executed and were successful.
verifyPath(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.GcsPathValidator
 
verifyPath(String) - Method in class org.apache.beam.sdk.extensions.gcp.storage.NoopPathValidator
 
verifyPath(String) - Method in interface org.apache.beam.sdk.extensions.gcp.storage.PathValidator
Validate that a path is a valid path and that the path is accessible.
via(Contextful<Contextful.Fn<UserT, OutputT>>, Contextful<Contextful.Fn<DestinationT, FileIO.Sink<OutputT>>>) - Method in class org.apache.beam.sdk.io.FileIO.Write
Specifies how to create a FileIO.Sink for a particular destination and how to map the element type to the sink's output type.
via(Contextful<Contextful.Fn<UserT, OutputT>>, FileIO.Sink<OutputT>) - Method in class org.apache.beam.sdk.io.FileIO.Write
Like FileIO.Write.via(Contextful, Contextful), but uses the same sink for all destinations.
via(Contextful<Contextful.Fn<DestinationT, FileIO.Sink<UserT>>>) - Method in class org.apache.beam.sdk.io.FileIO.Write
Like FileIO.Write.via(Contextful, Contextful), but the output type of the sink is the same as the type of the input collection.
via(FileIO.Sink<UserT>) - Method in class org.apache.beam.sdk.io.FileIO.Write
Like FileIO.Write.via(Contextful), but uses the same FileIO.Sink for all destinations.
via(InferableFunction<? super InputT, ? extends Iterable<OutputT>>) - Static method in class org.apache.beam.sdk.transforms.FlatMapElements
For a InferableFunction<InputT, ? extends Iterable<OutputT>> fn, return a PTransform that applies fn to every element of the input PCollection<InputT> and outputs all of the elements to the output PCollection<OutputT>.
via(SimpleFunction<? super InputT, ? extends Iterable<OutputT>>) - Static method in class org.apache.beam.sdk.transforms.FlatMapElements
Binary compatibility adapter for FlatMapElements.via(ProcessFunction).
via(ProcessFunction<NewInputT, ? extends Iterable<OutputT>>) - Method in class org.apache.beam.sdk.transforms.FlatMapElements
For a ProcessFunction<InputT, ? extends Iterable<OutputT>> fn, returns a PTransform that applies fn to every element of the input PCollection<InputT> and outputs all of the elements to the output PCollection<OutputT>.
via(SerializableFunction<NewInputT, ? extends Iterable<OutputT>>) - Method in class org.apache.beam.sdk.transforms.FlatMapElements
Binary compatibility adapter for FlatMapElements.via(ProcessFunction).
via(Contextful<Contextful.Fn<NewInputT, Iterable<OutputT>>>) - Method in class org.apache.beam.sdk.transforms.FlatMapElements
Like FlatMapElements.via(ProcessFunction), but allows access to additional context.
via(InferableFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.MapElements
For InferableFunction<InputT, OutputT> fn, returns a PTransform that takes an input PCollection<InputT> and returns a PCollection<OutputT> containing fn.apply(v) for every element v in the input.
via(SimpleFunction<InputT, OutputT>) - Static method in class org.apache.beam.sdk.transforms.MapElements
Binary compatibility adapter for MapElements.via(InferableFunction).
via(ProcessFunction<NewInputT, OutputT>) - Method in class org.apache.beam.sdk.transforms.MapElements
For a ProcessFunction<InputT, OutputT> fn and output type descriptor, returns a PTransform that takes an input PCollection<InputT> and returns a PCollection<OutputT> containing fn.apply(v) for every element v in the input.
via(SerializableFunction<NewInputT, OutputT>) - Method in class org.apache.beam.sdk.transforms.MapElements
Binary compatibility adapter for MapElements.via(ProcessFunction).
via(Contextful<Contextful.Fn<NewInputT, OutputT>>) - Method in class org.apache.beam.sdk.transforms.MapElements
Like MapElements.via(ProcessFunction), but supports access to context, such as side inputs.
viaRandomKey() - Static method in class org.apache.beam.sdk.transforms.Reshuffle
Deprecated.
Encapsulates the sequence "pair input with unique key, apply Reshuffle.of(), drop the key" commonly used to break fusion.
VideoIntelligence - Class in org.apache.beam.sdk.extensions.ml
Factory class for AnnotateVideo subclasses.
VideoIntelligence() - Constructor for class org.apache.beam.sdk.extensions.ml.VideoIntelligence
 
VideoIntelligence.AnnotateVideoBytesWithContext - Class in org.apache.beam.sdk.extensions.ml
Implementation of AnnotateVideo accepting KVs as contents of input PCollection.
VideoIntelligence.AnnotateVideoFromBytes - Class in org.apache.beam.sdk.extensions.ml
Implementation of AnnotateVideo accepting ByteStrings as contents of input PCollection.
VideoIntelligence.AnnotateVideoFromURI - Class in org.apache.beam.sdk.extensions.ml
Implementation of AnnotateVideo accepting Strings as contents of input PCollection.
VideoIntelligence.AnnotateVideoURIWithContext - Class in org.apache.beam.sdk.extensions.ml
Implementation of AnnotateVideo accepting KVs as contents of input PCollection.
View - Class in org.apache.beam.sdk.transforms
Transforms for creating PCollectionViews from PCollections (to read them as side inputs).
View.AsIterable<T> - Class in org.apache.beam.sdk.transforms
For internal use only; no backwards-compatibility guarantees.
View.AsList<T> - Class in org.apache.beam.sdk.transforms
For internal use only; no backwards-compatibility guarantees.
View.AsMap<K,V> - Class in org.apache.beam.sdk.transforms
For internal use only; no backwards-compatibility guarantees.
View.AsMultimap<K,V> - Class in org.apache.beam.sdk.transforms
For internal use only; no backwards-compatibility guarantees.
View.AsSingleton<T> - Class in org.apache.beam.sdk.transforms
For internal use only; no backwards-compatibility guarantees.
View.CreatePCollectionView<ElemT,ViewT> - Class in org.apache.beam.sdk.transforms
For internal use only; no backwards-compatibility guarantees.
viewAsValues(PCollectionView<V>, Coder<V>) - Static method in class org.apache.beam.sdk.transforms.Reify
Pairs each element in a collection with the value of a side input associated with the element's window.
ViewFn<PrimitiveViewT,ViewT> - Class in org.apache.beam.sdk.transforms
For internal use only; no backwards-compatibility guarantees.
ViewFn() - Constructor for class org.apache.beam.sdk.transforms.ViewFn
 
viewInGlobalWindow(PCollectionView<V>, Coder<V>) - Static method in class org.apache.beam.sdk.transforms.Reify
Returns a PCollection consisting of a single element, containing the value of the given view in the global window.
ViewP - Class in org.apache.beam.runners.jet.processors
Jet Processor implementation for Beam's side input producing primitives.
visitArrayQualifier(FieldSpecifierNotationParser.ArrayQualifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
Visit a parse tree produced by FieldSpecifierNotationParser.arrayQualifier().
visitArrayQualifier(FieldSpecifierNotationParser.ArrayQualifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
Visit a parse tree produced by FieldSpecifierNotationParser.arrayQualifier().
visitArrayQualifierList(FieldSpecifierNotationParser.ArrayQualifierListContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
Visit a parse tree produced by the arrayQualifierList labeled alternative in FieldSpecifierNotationParser.qualifierList().
visitArrayQualifierList(FieldSpecifierNotationParser.ArrayQualifierListContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
Visit a parse tree produced by the arrayQualifierList labeled alternative in FieldSpecifierNotationParser.qualifierList().
visitDotExpression(FieldSpecifierNotationParser.DotExpressionContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
Visit a parse tree produced by FieldSpecifierNotationParser.dotExpression().
visitDotExpression(FieldSpecifierNotationParser.DotExpressionContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
Visit a parse tree produced by FieldSpecifierNotationParser.dotExpression().
visitErrorNode(ErrorNode) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
visitFieldSpecifier(FieldSpecifierNotationParser.FieldSpecifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
Visit a parse tree produced by FieldSpecifierNotationParser.fieldSpecifier().
visitFieldSpecifier(FieldSpecifierNotationParser.FieldSpecifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
Visit a parse tree produced by FieldSpecifierNotationParser.fieldSpecifier().
visitMapQualifier(FieldSpecifierNotationParser.MapQualifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
Visit a parse tree produced by FieldSpecifierNotationParser.mapQualifier().
visitMapQualifier(FieldSpecifierNotationParser.MapQualifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
Visit a parse tree produced by FieldSpecifierNotationParser.mapQualifier().
visitMapQualifierList(FieldSpecifierNotationParser.MapQualifierListContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
Visit a parse tree produced by the mapQualifierList labeled alternative in FieldSpecifierNotationParser.qualifierList().
visitMapQualifierList(FieldSpecifierNotationParser.MapQualifierListContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
Visit a parse tree produced by the mapQualifierList labeled alternative in FieldSpecifierNotationParser.qualifierList().
visitPrimitiveTransform(TransformHierarchy.Node) - Method in class org.apache.beam.runners.gearpump.translators.GearpumpPipelineTranslator
 
visitPrimitiveTransform(TransformHierarchy.Node) - Method in class org.apache.beam.runners.spark.SparkRunner.Evaluator
 
visitPrimitiveTransform(TransformHierarchy.Node) - Method in class org.apache.beam.runners.spark.structuredstreaming.translation.PipelineTranslator
 
visitPrimitiveTransform(TransformHierarchy.Node) - Method in class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
 
visitPrimitiveTransform(TransformHierarchy.Node) - Method in interface org.apache.beam.sdk.Pipeline.PipelineVisitor
Called for each primitive transform after all of its topological predecessors and inputs have been visited.
visitQualifiedComponent(FieldSpecifierNotationParser.QualifiedComponentContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
visitQualifiedComponent(FieldSpecifierNotationParser.QualifiedComponentContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
visitQualifyComponent(FieldSpecifierNotationParser.QualifyComponentContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
Visit a parse tree produced by the qualifyComponent labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
visitQualifyComponent(FieldSpecifierNotationParser.QualifyComponentContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
Visit a parse tree produced by the qualifyComponent labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
visitSimpleIdentifier(FieldSpecifierNotationParser.SimpleIdentifierContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
Visit a parse tree produced by the simpleIdentifier labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
visitSimpleIdentifier(FieldSpecifierNotationParser.SimpleIdentifierContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
Visit a parse tree produced by the simpleIdentifier labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
visitTerminal(TerminalNode) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseListener
visitValue(PValue, TransformHierarchy.Node) - Method in class org.apache.beam.runners.gearpump.translators.GearpumpPipelineTranslator
 
visitValue(PValue, TransformHierarchy.Node) - Method in class org.apache.beam.sdk.Pipeline.PipelineVisitor.Defaults
 
visitValue(PValue, TransformHierarchy.Node) - Method in interface org.apache.beam.sdk.Pipeline.PipelineVisitor
Called for each value after the transform that produced the value has been visited.
visitWildcard(FieldSpecifierNotationParser.WildcardContext) - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationBaseVisitor
Visit a parse tree produced by the wildcard labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
visitWildcard(FieldSpecifierNotationParser.WildcardContext) - Method in interface org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationVisitor
Visit a parse tree produced by the wildcard labeled alternative in FieldSpecifierNotationParser.dotExpressionComponent().
VOCABULARY - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
VOCABULARY - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
VoidCoder - Class in org.apache.beam.sdk.coders
A Coder for Void.
voids() - Static method in class org.apache.beam.sdk.values.TypeDescriptors
 

W

Wait - Class in org.apache.beam.sdk.transforms
Delays processing of each window in a PCollection until signaled.
Wait() - Constructor for class org.apache.beam.sdk.transforms.Wait
 
Wait.OnSignal<T> - Class in org.apache.beam.sdk.transforms
waitForNMessages(int, Duration) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsub
Repeatedly pull messages from TestPubsub.subscriptionPath(), returns after receiving n messages or after waiting for timeoutDuration.
waitForStart(Duration) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal
Future that waits for a start signal for duration.
waitForSuccess(Duration) - Method in class org.apache.beam.sdk.io.gcp.pubsub.TestPubsubSignal
Wait for a success signal for duration.
waitForUpTo(Duration) - Method in interface org.apache.beam.sdk.io.gcp.pubsub.TestPubsub.PollingAssertion
 
waitUntilFinish(Duration) - Method in class org.apache.beam.runners.apex.ApexRunnerResult
 
waitUntilFinish() - Method in class org.apache.beam.runners.apex.ApexRunnerResult
 
waitUntilFinish() - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
 
waitUntilFinish(Duration) - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
 
waitUntilFinish(Duration, MonitoringUtil.JobMessagesHandler) - Method in class org.apache.beam.runners.dataflow.DataflowPipelineJob
Waits until the pipeline finishes and returns the final status.
waitUntilFinish() - Method in class org.apache.beam.runners.direct.DirectRunner.DirectPipelineResult
Waits until the pipeline finishes and returns the final status.
waitUntilFinish(Duration) - Method in class org.apache.beam.runners.direct.DirectRunner.DirectPipelineResult
Waits until the pipeline finishes and returns the final status.
waitUntilFinish() - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
 
waitUntilFinish(Duration) - Method in class org.apache.beam.runners.flink.FlinkDetachedRunnerResult
 
waitUntilFinish() - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
 
waitUntilFinish(Duration) - Method in class org.apache.beam.runners.flink.FlinkRunnerResult
 
waitUntilFinish(Duration) - Method in class org.apache.beam.runners.gearpump.GearpumpPipelineResult
 
waitUntilFinish() - Method in class org.apache.beam.runners.gearpump.GearpumpPipelineResult
 
waitUntilFinish(Duration) - Method in class org.apache.beam.runners.jet.FailedRunningPipelineResults
 
waitUntilFinish() - Method in class org.apache.beam.runners.jet.FailedRunningPipelineResults
 
waitUntilFinish(Duration) - Method in class org.apache.beam.runners.jet.JetPipelineResult
 
waitUntilFinish() - Method in class org.apache.beam.runners.jet.JetPipelineResult
 
waitUntilFinish() - Method in class org.apache.beam.runners.spark.SparkPipelineResult
 
waitUntilFinish(Duration) - Method in class org.apache.beam.runners.spark.SparkPipelineResult
 
waitUntilFinish() - Method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineResult
 
waitUntilFinish(Duration) - Method in class org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingPipelineResult
 
waitUntilFinish(Duration) - Method in interface org.apache.beam.sdk.PipelineResult
Waits until the pipeline finishes and returns the final status.
waitUntilFinish() - Method in interface org.apache.beam.sdk.PipelineResult
Waits until the pipeline finishes and returns the final status.
WallTime(Instant) - Constructor for class org.apache.beam.sdk.transforms.splittabledofn.WatermarkEstimators.WallTime
 
Watch - Class in org.apache.beam.sdk.transforms
Given a "poll function" that produces a potentially growing set of outputs for an input, this transform simultaneously continuously watches the growth of output sets of all inputs, until a per-input termination condition is reached.
Watch() - Constructor for class org.apache.beam.sdk.transforms.Watch
 
Watch.Growth<InputT,OutputT,KeyT> - Class in org.apache.beam.sdk.transforms
Watch.Growth.PollFn<InputT,OutputT> - Class in org.apache.beam.sdk.transforms
A function that computes the current set of outputs for the given input, in the form of a Watch.Growth.PollResult.
Watch.Growth.PollResult<OutputT> - Class in org.apache.beam.sdk.transforms
The result of a single invocation of a Watch.Growth.PollFn.
Watch.Growth.TerminationCondition<InputT,StateT> - Interface in org.apache.beam.sdk.transforms
A strategy for determining whether it is time to stop polling the current input regardless of whether its output is complete or not.
watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
Like Read#watchForNewFiles.
watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.AvroIO.ParseAll
Deprecated.
Like Read#watchForNewFiles.
watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.AvroIO.Read
Continuously watches for new files matching the filepattern, polling it at the given interval, until the given termination condition is reached.
watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.AvroIO.ReadAll
Deprecated.
Like Read#watchForNewFiles.
watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.TextIO.Read
watchForNewFiles(Duration, Watch.Growth.TerminationCondition<String, ?>) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
Deprecated.
Same as Read#watchForNewFiles(Duration, TerminationCondition).
WatermarkAdvancingStreamingListener() - Constructor for class org.apache.beam.runners.spark.util.GlobalWatermarkHolder.WatermarkAdvancingStreamingListener
 
WatermarkEstimator<WatermarkEstimatorStateT> - Interface in org.apache.beam.sdk.transforms.splittabledofn
A WatermarkEstimator which is used for estimating output watermarks of a splittable DoFn.
WatermarkEstimators - Class in org.apache.beam.sdk.fn.splittabledofn
Support utilties for interacting with WatermarkEstimators.
WatermarkEstimators() - Constructor for class org.apache.beam.sdk.fn.splittabledofn.WatermarkEstimators
 
WatermarkEstimators - Class in org.apache.beam.sdk.transforms.splittabledofn
A set of WatermarkEstimators that users can use to advance the output watermark for their associated splittable DoFns.
WatermarkEstimators.Manual - Class in org.apache.beam.sdk.transforms.splittabledofn
Concrete implementation of a ManualWatermarkEstimator.
WatermarkEstimators.MonotonicallyIncreasing - Class in org.apache.beam.sdk.transforms.splittabledofn
A watermark estimator that observes timestamps of records output from a DoFn reporting the timestamp of the last element seen as the current watermark.
WatermarkEstimators.WallTime - Class in org.apache.beam.sdk.transforms.splittabledofn
A watermark estimator that tracks wall time.
WatermarkEstimators.WatermarkAndStateObserver<WatermarkEstimatorStateT> - Interface in org.apache.beam.sdk.fn.splittabledofn
Interface which allows for accessing the current watermark and watermark estimator state.
WatermarkEvent() - Constructor for class org.apache.beam.sdk.testing.TestStream.WatermarkEvent
 
WatermarkHoldState - Interface in org.apache.beam.sdk.state
For internal use only; no backwards-compatibility guarantees.
WatermarkManager<ExecutableT,CollectionT> - Class in org.apache.beam.runners.direct
Manages watermarks of PCollections and input and output watermarks of AppliedPTransforms to provide event-time and completion tracking for in-memory execution.
WatermarkManager.FiredTimers<ExecutableT> - Class in org.apache.beam.runners.direct
A pair of TimerInternals.TimerData and key which can be delivered to the appropriate AppliedPTransform.
WatermarkManager.TimerUpdate - Class in org.apache.beam.runners.direct
A collection of newly set, deleted, and completed timers.
WatermarkManager.TimerUpdate.TimerUpdateBuilder - Class in org.apache.beam.runners.direct
A WatermarkManager.TimerUpdate builder that needs to be provided with set timers and deleted timers.
WatermarkManager.TransformWatermarks - Class in org.apache.beam.runners.direct
A reference to the input and output watermarks of an AppliedPTransform.
WatermarkParameters - Class in org.apache.beam.sdk.io.kinesis
WatermarkParameters contains the parameters used for watermark computation.
WatermarkParameters() - Constructor for class org.apache.beam.sdk.io.kinesis.WatermarkParameters
 
WatermarkPolicy - Interface in org.apache.beam.sdk.io.kinesis
Implement this interface to define a custom watermark calculation heuristic.
WatermarkPolicyFactory - Interface in org.apache.beam.sdk.io.kinesis
Implement this interface to create a WatermarkPolicy.
WatermarkPolicyFactory.ArrivalTimeWatermarkPolicy - Class in org.apache.beam.sdk.io.kinesis
ArrivalTimeWatermarkPolicy uses WatermarkPolicyFactory.CustomWatermarkPolicy for watermark computation.
WatermarkPolicyFactory.CustomWatermarkPolicy - Class in org.apache.beam.sdk.io.kinesis
CustomWatermarkPolicy uses parameters defined in WatermarkParameters to compute watermarks.
WatermarkPolicyFactory.ProcessingTimeWatermarkPolicy - Class in org.apache.beam.sdk.io.kinesis
Watermark policy where the processing time is used as the event time.
watermarkStateInternal(TimestampCombiner) - Static method in class org.apache.beam.sdk.state.StateSpecs
For internal use only; no backwards-compatibility guarantees.
weeks(int, int) - Static method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows
Returns a WindowFn that windows elements into periods measured by weeks.
where(TypeParameter<X>, TypeDescriptor<X>) - Method in class org.apache.beam.sdk.values.TypeDescriptor
Returns a new TypeDescriptor where the type variable represented by typeParameter are substituted by type.
where(Type, Type) - Method in class org.apache.beam.sdk.values.TypeDescriptor
A more general form of TypeDescriptor.where(TypeParameter, TypeDescriptor) that returns a new TypeDescriptor by matching formal against actual to resolve type variables in the current TypeDescriptor.
whereFieldId(int, SerializableFunction<FieldT, Boolean>) - Method in class org.apache.beam.sdk.schemas.transforms.Filter.Inner
Set a predicate based on the value of a field, where the field is specified by id.
whereFieldIds(List<Integer>, SerializableFunction<Row, Boolean>) - Method in class org.apache.beam.sdk.schemas.transforms.Filter.Inner
Set a predicate based on the value of multipled fields, specified by id.
whereFieldName(String, SerializableFunction<FieldT, Boolean>) - Method in class org.apache.beam.sdk.schemas.transforms.Filter.Inner
Set a predicate based on the value of a field, where the field is specified by name.
whereFieldNames(List<String>, SerializableFunction<Row, Boolean>) - Method in class org.apache.beam.sdk.schemas.transforms.Filter.Inner
Set a predicate based on the value of multipled fields, specified by name.
widening(Schema) - Static method in class org.apache.beam.sdk.schemas.transforms.Cast
 
Widening() - Constructor for class org.apache.beam.sdk.schemas.transforms.Cast.Widening
 
WILDCARD - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
WILDCARD - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
WILDCARD() - Method in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.WildcardContext
 
WildcardContext(FieldSpecifierNotationParser.DotExpressionComponentContext) - Constructor for class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser.WildcardContext
 
wildcardToRegexp(String) - Static method in class org.apache.beam.sdk.extensions.gcp.util.GcsUtil
Expands glob expressions to regular expressions.
WINDMILL_SERVICE_EXPERIMENT - Static variable in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions
Deprecated.
Use STREAMING_ENGINE_EXPERIMENT instead.
window() - Method in interface org.apache.beam.sdk.state.StateContext
Returns the window corresponding to the state.
window() - Method in class org.apache.beam.sdk.transforms.DoFn.OnTimerContext
Returns the window in which the timer is firing.
Window<T> - Class in org.apache.beam.sdk.transforms.windowing
Window logically divides up or groups the elements of a PCollection into finite windows according to a WindowFn.
Window() - Constructor for class org.apache.beam.sdk.transforms.windowing.Window
 
window() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn.AssignContext
Returns the window of the current element prior to this WindowFn being called.
Window.Assign<T> - Class in org.apache.beam.sdk.transforms.windowing
A Primitive PTransform that assigns windows to elements based on a WindowFn.
Window.ClosingBehavior - Enum in org.apache.beam.sdk.transforms.windowing
Specifies the conditions under which a final pane will be created when a window is permanently closed.
Window.OnTimeBehavior - Enum in org.apache.beam.sdk.transforms.windowing
Specifies the conditions under which an on-time pane will be created when a window is closed.
WindowAssignTranslator<T> - Class in org.apache.beam.runners.gearpump.translators
Window.Assign is translated to Gearpump flatMap function.
WindowAssignTranslator() - Constructor for class org.apache.beam.runners.gearpump.translators.WindowAssignTranslator
 
WindowAssignTranslator.AssignWindows<T> - Class in org.apache.beam.runners.gearpump.translators
A Function used internally by Gearpump to wrap the actual Beam's WindowFn.
windowCoder() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.BagUserStateSpec
 
windowCoder() - Method in class org.apache.beam.runners.fnexecution.control.ProcessBundleDescriptors.SideInputSpec
 
windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
 
windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
 
windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
 
windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
 
windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.GlobalWindows
 
windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.InvalidWindows
 
windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.Sessions
 
windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
 
windowCoder() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn
Returns the Coder used for serializing the windows used by this windowFn.
WindowedContext() - Constructor for class org.apache.beam.sdk.transforms.DoFn.WindowedContext
 
windowedFilename(int, int, BoundedWindow, PaneInfo, FileBasedSink.OutputFileHints) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy
 
windowedFilename(int, int, BoundedWindow, PaneInfo, FileBasedSink.OutputFileHints) - Method in class org.apache.beam.sdk.io.FileBasedSink.FilenamePolicy
When a sink has requested windowed or triggered output, this method will be invoked to return the file resource to be created given the base output directory and a FileBasedSink.OutputFileHints containing information about the file, including a suggested extension (e.g.
windowedMultiReceiver(DoFn<?, ?>.WindowedContext, Map<TupleTag<?>, Coder<?>>) - Static method in class org.apache.beam.sdk.transforms.DoFnOutputReceivers
Returns a DoFn.MultiOutputReceiver that delegates to a DoFn.WindowedContext.
windowedMultiReceiver(DoFn<?, ?>.WindowedContext) - Static method in class org.apache.beam.sdk.transforms.DoFnOutputReceivers
Returns a DoFn.MultiOutputReceiver that delegates to a DoFn.WindowedContext.
windowedReceiver(DoFn<?, ?>.WindowedContext, TupleTag<T>) - Static method in class org.apache.beam.sdk.transforms.DoFnOutputReceivers
Returns a DoFn.OutputReceiver that delegates to a DoFn.WindowedContext.
windowedWrites - Variable in class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
Whether windowed writes are being used.
WindowFn<T,W extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
The argument to the Window transform used to assign elements into windows and to determine how windows are merged.
WindowFn() - Constructor for class org.apache.beam.sdk.transforms.windowing.WindowFn
 
WindowFn.AssignContext - Class in org.apache.beam.sdk.transforms.windowing
WindowFn.MergeContext - Class in org.apache.beam.sdk.transforms.windowing
WindowFnTestUtils - Class in org.apache.beam.sdk.testing
A utility class for testing WindowFns.
WindowFnTestUtils() - Constructor for class org.apache.beam.sdk.testing.WindowFnTestUtils
 
WindowGroupP<K,V> - Class in org.apache.beam.runners.jet.processors
Jet Processor implementation for Beam's GroupByKeyOnly + GroupAlsoByWindow primitives.
WINDOWING_STRATEGY - Static variable in class org.apache.beam.runners.dataflow.util.PropertyNames
 
WindowingHelpers - Class in org.apache.beam.runners.spark.structuredstreaming.translation.helpers
Helper functions for working with windows.
WindowingHelpers() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.translation.helpers.WindowingHelpers
 
WindowingStrategy<T,W extends BoundedWindow> - Class in org.apache.beam.sdk.values
A WindowingStrategy describes the windowing behavior for a specific collection of values.
WindowingStrategy.AccumulationMode - Enum in org.apache.beam.sdk.values
The accumulation modes that can be used with windowing.
WindowMappingFn<TargetWindowT extends BoundedWindow> - Class in org.apache.beam.sdk.transforms.windowing
Experimental! This will be ready for users eventually, but should be considered internal for now.
WindowMappingFn() - Constructor for class org.apache.beam.sdk.transforms.windowing.WindowMappingFn
Create a new WindowMappingFn with zero maximum lookback.
WindowMappingFn(Duration) - Constructor for class org.apache.beam.sdk.transforms.windowing.WindowMappingFn
Create a new WindowMappingFn with the specified maximum lookback.
windowOnlyContext(W) - Static method in class org.apache.beam.sdk.state.StateContexts
 
windows() - Static method in class org.apache.beam.sdk.transforms.Reify
Create a PTransform that will reify information from the processing context into instances of ValueInSingleWindow.
windows() - Method in class org.apache.beam.sdk.transforms.windowing.WindowFn.MergeContext
Returns the current set of windows.
windowsInValue() - Static method in class org.apache.beam.sdk.transforms.Reify
Create a PTransform that will output all input KVs with the window pane info inside the value.
WireCoders - Class in org.apache.beam.runners.fnexecution.wire
Helpers to construct coders for gRPC port reads and writes.
with(PCollection<KV<K, V2>>, V1, V2) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join.FullOuterJoin
 
with(PCollection<KV<K, V2>>) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join.InnerJoin
 
with(PCollection<KV<K, V2>>, V2) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join.LeftOuterJoin
 
with(PCollection<KV<K, V2>>, V1) - Static method in class org.apache.beam.sdk.extensions.joinlibrary.Join.RightOuterJoin
 
with(SimpleFunction<DataT, InputT>, Combine.CombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposeCombineFnBuilder
Returns a CombineFns.ComposedCombineFn that can take additional GlobalCombineFns and apply them as a single combine function.
with(SimpleFunction<DataT, InputT>, Coder, Combine.CombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposeCombineFnBuilder
Like #with(SimpleFunction, CombineFn, TupleTag) but with an explicit input coder.
with(SimpleFunction<DataT, InputT>, CombineWithContext.CombineFnWithContext<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposeCombineFnBuilder
Returns a CombineFns.ComposedCombineFnWithContext that can take additional GlobalCombineFns and apply them as a single combine function.
with(SimpleFunction<DataT, InputT>, Coder, CombineWithContext.CombineFnWithContext<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposeCombineFnBuilder
Like #with(SimpleFunction, CombineFnWithContext, TupleTag) but with input coder.
with(SimpleFunction<DataT, InputT>, Combine.CombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
Returns a CombineFns.ComposedCombineFn with an additional Combine.CombineFn.
with(SimpleFunction<DataT, InputT>, CombineWithContext.CombineFnWithContext<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
with(SimpleFunction<DataT, InputT>, Coder, Combine.CombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
Returns a CombineFns.ComposedCombineFn with an additional Combine.CombineFn.
with(SimpleFunction<DataT, InputT>, Coder, CombineWithContext.CombineFnWithContext<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFn
with(SimpleFunction<DataT, InputT>, CombineFnBase.GlobalCombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
with(SimpleFunction<DataT, InputT>, Coder<InputT>, CombineFnBase.GlobalCombineFn<InputT, ?, OutputT>, TupleTag<OutputT>) - Method in class org.apache.beam.sdk.transforms.CombineFns.ComposedCombineFnWithContext
withAccuracy(double, double) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.CountMinSketchFn
Returns a new SketchFrequencies.CountMinSketchFn combiner with new precision accuracy parameters epsilon and confidence.
withAddresses(List<String>) - Method in class org.apache.beam.sdk.io.amqp.AmqpIO.Read
Define the AMQP addresses where to receive messages.
withAllFields() - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
 
withAllowedLateness(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.Window
Override the amount of lateness allowed for data elements in the output PCollection and downstream PCollections until explicitly set again.
withAllowedLateness(Duration, Window.ClosingBehavior) - Method in class org.apache.beam.sdk.transforms.windowing.Window
Override the amount of lateness allowed for data elements in the pipeline.
withAllowedLateness(Duration) - Method in class org.apache.beam.sdk.values.WindowingStrategy
Returns a WindowingStrategy identical to this but with the allowed lateness set to allowedLateness.
withAllowedTimestampSkew(Duration) - Method in class org.apache.beam.sdk.transforms.WithTimestamps
Deprecated.
This method permits a to elements to be emitted behind the watermark. These elements are considered late, and if behind the allowed lateness of a downstream PCollection may be silently dropped. See https://issues.apache.org/jira/browse/BEAM-644 for details on a replacement.
withArrivalTimePolicy() - Static method in interface org.apache.beam.sdk.io.kinesis.WatermarkPolicyFactory
Returns an ArrivalTimeWatermarkPolicy.
withArrivalTimePolicy(Duration) - Static method in interface org.apache.beam.sdk.io.kinesis.WatermarkPolicyFactory
Returns an ArrivalTimeWatermarkPolicy.
withArrivalTimeWatermarkPolicy() - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specifies the WatermarkPolicyFactory as ArrivalTimeWatermarkPolicyFactory.
withArrivalTimeWatermarkPolicy(Duration) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specifies the WatermarkPolicyFactory as ArrivalTimeWatermarkPolicyFactory.
withAuth(String) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
Define the password to authenticate on the Redis server.
withAuth(String) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
 
withAuth(String) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadAll
 
withAuth(String) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
 
withAutoUdfUdafLoad(boolean) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
 
withAvroDataModel(GenericData) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Read
Define the Avro data model; see AvroParquetReader.Builder#withDataModel(GenericData).
withAvroDataModel(GenericData) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.ReadFiles
Define the Avro data model; see AvroParquetReader.Builder#withDataModel(GenericData).
withAvroFormatFunction(SerializableFunction<AvroWriteRequest<T>, GenericRecord>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Formats the user's type into a GenericRecord to be written to BigQuery.
withAvroSchemaFactory(SerializableFunction<TableSchema, Schema>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Uses the specified function to convert a TableSchema to a Schema.
withAwsClientsProvider(AwsClientsProvider) - Method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Read
 
withAwsClientsProvider(String, String, Regions, String) - Method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Read
 
withAwsClientsProvider(String, String, Regions) - Method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Read
 
withAwsClientsProvider(AwsClientsProvider) - Method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Write
 
withAwsClientsProvider(String, String, Regions, String) - Method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Write
 
withAwsClientsProvider(String, String, Regions) - Method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Write
 
withAWSClientsProvider(AwsClientsProvider) - Method in class org.apache.beam.sdk.io.aws.sns.SnsIO.Write
Allows to specify custom AwsClientsProvider.
withAWSClientsProvider(String, String, Regions) - Method in class org.apache.beam.sdk.io.aws.sns.SnsIO.Write
Specify credential details and region to be used to write to SNS.
withAWSClientsProvider(String, String, Regions, String) - Method in class org.apache.beam.sdk.io.aws.sns.SnsIO.Write
Specify credential details and region to be used to write to SNS.
withAWSClientsProvider(AWSClientsProvider) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Allows to specify custom AWSClientsProvider.
withAWSClientsProvider(String, String, Regions) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specify credential details and region to be used to read from Kinesis.
withAWSClientsProvider(String, String, Regions, String) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specify credential details and region to be used to read from Kinesis.
withAWSClientsProvider(AWSClientsProvider) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Write
Allows to specify custom AWSClientsProvider.
withAWSClientsProvider(String, String, Regions) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Write
Specify credential details and region to be used to write to Kinesis.
withAWSClientsProvider(String, String, Regions, String) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Write
Specify credential details and region to be used to write to Kinesis.
withBaseFilename(ResourceId) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
Sets the base filename.
withBaseFilename(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
withBasicCredentials(String, String) - Method in class org.apache.beam.sdk.io.solr.SolrIO.ConnectionConfiguration
If Solr basic authentication is enabled, provide the username and password.
withBatching(boolean) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
If true the uses Cloud Spanner batch API.
withBatching(boolean) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
By default Batch API is used to read data from Cloud Spanner.
withBatchSize(long) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
Provide a size for the scroll read.
withBatchSize(long) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
Sets batch size for the write operation.
withBatchSize(long) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
withBatchSize(long) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
Provide a maximum size in number of SQL statement for the batch.
withBatchSize(long) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
Define the size of the batch to group write operations.
withBatchSize(int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
 
withBatchSize(int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadAll
 
withBatchSizeBytes(long) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the batch size limit (max number of bytes mutated per batch).
withBeamRowConverters(TypeDescriptor<T>, BigQueryIO.TypedRead.ToBeamRowFunction<T>, BigQueryIO.TypedRead.FromBeamRowFunction<T>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
Sets the functions to convert elements to/from Row objects.
withBeamSchemas(boolean) - Method in class org.apache.beam.sdk.io.AvroIO.Read
If set to true, a Beam schema will be inferred from the AVRO schema.
withBeamSchemas(boolean) - Method in class org.apache.beam.sdk.io.AvroIO.ReadAll
Deprecated.
If set to true, a Beam schema will be inferred from the AVRO schema.
withBeamSchemas(boolean) - Method in class org.apache.beam.sdk.io.AvroIO.ReadFiles
If set to true, a Beam schema will be inferred from the AVRO schema.
withBigtableOptions(BigtableOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Deprecated.
will be replaced by bigtable options configurator.
withBigtableOptions(BigtableOptions.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Deprecated.
will be replaced by bigtable options configurator.
withBigtableOptions(BigtableOptions) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
Deprecated.
will be replaced by bigtable options configurator.
withBigtableOptions(BigtableOptions.Builder) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
Deprecated.
will be replaced by bigtable options configurator.
withBigtableOptionsConfigurator(SerializableFunction<BigtableOptions.Builder, BigtableOptions.Builder>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Returns a new BigtableIO.Read that will read from the Cloud Bigtable instance with customized options provided by given configurator.
withBigtableOptionsConfigurator(SerializableFunction<BigtableOptions.Builder, BigtableOptions.Builder>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
Returns a new BigtableIO.Write that will read from the Cloud Bigtable instance with customized options provided by given configurator.
withBootstrapServers(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Sets the bootstrap servers for the Kafka consumer.
withBootstrapServers(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
Wrapper method over KafkaIO.WriteRecords.withBootstrapServers(String), used to keep the compatibility with old API based on KV type of element.
withBootstrapServers(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
Returns a new KafkaIO.Write transform with Kafka producer pointing to bootstrapServers.
withBucket(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
 
withBucket(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
 
withBucketAuto(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
Sets weather to use $bucketAuto or not.
withCharset(Charset) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
Sets the XML file charset.
withCharset(Charset) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
withCharset(Charset) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Sink
 
withCharset(Charset) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
Sets the charset used to write the file.
withChunkSize(Long) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
 
withClientFactory(PubsubClient.PubsubClientFactory) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
The default client to write to Pub/Sub is the PubsubJsonClient, created by the PubsubJsonClient.PubsubJsonClientFactory.
withClientFactory(PubsubClient.PubsubClientFactory) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
The default client to write to Pub/Sub is the PubsubJsonClient, created by the PubsubJsonClient.PubsubJsonClientFactory.
withClientId(String) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
Set up the client ID prefix, which is used to construct a unique client ID.
withClosingBehavior(Window.ClosingBehavior) - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
withClustering(Clustering) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Specifies the clustering fields to use when writing to a single output table.
withClustering() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
withCodec(CodecFactory) - Method in class org.apache.beam.sdk.io.AvroIO.Sink
Specifies to use the given CodecFactory for each generated file.
withCodec(CodecFactory) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Writes to Avro file(s) compressed using specified codec.
withCodec(CodecFactory) - Method in class org.apache.beam.sdk.io.AvroIO.Write
withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
Sets a coder for the result of the parse function.
withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.AvroIO.ParseAll
Deprecated.
Specifies the coder for the result of the parseFn.
withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.AvroIO.ParseFiles
Specifies the coder for the result of the parseFn.
withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Read
 
withCoder(Coder<PublishResult>) - Method in class org.apache.beam.sdk.io.aws.sns.SnsIO.Write
Encode the PublishResult with the given coder.
withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
 
withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the Coder used to serialize the entity in the PCollection.
withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
Sets a Coder for the result of the parse function.
withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
 
withCoder(Coder<OutputT>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
 
withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
 
withCoder(Coder<T>) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
 
withCoder(Coder<T>) - Method in class org.apache.beam.sdk.transforms.Create.TimestampedValues
Returns a Create.TimestampedValues PTransform like this one that uses the given Coder<T> to decode each of the objects into a value of type T.
withCoder(Coder<T>) - Method in class org.apache.beam.sdk.transforms.Create.Values
Returns a Create.Values PTransform like this one that uses the given Coder<T> to decode each of the objects into a value of type T.
withCoderAndParseFn(Coder<T>, SimpleFunction<PubsubMessage, T>) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
Causes the source to return a PubsubMessage that includes Pubsub attributes, and uses the given parsing function to transform the PubsubMessage into an output type.
withCollection(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
Sets the collection to consider in the database.
withCollection(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
Sets the collection where to write data in the database.
withColumns(String...) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
withColumns(List<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
withColumns(String...) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
withColumns(List<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
withCommitDeadline(Duration) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
withCommitDeadline(ValueProvider<Duration>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
withCommitDeadline(Duration) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the deadline for the Commit API call.
withCompletedTimers(Iterable<TimerInternals.TimerData>) - Method in class org.apache.beam.runners.direct.WatermarkManager.TimerUpdate.TimerUpdateBuilder
Adds all of the provided timers to the collection of completed timers, and returns this WatermarkManager.TimerUpdate.TimerUpdateBuilder.
withCompletedTimers(Iterable<TimerInternals.TimerData>) - Method in class org.apache.beam.runners.direct.WatermarkManager.TimerUpdate
Returns a WatermarkManager.TimerUpdate that is like this one, but with the specified completed timers.
withCompression(double) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.GlobalDigest
Sets the compression factor cf.
withCompression(double) - Method in class org.apache.beam.sdk.extensions.sketching.TDigestQuantiles.PerKeyDigest
Sets the compression factor cf.
withCompression(Compression) - Method in class org.apache.beam.sdk.io.CompressedSource
withCompression(Compression) - Method in class org.apache.beam.sdk.io.FileIO.ReadMatches
Reads files using the given Compression.
withCompression(Compression) - Method in class org.apache.beam.sdk.io.FileIO.Write
Specifies to compress all generated shard files using the given Compression and, by default, append the respective extension to the filename.
withCompression(Compression) - Method in class org.apache.beam.sdk.io.TextIO.Read
Reads from input sources using the specified compression type.
withCompression(Compression) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
Deprecated.
Reads from input sources using the specified compression type.
withCompression(Compression) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Returns a transform for writing to text files like this one but that compresses output using the given Compression.
withCompression(Compression) - Method in class org.apache.beam.sdk.io.TextIO.Write
withCompression(Compression) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
Returns a transform for reading TFRecord files that decompresses all input files using the specified compression type.
withCompression(Compression) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
Writes to output files using the specified compression type.
withCompression(Compression) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
Decompresses all input files using the specified compression type.
withCompressionCodec(CompressionCodecName) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Sink
Specifies compression codec.
withCompressionType(TextIO.CompressionType) - Method in class org.apache.beam.sdk.io.TextIO.Read
withCompressionType(TextIO.CompressionType) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
withCompressionType(TFRecordIO.CompressionType) - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
withCompressionType(TFRecordIO.CompressionType) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
withCompressionType(XmlIO.Read.CompressionType) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
withConfidence(double) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.GlobalSketch
Sets the confidence value, i.e.
withConfidence(double) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.PerKeySketch
Sets the confidence value, i.e.
withConfigProperties(Map<String, String>) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
Sets the configuration properties like metastore URI.
withConfigProperties(Map<String, String>) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
Sets the configuration properties like metastore URI.
withConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.FileIO.Match
withConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.FileIO.MatchAll
withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
Reads from the source using the options provided by the given configuration.
withConfiguration(Configuration) - Method in interface org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write.WriteBuilder
Writes to the sink using the options provided by the given hadoop configuration.
withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
Reads from the HBase instance indicated by the* given configuration.
withConfiguration(Configuration) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
Writes to the HBase instance indicated by the* given Configuration.
withConfigurationTransform(PTransform<PCollection<? extends KV<KeyT, ValueT>>, PCollectionView<Configuration>>) - Method in interface org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write.WriteBuilder
Writes to the sink using configuration created by provided configurationTransformation.
withConnectionConfiguration(ElasticsearchIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
Provide the Elasticsearch connection configuration object.
withConnectionConfiguration(ElasticsearchIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
Provide the Elasticsearch connection configuration object.
withConnectionConfiguration(MqttIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Read
Define the MQTT connection configuration used to connect to the MQTT broker.
withConnectionConfiguration(MqttIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Write
Define MQTT connection configuration used to connect to the MQTT broker.
withConnectionConfiguration(RedisConnectionConfiguration) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
 
withConnectionConfiguration(RedisConnectionConfiguration) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadAll
 
withConnectionConfiguration(RedisConnectionConfiguration) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
 
withConnectionConfiguration(SolrIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Read
Provide the Solr connection configuration object.
withConnectionConfiguration(SolrIO.ConnectionConfiguration) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Write
Provide the Solr connection configuration object.
withConnectionFactory(ConnectionFactory) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
Specify the JMS connection factory to connect to the JMS broker.
withConnectionFactory(ConnectionFactory) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
Specify the JMS connection factory to connect to the JMS broker.
withConnectionInitSqls(Collection<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
Sets the connection init sql statements to driver.connect(...).
withConnectionInitSqls(ValueProvider<Collection<String>>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
withConnectionProperties(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
Sets the connection properties passed to driver.connect(...).
withConnectionProperties(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
withConnectTimeout(Integer) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
If set, overwrites the default connect timeout (1000ms) in the RequestConfig of the Elastic RestClient.
withConsistencyLevel(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
 
withConsistencyLevel(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
 
withConsistencyLevel(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
 
withConsistencyLevel(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
 
withConsumerConfigUpdates(Map<String, Object>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Update configuration for the backend main consumer.
withConsumerFactoryFn(SerializableFunction<Map<String, Object>, Consumer<byte[], byte[]>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
A factory to create Kafka Consumer from consumer configuration.
withConsumerFactoryFn(SerializableFunction<Map<String, Object>, ? extends Consumer<?, ?>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
Wrapper method over KafkaIO.WriteRecords.withConsumerFactoryFn(SerializableFunction), used to keep the compatibility with old API based on KV type of element.
withConsumerFactoryFn(SerializableFunction<Map<String, Object>, ? extends Consumer<?, ?>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
When exactly-once semantics are enabled (see KafkaIO.WriteRecords.withEOS(int, String)), the sink needs to fetch previously stored state with Kafka topic.
withContentTypeHint(String) - Method in class org.apache.beam.sdk.io.tika.TikaIO.ParseFiles
Sets a content type hint to make the file parser detection more efficient.
withCreateDisposition(BigQueryIO.Write.CreateDisposition) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Specifies whether the table should be created if it does not exist.
withCreateTime(Duration) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Sets the timestamps policy based on KafkaTimestampType.CREATE_TIME timestamp of the records.
withCreateTime(Duration) - Static method in interface org.apache.beam.sdk.io.kafka.TimestampPolicyFactory
withCustomGcsTempLocation(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Provides a custom location on GCS for storing temporary files to be loaded via BigQuery batch load jobs.
withCustomRateLimitPolicy(RateLimitPolicyFactory) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specifies the RateLimitPolicyFactory for a custom rate limiter.
withCustomWatermarkPolicy(WatermarkPolicyFactory) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specifies the WatermarkPolicyFactory as a custom watermarkPolicyFactory.
withCustomWatermarkPolicy(WatermarkParameters) - Static method in interface org.apache.beam.sdk.io.kinesis.WatermarkPolicyFactory
Returns an custom WatermarkPolicyFactory.
withDatabase(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
Sets the database name.
withDatabase(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
Sets the database name.
withDatabase(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
 
withDatabase(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
 
withDatabase(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
Sets the database to use.
withDatabase(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
Sets the database to use.
withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
Specifies the Cloud Spanner database.
withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
Specifies the Cloud Spanner database.
withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
Specifies the Cloud Spanner database.
withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
Specifies the Cloud Spanner database.
withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
Specifies the Cloud Spanner database.
withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
Specifies the Cloud Spanner database.
withDatabaseId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the Cloud Spanner database.
withDatabaseId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the Cloud Spanner database.
withDatasetService(FakeDatasetService) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
 
withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
 
withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
 
withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
 
withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
See WriteVoid#withDataSourceConfiguration(DataSourceConfiguration).
withDataSourceConfiguration(JdbcIO.DataSourceConfiguration) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
 
withDataSourceProviderFn(SerializableFunction<Void, DataSource>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
 
withDataSourceProviderFn(SerializableFunction<Void, DataSource>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
 
withDataSourceProviderFn(SerializableFunction<Void, DataSource>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
 
withDataSourceProviderFn(SerializableFunction<Void, DataSource>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
withDataSourceProviderFn(SerializableFunction<Void, DataSource>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
 
withDecompression(CompressedSource.DecompressingChannelFactory) - Method in class org.apache.beam.sdk.io.CompressedSource
Return a CompressedSource that is like this one but will decompress its underlying file with the given CompressedSource.DecompressingChannelFactory.
WithDefault() - Constructor for class org.apache.beam.sdk.state.StateSpec.Cases.WithDefault
 
withDefaultTableProvider(String, TableProvider) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
 
withDefaultValue(T) - Method in class org.apache.beam.sdk.transforms.View.AsSingleton
Default value to return for windows with no value in them.
withDelay(Supplier<Duration>) - Static method in interface org.apache.beam.sdk.io.kinesis.RateLimitPolicyFactory
 
withDelimiter(byte[]) - Method in class org.apache.beam.sdk.io.TextIO.Read
Set the custom delimiter to be used in place of the default ones ('\r', '\n' or '\r\n').
withDelimiter(byte[]) - Method in class org.apache.beam.sdk.io.TextIO.ReadFiles
Like Read#withDelimiter.
withDelimiter(char[]) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Specifies the delimiter after each string written.
withDelimiter(char[]) - Method in class org.apache.beam.sdk.io.TextIO.Write
withDescription(String) - Method in class org.apache.beam.sdk.schemas.Schema.Field
Returns a copy of the Field with the description set.
withDestinationCoder(Coder<DestinationT>) - Method in class org.apache.beam.sdk.io.FileIO.Write
Specifies a Coder for the destination type, if it can not be inferred from FileIO.Write.by(org.apache.beam.sdk.transforms.SerializableFunction<UserT, DestinationT>).
withDirectoryTreatment(FileIO.ReadMatches.DirectoryTreatment) - Method in class org.apache.beam.sdk.io.FileIO.ReadMatches
Controls how to handle directories in the input PCollection.
withDuration(Duration) - Method in class org.apache.beam.sdk.transforms.Deduplicate.KeyedValues
Returns a KeyedValues PTransform like this one but with the specified duration.
withDuration(Duration) - Method in class org.apache.beam.sdk.transforms.Deduplicate.Values
Returns a Values PTransform like this one but with the specified duration.
withDuration(Duration) - Method in class org.apache.beam.sdk.transforms.Deduplicate.WithRepresentativeValues
Return a WithRepresentativeValues PTransform that is like this one, but with the specified deduplication duration.
withDynamicDelayRateLimitPolicy(Supplier<Duration>) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specifies a dynamic delay rate limit policy with the given function being called at each polling interval to get the next delay value.
withDynamoDbClientProvider(DynamoDbClientProvider) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
 
withDynamoDbClientProvider(AwsCredentialsProvider, String, URI) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
 
withDynamoDbClientProvider(AwsCredentialsProvider, String) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
 
withDynamoDbClientProvider(DynamoDbClientProvider) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Write
 
withDynamoDbClientProvider(AwsCredentialsProvider, String, URI) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Write
 
withDynamoDbClientProvider(AwsCredentialsProvider, String) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Write
 
withEarlyFirings(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
 
withEarlyFirings(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
Creates a new Trigger like the this, except that it fires repeatedly whenever the given Trigger fires before the watermark has passed the end of the window.
withElementTimestamp() - Static method in interface org.apache.beam.sdk.io.kafka.KafkaPublishTimestampFunction
Returns KafkaPublishTimestampFunction returns element timestamp from ProcessContext.
withEmptyGlobalWindowDestination(DestinationT) - Method in class org.apache.beam.sdk.io.FileIO.Write
If FileIO.Write.withIgnoreWindowing() is specified, specifies a destination to be used in case the collection is empty, to generate the (only, empty) output file.
withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
Like Read#withEmptyMatchTreatment.
withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.AvroIO.ParseAll
Deprecated.
Like Read#withEmptyMatchTreatment.
withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.AvroIO.Read
Configures whether or not a filepattern matching no files is allowed.
withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.AvroIO.ReadAll
Deprecated.
Like Read#withEmptyMatchTreatment.
withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.AvroSource
 
withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.FileIO.Match
withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.FileIO.MatchAll
withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.FileIO.MatchConfiguration
withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.TextIO.Read
withEmptyMatchTreatment(EmptyMatchTreatment) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
Deprecated.
Same as Read#withEmptyMatchTreatment.
withEndKey(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
Returns new ByteKeyRange like this one, but with the specified end key.
withEndpoint(String, int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
 
withEndpoint(String, int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadAll
 
withEndpoint(String, int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
 
withEntity(Class<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the entity class (annotated POJO).
withEntity(Class<T>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
Specify the entity class in the input PCollection.
withEntries - Variable in class org.apache.beam.sdk.extensions.sql.zetasql.QueryTrait
 
withEOS(int, String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
Wrapper method over KafkaIO.WriteRecords.withEOS(int, String), used to keep the compatibility with old API based on KV type of element.
withEOS(int, String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
Provides exactly-once semantics while writing to Kafka, which enables applications with end-to-end exactly-once guarantees on top of exactly-once semantics within Beam pipelines.
withEpsilon(double) - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
Returns an ApproximateQuantilesCombineFn that's like this one except that it uses the specified epsilon value.
withExpireTime(Long) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
 
withExtendedErrorInfo() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Enables extended error information by enabling WriteResult.getFailedInsertsWithErr()
withExtendedErrorInfo(boolean) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StreamingInserts
Specify whether to use extended error info or not.
withExtensionsFrom(Iterable<Class<?>>) - Method in class org.apache.beam.sdk.extensions.protobuf.DynamicProtoCoder
Returns a DynamicProtoCoder like this one, but with the extensions from the given classes registered.
withExtensionsFrom(Iterable<Class<?>>) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
Returns a ProtoCoder like this one, but with the extensions from the given classes registered.
withExtensionsFrom(Class<?>...) - Method in class org.apache.beam.sdk.extensions.protobuf.ProtoCoder
withExternalSorterType(ExternalSorter.Options.SorterType) - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
Sets the external sorter type.
withExternalSynchronization(ExternalSynchronization) - Method in interface org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write.ExternalSynchronizationBuilder
Specifies class which will provide external synchronization required for hadoop write operation.
withFailedInsertRetryPolicy(InsertRetryPolicy) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Specfies a policy for handling failed inserts.
withFailureMode(SpannerIO.FailureMode) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies failure mode.
WithFailures - Class in org.apache.beam.sdk.transforms
A collection of utilities for writing transforms that can handle exceptions raised during processing of elements.
WithFailures() - Constructor for class org.apache.beam.sdk.transforms.WithFailures
 
WithFailures.ExceptionAsMapHandler<T> - Class in org.apache.beam.sdk.transforms
A simple handler that extracts information from an exception to a Map<String, String> and returns a KV where the key is the input element that failed processing, and the value is the map of exception attributes.
WithFailures.ExceptionElement<T> - Class in org.apache.beam.sdk.transforms
The value type passed as input to exception handlers.
WithFailures.Result<OutputT extends POutput,FailureElementT> - Class in org.apache.beam.sdk.transforms
An intermediate output type for PTransforms that allows an output collection to live alongside a collection of elements that failed the transform.
withFanout(int) - Method in class org.apache.beam.sdk.transforms.Combine.Globally
Returns a PTransform identical to this, but that uses an intermediate node to combine parts of the data to reduce load on the final global combine step.
withFetchSize(int) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
This method is used to set the size of the data that is going to be fetched and loaded in memory per every database call.
withFetchSize(int) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
This method is used to set the size of the data that is going to be fetched and loaded in memory per every database call.
withFetchSize(int) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
This method is used to set the size of the data that is going to be fetched and loaded in memory per every database call.
withFieldAccessDescriptors(Map<FieldAccessDescriptor, Object>) - Method in class org.apache.beam.sdk.values.Row.FieldValueBuilder
Sets field values using the FieldAccessDescriptors.
withFieldIds(FieldAccessDescriptor, Integer...) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return a descriptor that accesses the specified field ids as nested subfields of the baseDescriptor.
withFieldIds(FieldAccessDescriptor, Iterable<Integer>) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return a descriptor that accesses the specified field ids as nested subfields of the baseDescriptor.
withFieldIds(Integer...) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return a descriptor that access the specified fields.
withFieldIds(Iterable<Integer>) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return a descriptor that access the specified fields.
withFieldNameAs(String, String) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Add a field with a new name.
withFieldNameAs(String, String) - Method in class org.apache.beam.sdk.schemas.transforms.Select.Fields
Add a single field to the selection, along with the name the field should take in the selected schema.
withFieldNameAs(String, String) - Method in class org.apache.beam.sdk.schemas.transforms.Select.Flattened
Allows renaming a specific nested field.
withFieldNames(String...) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return a descriptor that access the specified fields.
withFieldNames(Iterable<String>) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return a descriptor that access the specified fields.
withFieldNames(FieldAccessDescriptor, String...) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return a descriptor that accesses the specified field names as nested subfields of the baseDescriptor.
withFieldNames(FieldAccessDescriptor, Iterable<String>) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return a descriptor that accesses the specified field names as nested subfields of the baseDescriptor.
withFieldNamesAs(Map<String, String>) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return a descriptor that accesses the specified fields, renaming those fields.
withFieldReordering() - Method in enum org.apache.beam.sdk.extensions.sql.meta.ProjectSupport
 
withFields(FieldAccessDescriptor.FieldDescriptor...) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Returns a FieldAccessDescriptor that accesses the specified fields.
withFields(Iterable<FieldAccessDescriptor.FieldDescriptor>) - Static method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Returns a FieldAccessDescriptor that accesses the specified fields.
withFieldValue(String, Object) - Method in class org.apache.beam.sdk.values.Row.Builder
Set a field value using the field name.
withFieldValue(Integer, Object) - Method in class org.apache.beam.sdk.values.Row.Builder
Set a field value using the field id.
withFieldValue(FieldAccessDescriptor, Object) - Method in class org.apache.beam.sdk.values.Row.Builder
Set a field value using a FieldAccessDescriptor.
withFieldValue(String, Object) - Method in class org.apache.beam.sdk.values.Row.FieldValueBuilder
Set a field value using the field name.
withFieldValue(Integer, Object) - Method in class org.apache.beam.sdk.values.Row.FieldValueBuilder
Set a field value using the field id.
withFieldValue(FieldAccessDescriptor, Object) - Method in class org.apache.beam.sdk.values.Row.FieldValueBuilder
Set a field value using a FieldAccessDescriptor.
withFieldValueGetters(Factory<List<FieldValueGetter>>, Object) - Method in class org.apache.beam.sdk.values.Row.Builder
 
withFieldValues(Map<String, Object>) - Method in class org.apache.beam.sdk.values.Row.Builder
Sets field values using the field names.
withFieldValues(Map<String, Object>) - Method in class org.apache.beam.sdk.values.Row.FieldValueBuilder
Sets field values using the field names.
withFilename(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
 
withFilter(Filter) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
Filters the rows read from HBase using the given* row filter.
withFilter(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
Sets the filter details.
withFilter(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
 
withFilter(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
withFilters(Bson) - Method in class org.apache.beam.sdk.io.mongodb.FindQuery
Sets the filters to find.
withFixedDelay() - Static method in interface org.apache.beam.sdk.io.kinesis.RateLimitPolicyFactory
 
withFixedDelay(Duration) - Static method in interface org.apache.beam.sdk.io.kinesis.RateLimitPolicyFactory
 
withFixedDelayRateLimitPolicy() - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specifies a fixed delay rate limit policy with the default delay of 1 second.
withFixedDelayRateLimitPolicy(Duration) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specifies a fixed delay rate limit policy with the given delay.
withFooter(String) - Method in class org.apache.beam.sdk.io.TextIO.Sink
 
withFooter(String) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Adds a footer string to each file.
withFooter(String) - Method in class org.apache.beam.sdk.io.TextIO.Write
withFormatFunction(SerializableFunction<UserT, OutputT>) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Specifies a format function to convert UserT to the output type.
withFormatFunction(SerializableFunction<T, TableRow>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Formats the user's type into a TableRow to be written to BigQuery.
withFormatFunction(SerializableFunction<UserT, String>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Deprecated.
withFullPublishResult() - Method in class org.apache.beam.sdk.io.aws.sns.SnsIO.Write
Encode the full PublishResult object, including sdkResponseMetadata and sdkHttpMetadata with the HTTP response headers.
withFullPublishResultWithoutHeaders() - Method in class org.apache.beam.sdk.io.aws.sns.SnsIO.Write
Encode the full PublishResult object, including sdkResponseMetadata and sdkHttpMetadata but excluding the HTTP response headers.
withGapDuration(Duration) - Static method in class org.apache.beam.sdk.transforms.windowing.Sessions
Creates a Sessions WindowFn with the specified gap duration.
withGroupingFactor(int) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the multiple of max mutation (in terms of both bytes per batch and cells per batch) that is used to select a set of mutations to sort by key for batching.
withHeader(String) - Method in class org.apache.beam.sdk.io.TextIO.Sink
 
withHeader(String) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Adds a header string to each file.
withHeader(String) - Method in class org.apache.beam.sdk.io.TextIO.Write
withHintMatchesManyFiles() - Method in class org.apache.beam.sdk.io.AvroIO.Parse
Like Read#withHintMatchesManyFiles().
withHintMatchesManyFiles() - Method in class org.apache.beam.sdk.io.AvroIO.Read
Hints that the filepattern specified in AvroIO.Read.from(String) matches a very large number of files.
withHintMatchesManyFiles() - Method in class org.apache.beam.sdk.io.TextIO.Read
Hints that the filepattern specified in TextIO.Read.from(String) matches a very large number of files.
withHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
withHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
Specifies the Cloud Spanner host.
withHost(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
 
withHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
Specifies the Cloud Spanner host.
withHost(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
withHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
Specifies the Cloud Spanner host.
withHost(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
 
withHost(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the Cloud Spanner host.
withHost(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the Cloud Spanner host.
withHost(String) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
Define the host name of the Redis server.
withHosts(List<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the hosts of the Apache Cassandra instances.
withHosts(ValueProvider<List<String>>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the hosts of the Apache Cassandra instances.
withHosts(List<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
Specify the Cassandra instance hosts where to write data.
withHosts(ValueProvider<List<String>>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
Specify the hosts of the Apache Cassandra instances.
withHotKeyFanout(SerializableFunction<? super K, Integer>) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
If a single key has disproportionately many values, it may become a bottleneck, especially in streaming mode.
withHotKeyFanout(int) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
Like Combine.PerKey.withHotKeyFanout(SerializableFunction), but returning the given constant value for every key.
withIdAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
When reading from Cloud Pub/Sub where unique record identifiers are provided as Pub/Sub message attributes, specifies the name of the attribute containing the unique identifier.
withIdAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
Writes to Pub/Sub, adding each record's unique identifier to the published messages in an attribute with the specified name.
withIdFn(ElasticsearchIO.Write.FieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
Provide a function to extract the id from the document.
withIdGenerator(IdGenerator) - Method in class org.apache.beam.runners.fnexecution.control.SdkHarnessClient
 
withIgnoreSSLCertificate(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
Enable ignoreSSLCertificate for ssl for connection (allow for self signed certificates).
withIgnoreSSLCertificate(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
Enable ignoreSSLCertificate for ssl for connection (allow for self signed certificates).
withIgnoreWindowing() - Method in class org.apache.beam.sdk.io.FileIO.Write
Deprecated.
Avoid usage of this method: its effects are complex and it will be removed in future versions of Beam. Right now it exists for compatibility with WriteFiles.
withIndex(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
withIndex(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
withIndexFn(ElasticsearchIO.Write.FieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
Provide a function to extract the target index from the document allowing for dynamic document routing.
withInitialBackoff(Duration) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
Set initial backoff duration.
withInitialPositionInStream(InitialPositionInStream) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specify reading from some initial position in stream.
withInitialTimestampInStream(Instant) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specify reading beginning at given Instant.
withInputMetadata(Metadata) - Method in class org.apache.beam.sdk.io.tika.TikaIO.ParseFiles
Sets the input metadata for Parser.parse(java.io.InputStream, org.xml.sax.ContentHandler, org.apache.tika.metadata.Metadata, org.apache.tika.parser.ParseContext).
withInputTimestamp() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
Wrapper method over KafkaIO.WriteRecords.withInputTimestamp(), used to keep the compatibility with old API based on KV type of element.
withInputTimestamp() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
The timestamp for each record being published is set to timestamp of the element in the pipeline.
withInsertDeduplicate(Boolean) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
For INSERT queries in the replicated table, specifies that deduplication of inserting blocks should be performed.
withInsertDistributedSync(Boolean) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
If setting is enabled, insert query into distributed waits until data will be sent to all nodes in cluster.
withInsertQuorum(Long) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
For INSERT queries in the replicated table, wait writing for the specified number of replicas and linearize the addition of the data.
withInsertRetryPolicy(InsertRetryPolicy) - Method in class org.apache.beam.sdk.io.gcp.bigquery.StreamingInserts
Specify a retry policy for failed inserts.
withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Returns a new BigtableIO.Read that will read from the Cloud Bigtable instance indicated by given parameter, requires BigtableIO.Read.withProjectId(org.apache.beam.sdk.options.ValueProvider<java.lang.String>) to be called to determine the project.
withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Returns a new BigtableIO.Read that will read from the Cloud Bigtable instance indicated by given parameter, requires BigtableIO.Read.withProjectId(org.apache.beam.sdk.options.ValueProvider<java.lang.String>) to be called to determine the project.
withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
Returns a new BigtableIO.Write that will write into the Cloud Bigtable instance indicated by given parameter, requires BigtableIO.Write.withProjectId(org.apache.beam.sdk.options.ValueProvider<java.lang.String>) to be called to determine the project.
withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
Returns a new BigtableIO.Write that will write into the Cloud Bigtable instance indicated by given parameter, requires BigtableIO.Write.withProjectId(org.apache.beam.sdk.options.ValueProvider<java.lang.String>) to be called to determine the project.
withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
Specifies the Cloud Spanner instance.
withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
Specifies the Cloud Spanner instance.
withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
Specifies the Cloud Spanner instance.
withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
Specifies the Cloud Spanner instance.
withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
Specifies the Cloud Spanner instance.
withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
Specifies the Cloud Spanner instance.
withInstanceId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the Cloud Spanner instance.
withInstanceId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the Cloud Spanner instance.
withInterceptors(List<ClientInterceptor>) - Method in class org.apache.beam.sdk.fn.channel.ManagedChannelFactory
Returns a ManagedChannelFactory like this one, but which will apply the provided ClientInterceptors to any channel it creates.
withIsReady(Supplier<Boolean>) - Method in class org.apache.beam.sdk.fn.test.TestStreams.Builder
Returns a new TestStreams.Builder like this one with the specified CallStreamObserver.isReady() callback.
withJobService(BigQueryServices.JobService) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
 
withJsonSchema(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Similar to BigQueryIO.Write.withSchema(TableSchema) but takes in a JSON-serialized TableSchema.
withJsonSchema(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
withJsonTimePartitioning(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
withKeyDeserializer(Class<? extends Deserializer<K>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Sets a Kafka Deserializer to interpret key bytes read from Kafka.
withKeyDeserializer(DeserializerProvider<K>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
 
withKeyDeserializerAndCoder(Class<? extends Deserializer<K>>, Coder<K>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Sets a Kafka Deserializer for interpreting key bytes read from Kafka along with a Coder for helping the Beam runner materialize key objects at runtime if necessary.
withKeyField(String) - Method in class org.apache.beam.sdk.schemas.transforms.CoGroup.Impl
 
withKeyField(String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
 
withKeyField(String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
Set the name of the key field in the resulting schema.
withKeyPattern(String) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
 
withKeyRange(ByteKeyRange) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Returns a new BigtableIO.Read that will read only rows in the specified range.
withKeyRange(ByteKeyRange) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
Reads only rows in the specified range.
withKeyRange(byte[], byte[]) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
Reads only rows in the specified range.
withKeyRanges(List<ByteKeyRange>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Returns a new BigtableIO.Read that will read only rows in the specified ranges.
WithKeys<K,V> - Class in org.apache.beam.sdk.transforms
WithKeys<K, V> takes a PCollection<V>, and either a constant key of type K or a function from V to K, and returns a PCollection<KV<K, V>>, where each of the values in the input PCollection has been paired with either the constant key or a key computed from the value.
withKeySerializer(Class<? extends Serializer<K>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
Wrapper method over KafkaIO.WriteRecords.withKeySerializer(Class), used to keep the compatibility with old API based on KV type of element.
withKeySerializer(Class<? extends Serializer<K>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
Sets a Serializer for serializing key (if any) to bytes.
withKeySet(KeySet) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
withKeySet(KeySet) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
withKeyspace(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the Cassandra keyspace where to read data.
withKeyspace(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the Cassandra keyspace where to read data.
withKeyspace(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
Specify the Cassandra keyspace where to write data.
withKeyspace(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
Specify the Cassandra keyspace where to read data.
withKeystorePassword(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
If Elasticsearch uses SSL/TLS with mutual authentication (via shield), provide the password to open the client keystore.
withKeystorePath(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
If Elasticsearch uses SSL/TLS with mutual authentication (via shield), provide the keystore containing the client key.
withKeyTranslation(SimpleFunction<?, K>) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
Transforms the keys read from the source using the given key translation function.
withKeyType(TypeDescriptor<K>) - Method in class org.apache.beam.sdk.transforms.WithKeys
Return a WithKeys that is like this one with the specified key type descriptor.
withKmsKey(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
For query sources, use this Cloud KMS key to encrypt any temporary tables created.
withKmsKey(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
 
withLabel(String) - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
Set the item label.
withLateFirings(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.AfterWatermarkEarlyAndLate
 
withLateFirings(Trigger.OnceTrigger) - Method in class org.apache.beam.sdk.transforms.windowing.AfterWatermark.FromEndOfWindow
Creates a new Trigger like the this, except that it fires repeatedly whenever the given Trigger fires after the watermark has passed the end of the window.
withLimit(int) - Method in class org.apache.beam.sdk.io.mongodb.FindQuery
Sets the limit of documents to find.
withLinkUrl(String) - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
Set the item link url.
withLiteralGqlQuery(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
Returns a new DatastoreV1.Read that reads the results of the specified GQL query.
withLiteralGqlQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
withLoadJobProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Set the project the BigQuery load job will be initiated from.
withLoadJobProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
 
withLocalDc(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the local DC used for the load balancing.
withLocalDc(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the local DC used for the load balancing.
withLocalDc(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
Specify the local DC used by the load balancing policy.
withLocalDc(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
Specify the local DC used for the load balancing.
withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
Returns a new DatastoreV1.DeleteEntity that deletes entities from the Cloud Datastore Emulator running locally on the specified host port.
withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
Returns a new DatastoreV1.DeleteKey that deletes entities from the Cloud Datastore Emulator running locally on the specified host port.
withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
Returns a new DatastoreV1.Read that reads from a Datastore Emulator running at the given localhost address.
withLocalhost(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
Returns a new DatastoreV1.Write that writes to the Cloud Datastore Emulator running locally on the specified host port.
withLogAppendTime() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
withLogAppendTime() - Static method in interface org.apache.beam.sdk.io.kafka.TimestampPolicyFactory
A TimestampPolicy that assigns Kafka's log append time (server side ingestion time) to each record.
withMapper(ObjectMapper) - Method in class org.apache.beam.sdk.extensions.jackson.AsJsons
Use custom Jackson ObjectMapper instead of the default one.
withMapper(ObjectMapper) - Method in class org.apache.beam.sdk.extensions.jackson.ParseJsons
Use custom Jackson ObjectMapper instead of the default one.
withMapperFactoryFn(SerializableFunction<Session, Mapper>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
A factory to create a specific Mapper for a given Cassandra Session.
withMapperFactoryFn(SerializableFunction<Session, Mapper>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
 
withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.AvroIO.Parse
withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.AvroIO.ParseAll
Deprecated.
withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.AvroIO.Read
withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.AvroIO.ReadAll
Deprecated.
withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.TextIO.Read
withMatchConfiguration(FileIO.MatchConfiguration) - Method in class org.apache.beam.sdk.io.TextIO.ReadAll
Deprecated.
withMaxBatchBytesSize(int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
Writes to Pub/Sub are limited by 10mb in general.
withMaxBatchSize(long) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
Provide a maximum size in number of documents for the batch see bulk API (https://www.elastic.co/guide/en/elasticsearch/reference/2.4/docs-bulk.html).
withMaxBatchSize(int) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
Writes to Pub/Sub are batched to efficiently send data.
withMaxBatchSize(int) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Write
Provide a maximum size in number of documents for the batch.
withMaxBatchSizeBytes(long) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
Provide a maximum size in bytes for the batch see bulk API (https://www.elastic.co/guide/en/elasticsearch/reference/2.4/docs-bulk.html).
withMaxBytesPerPartition(long) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Control how much data will be assigned to a single BigQuery load job.
withMaxCapacityPerShard(Integer) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specifies the maximum number of messages per one shard.
withMaxConnectionIdleTime(int) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
Sets the maximum idle time for a pooled connection.
withMaxConnectionIdleTime(int) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
Sets the maximum idle time for a pooled connection.
withMaxCumulativeBackoff(Duration) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
Limits total time spent in backoff.
withMaxCumulativeBackoff(Duration) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
withMaxCumulativeBackoff(ValueProvider<Duration>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
withMaxCumulativeBackoff(Duration) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the maximum cumulative backoff time when retrying after DEADLINE_EXCEEDED errors.
withMaxFilesPerBundle(int) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Control how many files will be written concurrently by a single worker when using BigQuery load jobs before spilling to a shuffle.
withMaxInputSize(long) - Method in class org.apache.beam.sdk.transforms.ApproximateQuantiles.ApproximateQuantilesCombineFn
Returns an ApproximateQuantilesCombineFn that's like this one except that it uses the specified maxNumElements value.
withMaxInsertBlockSize(long) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
The maximum block size for insertion, if we control the creation of blocks for insertion.
withMaxNumMutations(long) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the cell mutation limit (maximum number of mutated cells per batch).
withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.amqp.AmqpIO.Read
Define the max number of records received by the AmqpIO.Read.
withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.aws.sqs.SqsIO.Read
Define the max number of records received by the SqsIO.Read.
withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Read
Define the max number of records received by the SqsIO.Read.
withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.BoundedReadFromUnboundedSource
Returns a new BoundedReadFromUnboundedSource that reads a bounded amount of data from the given UnboundedSource.
withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
Define the max number of records that the source will read.
withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specifies to read at most a given number of records.
withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Read
Define the max number of records received by the MqttIO.Read.
withMaxNumRecords(long) - Method in class org.apache.beam.sdk.io.Read.Unbounded
Returns a new BoundedReadFromUnboundedSource that reads a bounded amount of data from the given UnboundedSource.
withMaxNumRows(long) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the row mutation limit (maximum number of mutated rows per batch).
withMaxNumWritersPerBundle(int) - Method in class org.apache.beam.sdk.io.WriteFiles
Set the maximum number of writers created in a bundle before spilling to shuffle.
withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.amqp.AmqpIO.Read
Define the max read time (duration) while the AmqpIO.Read will receive messages.
withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.aws.sqs.SqsIO.Read
Define the max read time (duration) while the SqsIO.Read will receive messages.
withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Read
Define the max read time (duration) while the SqsIO.Read will receive messages.
withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.BoundedReadFromUnboundedSource
Returns a new BoundedReadFromUnboundedSource that reads a bounded amount of data from the given UnboundedSource.
withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.GenerateSequence
Specifies to stop generating elements after the given time.
withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
Define the max read time that the source will read.
withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specifies to read records during maxReadTime.
withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Read
Define the max read time (duration) while the MqttIO.Read will receive messages.
withMaxReadTime(Duration) - Method in class org.apache.beam.sdk.io.Read.Unbounded
Returns a new BoundedReadFromUnboundedSource that reads a bounded amount of data from the given UnboundedSource.
withMaxRetries(int) - Method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
Maximum number of retries per insert.
withMemoryMB(int) - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
Sets the size of the memory buffer in megabytes.
withMessageMapper(JmsIO.MessageMapper<T>) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
 
withMetadata(Map<String, Object>) - Method in class org.apache.beam.sdk.io.AvroIO.Sink
Specifies to put the given metadata into each generated file.
withMetadata(Map<String, Object>) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Writes to Avro file(s) with the specified metadata.
withMetadata(Map<String, Object>) - Method in class org.apache.beam.sdk.io.AvroIO.Write
withMetadata() - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
Include metadata in result json documents.
withMetadata(Map<String, byte[]>) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
Deprecated.
use schema options instead.
withMetadata(String, byte[]) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
Deprecated.
use schema options instead.
withMetadata(String, String) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
Deprecated.
use schema options instead.
withMethod(BigQueryIO.TypedRead.Method) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
withMethod(BigQueryIO.Write.Method) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Choose the method used to write data to BigQuery.
withMethod(RedisIO.Write.Method) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
 
WithMetricsSupport - Class in org.apache.beam.runners.spark.metrics
A MetricRegistry decorator-like that supports AggregatorMetric and SparkBeamMetric as Gauges.
WithMetricsSupport - Class in org.apache.beam.runners.spark.structuredstreaming.metrics
A MetricRegistry decorator-like that supports AggregatorMetric and SparkBeamMetric as Gauges.
withMinBundleSize(long) - Method in class org.apache.beam.sdk.io.AvroSource
Sets the minimum bundle size.
withMinBundleSize(long) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
Sets a parameter minBundleSize for the minimum bundle size of the source.
withMinNumberOfSplits(Integer) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
It's possible that system.size_estimates isn't populated or that the number of splits computed by Beam is still to low for Cassandra to handle it.
withMinNumberOfSplits(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
It's possible that system.size_estimates isn't populated or that the number of splits computed by Beam is still to low for Cassandra to handle it.
withMode(WindowingStrategy.AccumulationMode) - Method in class org.apache.beam.sdk.values.WindowingStrategy
Returns a WindowingStrategy identical to this but with the accumulation mode set to mode.
withMongoDbPipeline(List<BsonDocument>) - Method in class org.apache.beam.sdk.io.mongodb.AggregationQuery
 
withName(String) - Method in class org.apache.beam.sdk.schemas.FieldValueTypeInformation
 
withName(String) - Method in class org.apache.beam.sdk.schemas.Schema.Field
Returns a copy of the Field with the name set.
withNamedParameters(Map<String, ?>) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
 
withNamespace(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
Returns a new DatastoreV1.Read that reads from the given namespace.
withNamespace(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
withNamespace(Class<?>) - Method in class org.apache.beam.sdk.transforms.display.DisplayData.ItemSpec
Set the item namespace from the given Class.
withNaming(FileIO.Write.FileNaming) - Method in class org.apache.beam.sdk.io.FileIO.Write
Specifies a custom strategy for generating filenames.
withNaming(SerializableFunction<DestinationT, FileIO.Write.FileNaming>) - Method in class org.apache.beam.sdk.io.FileIO.Write
Specifies a custom strategy for generating filenames depending on the destination, similar to FileIO.Write.withNaming(FileNaming).
withNaming(Contextful<Contextful.Fn<DestinationT, FileIO.Write.FileNaming>>) - Method in class org.apache.beam.sdk.io.FileIO.Write
Like FileIO.Write.withNaming(SerializableFunction) but allows accessing context, such as side inputs, from the function.
withNestedField(int, FieldAccessDescriptor) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return a descriptor that access the specified nested field.
withNestedField(String, FieldAccessDescriptor) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
Return a descriptor that access the specified nested field.
withNestedField(FieldAccessDescriptor.FieldDescriptor, FieldAccessDescriptor) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
 
withNestedFieldAs(String, String, FieldAccessDescriptor) - Method in class org.apache.beam.sdk.schemas.FieldAccessDescriptor
withNoSpilling() - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
withNoSpilling() - Method in class org.apache.beam.sdk.io.FileIO.Write
withNoSpilling() - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
withNoSpilling() - Method in class org.apache.beam.sdk.io.TextIO.Write
withNoSpilling() - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
withNoSpilling() - Method in class org.apache.beam.sdk.io.WriteFiles
Returns a new WriteFiles that writes all data without spilling, simplifying the pipeline.
withNullable(boolean) - Method in class org.apache.beam.sdk.io.clickhouse.TableSchema.ColumnType
 
withNullable(boolean) - Method in class org.apache.beam.sdk.schemas.Schema.Field
Returns a copy of the Field with isNullable set.
withNullable(boolean) - Method in class org.apache.beam.sdk.schemas.Schema.FieldType
 
withNumFileShards(int) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Control how many file shards are written when using BigQuery load jobs.
withNumQuerySplits(int) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
Returns a new DatastoreV1.Read that reads by splitting the given query into numQuerySplits.
withNumShards(int) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Configures the number of output shards produced overall (when using unwindowed writes) or per-window (when using windowed writes).
withNumShards(int) - Method in class org.apache.beam.sdk.io.AvroIO.Write
withNumShards(int) - Method in class org.apache.beam.sdk.io.FileIO.Write
Specifies to use a given fixed number of shards per window.
withNumShards(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.FileIO.Write
withNumShards(int) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Configures the number of output shards produced overall (when using unwindowed writes) or per-window (when using windowed writes).
withNumShards(int) - Method in class org.apache.beam.sdk.io.TextIO.Write
withNumShards(int) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
Writes to the provided number of shards.
withNumShards(int) - Method in class org.apache.beam.sdk.io.WriteFiles
Returns a new WriteFiles that will write to the current FileBasedSink using the specified number of shards.
withNumShards(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.WriteFiles
Returns a new WriteFiles that will write to the current FileBasedSink using the ValueProvider specified number of shards.
withNumSplits(int) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
Sets the user defined number of splits.
withOffset(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.FixedWindows
Partitions the timestamp space into half-open intervals of the form [N * size + offset, (N + 1) * size + offset), where 0 is the epoch.
withOffset(Duration) - Method in class org.apache.beam.sdk.transforms.windowing.SlidingWindows
Assigns timestamps into half-open intervals of the form [N * period + offset, N * period + offset + size).
withOffsetConsumerConfigOverrides(Map<String, Object>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Set additional configuration for the backend offset consumer.
withOnCompleted(Runnable) - Method in class org.apache.beam.sdk.fn.test.TestStreams.Builder
Returns a new TestStreams.Builder like this one with the specified StreamObserver.onCompleted() callback.
withOnError(Runnable) - Method in class org.apache.beam.sdk.fn.test.TestStreams.Builder
Returns a new TestStreams.Builder like this one with the specified StreamObserver.onError(java.lang.Throwable) callback.
withOnError(Consumer<Throwable>) - Method in class org.apache.beam.sdk.fn.test.TestStreams.Builder
Returns a new TestStreams.Builder like this one with the specified StreamObserver.onError(java.lang.Throwable) consumer.
withOnNext(Consumer<T>) - Static method in class org.apache.beam.sdk.fn.test.TestStreams
Creates a test CallStreamObserver TestStreams.Builder that forwards StreamObserver.onNext(V) calls to the supplied Consumer.
withOnTimeBehavior(Window.OnTimeBehavior) - Method in class org.apache.beam.sdk.transforms.windowing.Window
(Experimental) Override the default Window.OnTimeBehavior, to control whether to output an empty on-time pane.
withOnTimeBehavior(Window.OnTimeBehavior) - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
withOptionalParticipation() - Method in class org.apache.beam.sdk.schemas.transforms.CoGroup.By
Means that this field will participate in a join even when not present, similar to SQL outer-join semantics.
withOptions(Schema.Options) - Method in class org.apache.beam.sdk.schemas.Schema.Field
Returns a copy of the Field with the options set.
withOptions(Schema.Options.Builder) - Method in class org.apache.beam.sdk.schemas.Schema.Field
Returns a copy of the Field with the options set.
withOptions(Schema.Options) - Method in class org.apache.beam.sdk.schemas.Schema
Returns a copy of the Schema with the options set.
withOptions(Schema.Options.Builder) - Method in class org.apache.beam.sdk.schemas.Schema
Returns a copy of the Schema with the options set.
withOrdered(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
Enables ordered bulk insertion (default: true).
withoutDefaults() - Method in class org.apache.beam.sdk.transforms.Combine.Globally
Returns a PTransform identical to this, but that does not attempt to provide a default value in the case of empty input.
withoutLimiter() - Static method in interface org.apache.beam.sdk.io.kinesis.RateLimitPolicyFactory
 
withoutMetadata() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Returns a PTransform for PCollection of KV, dropping Kafka metatdata.
withoutPartitioning() - Method in interface org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write.PartitionedWriterBuilder
Writes to the sink without need to partition output into specified number of partitions.
withOutputCoder(Coder<OutputT>) - Method in class org.apache.beam.sdk.transforms.Watch.Growth
Specifies a Coder to use for the outputs.
withOutputFilenames() - Method in class org.apache.beam.sdk.io.AvroIO.Write
Specify that output filenames are wanted.
withOutputFilenames() - Method in class org.apache.beam.sdk.io.TextIO.Write
Specify that output filenames are wanted.
withOutputKeyCoder(Coder<KeyT>) - Method in class org.apache.beam.sdk.transforms.Watch.Growth
Specifies the coder for the output key.
withOutputParallelization(boolean) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
Whether to reshuffle the resulting PCollection so results are distributed to all workers.
withOutputParallelization(boolean) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
Whether to reshuffle the resulting PCollection so results are distributed to all workers.
withOutputParallelization(boolean) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
Whether to reshuffle the resulting PCollection so results are distributed to all workers.
withOutputParallelization(boolean) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
Whether to reshuffle the resulting PCollection so results are distributed to all workers.
withOutputParallelization(boolean) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadAll
Whether to reshuffle the resulting PCollection so results are distributed to all workers.
withOutputs(List<TimestampedValue<OutputT>>) - Method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
Returns a new Watch.Growth.PollResult like this one with the provided outputs.
withOutputSchema(Schema) - Method in class org.apache.beam.sdk.schemas.transforms.Select.Fields
Rename all output fields to match the specified schema.
withOutputSchema(Schema) - Method in class org.apache.beam.sdk.schemas.transforms.Select.Flattened
Rename all output fields to match the specified schema.
withOutputTags(TupleTag<OutputT>, TupleTagList) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
Returns a new multi-output ParDo PTransform that's like this PTransform but with the specified output tags.
withOutputTimestamp(Instant) - Method in interface org.apache.beam.sdk.state.Timer
Sets event time timer's output timestamp.
withoutResultFlattening() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
withoutResultFlattening() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
withoutSharding() - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Forces a single file as output and empty shard name template.
withoutSharding() - Method in class org.apache.beam.sdk.io.AvroIO.Write
withoutSharding() - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Forces a single file as output and empty shard name template.
withoutSharding() - Method in class org.apache.beam.sdk.io.TextIO.Write
withoutSharding() - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
Forces a single file as output.
withoutStrictParsing() - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
During parsing of the arguments, we will skip over improperly formatted and unknown arguments.
withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
Disable validation that the table exists or the query succeeds prior to pipeline submission.
withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Disables BigQuery table validation.
withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Disables validation that the table being read from exists.
withoutValidation() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
Disables validation that the table being written to exists.
withoutValidation() - Method in class org.apache.beam.sdk.io.TFRecordIO.Read
Returns a transform for reading TFRecord files that has GCS path validation on pipeline creation disabled.
withParameterSetter(JdbcIO.PreparedStatementSetter<ParameterT>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
 
withParent(Schema.TypeName) - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold.Context
 
withParseFn(SerializableFunction<GenericRecord, X>, Coder<X>) - Method in class org.apache.beam.sdk.io.AvroSource
Reads GenericRecord of unspecified schema and maps them to instances of a custom type using the given parseFn and encoded using the given coder.
withParser(MongoDbGridFSIO.Parser<X>) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
 
withPartition(Map<String, String>) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
Sets the partition details.
withPartitionCols(List<String>) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
Set the names of the columns that are partitions.
withPartitioner(KinesisPartitioner) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Write
Allows to specify custom implementation of KinesisPartitioner.
withPartitioning() - Method in interface org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Write.PartitionedWriterBuilder
Writes to the sink with partitioning by Task Id.
withPartitionKey(String) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Write
Specify default partition key.
withPartitionOptions(PartitionOptions) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
withPartitionOptions(PartitionOptions) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
withPassword(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the password used for authentication.
withPassword(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the password used for authentication.
withPassword(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
Specify the password used for authentication.
withPassword(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
Specify the password used for authentication.
withPassword(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
If Elasticsearch authentication is enabled, provide the password.
withPassword(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
 
withPassword(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
 
withPassword(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
Define the password to connect to the JMS broker (authenticated).
withPassword(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
Define the password to connect to the JMS broker (authenticated).
withPassword(String) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
 
withPathPart(String) - Method in class org.apache.beam.sdk.schemas.utils.SchemaZipFold.Context
 
withPollingInterval(Duration) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
If specified, polling for new partitions will happen at this periodicity.
withPollInterval(Duration) - Method in class org.apache.beam.sdk.transforms.Watch.Growth
Specifies how long to wait after a call to Watch.Growth.PollFn before calling it again (if at all - according to Watch.Growth.PollResult and the Watch.Growth.TerminationCondition).
withPort(int) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the port number of the Apache Cassandra instances.
withPort(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the port number of the Apache Cassandra instances.
withPort(int) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
Specify the Cassandra instance port number where to write data.
withPort(ValueProvider<Integer>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
Specify the port number of the Apache Cassandra instances.
withPort(int) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
Define the port number of the Redis server.
withPositionalParameters(List<?>) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
 
withPrecision(int) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.ApproximateDistinctFn
Returns an ApproximateDistinct.ApproximateDistinctFn combiner with a new precision p.
withPrecision(int) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.GloballyDistinct
Sets the precision p.
withPrecision(int) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.PerKeyDistinct
Sets the precision p.
withPrecision(int) - Method in class org.apache.beam.sdk.extensions.zetasketch.HllCount.Init.Builder
Explicitly set the precision parameter used to compute HLL++ sketch.
withPrefix(String) - Method in class org.apache.beam.sdk.io.FileIO.Write
Specifies a common prefix to use for all generated filenames, if using the default file naming.
withPrefix(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.FileIO.Write
withPreparedStatementSetter(JdbcIO.PreparedStatementSetter<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
See WriteVoid#withPreparedStatementSetter(PreparedStatementSetter).
withPreparedStatementSetter(JdbcIO.PreparedStatementSetter<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
 
withProcessingTime() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
withProcessingTime() - Static method in interface org.apache.beam.sdk.io.kafka.TimestampPolicyFactory
A TimestampPolicy that assigns processing time to each record.
withProcessingTimePolicy() - Static method in interface org.apache.beam.sdk.io.kinesis.WatermarkPolicyFactory
Returns an ProcessingTimeWatermarkPolicy.
withProcessingTimeWatermarkPolicy() - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specifies the WatermarkPolicyFactory as ProcessingTimeWatermarkPolicyFactory.
withProducerConfigUpdates(Map<String, Object>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
Update configuration for the producer.
withProducerConfigUpdates(Map<String, Object>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
Update configuration for the producer.
withProducerFactoryFn(SerializableFunction<Map<String, Object>, Producer<K, V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
Wrapper method over KafkaIO.WriteRecords.withProducerFactoryFn(SerializableFunction), used to keep the compatibility with old API based on KV type of element.
withProducerFactoryFn(SerializableFunction<Map<String, Object>, Producer<K, V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
Sets a custom function to create Kafka producer.
withProducerProperties(Properties) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Write
Specify the configuration properties for Kinesis Producer Library (KPL).
withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Returns a new BigtableIO.Read that will read from the Cloud Bigtable project indicated by given parameter, requires BigtableIO.Read.withInstanceId(org.apache.beam.sdk.options.ValueProvider<java.lang.String>) to be called to determine the instance.
withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Returns a new BigtableIO.Read that will read from the Cloud Bigtable project indicated by given parameter, requires BigtableIO.Read.withInstanceId(org.apache.beam.sdk.options.ValueProvider<java.lang.String>) to be called to determine the instance.
withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
Returns a new BigtableIO.Write that will write into the Cloud Bigtable project indicated by given parameter, requires BigtableIO.Write.withInstanceId(org.apache.beam.sdk.options.ValueProvider<java.lang.String>) to be called to determine the instance.
withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
Returns a new BigtableIO.Write that will write into the Cloud Bigtable project indicated by given parameter, requires BigtableIO.Write.withInstanceId(org.apache.beam.sdk.options.ValueProvider<java.lang.String>) to be called to determine the instance.
withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
Returns a new DatastoreV1.DeleteEntity that deletes entities from the Cloud Datastore for the specified project.
withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteEntity
withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
Returns a new DatastoreV1.DeleteKey that deletes entities from the Cloud Datastore for the specified project.
withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.DeleteKey
withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
Returns a new DatastoreV1.Read that reads from the Cloud Datastore for the specified project.
withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
Returns a new DatastoreV1.Write that writes to the Cloud Datastore for the specified project.
withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Write
withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerConfig
 
withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
Specifies the Cloud Spanner project.
withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
Specifies the Cloud Spanner project.
withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
Specifies the Cloud Spanner project.
withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
Specifies the Cloud Spanner project.
withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
Specifies the Cloud Spanner project.
withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
Specifies the Cloud Spanner project.
withProjectId(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the Cloud Spanner project.
withProjectId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the Cloud Spanner project.
withProjection(List<String>) - Method in class org.apache.beam.sdk.io.mongodb.FindQuery
Sets the projection.
withProjection(String...) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
withProtocol(TProtocolFactory) - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.ReadFiles
Specifies the TProtocolFactory to be used to decode Thrift objects.
withPublishRequestFn(SerializableFunction<T, PublishRequest>) - Method in class org.apache.beam.sdk.io.aws2.sns.SnsIO.Write
Specify a function for converting a message into PublishRequest object, this function is mandatory.
withPublishTimestampFunction(KafkaPublishTimestampFunction<KV<K, V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
Deprecated.
use KafkaIO.WriteRecords and ProducerRecords to set publish timestamp.
withPublishTimestampFunction(KafkaPublishTimestampFunction<ProducerRecord<K, V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
Deprecated.
use ProducerRecords to set publish timestamp.
withPushedBackTimers(Iterable<TimerInternals.TimerData>) - Method in class org.apache.beam.runners.direct.WatermarkManager.TimerUpdate
Returns a WatermarkManager.TimerUpdate that is like this one, but with the pushedBackTimersare removed set by provided pushedBackTimers.
withQuery(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the query to read data.
withQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the query to read data.
withQuery(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
Provide a query used while reading from Elasticsearch.
withQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
Provide a ValueProvider that provides the query used while reading from Elasticsearch.
withQuery(Query) - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1.Read
Returns a new DatastoreV1.Read that reads the results of the specified query.
withQuery(Statement) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
withQuery(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
withQuery(Statement) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
withQuery(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
withQuery(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
 
withQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
 
withQuery(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
 
withQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
 
withQuery(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
 
withQuery(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
 
withQuery(String) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Read
Provide a query used while reading from Solr.
withQueryFn(SerializableFunction<MongoCollection<Document>, MongoCursor<Document>>) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
Sets a queryFn.
withQueryLocation(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
BigQuery geographic location where the query job will be executed.
withQueryPriority(BigQueryIO.TypedRead.QueryPriority) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
withQueryTempDataset(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
Temporary dataset reference when using BigQueryIO.TypedRead.fromQuery(String).
withQueue(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
Specify the JMS queue destination name where to read messages from.
withQueue(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
Specify the JMS queue destination name where to send messages to.
withQueueUrl(String) - Method in class org.apache.beam.sdk.io.aws.sqs.SqsIO.Read
Define the queueUrl used by the SqsIO.Read to receive messages from SQS.
withQueueUrl(String) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Read
Define the queueUrl used by the SqsIO.Read to receive messages from SQS.
withRate(long, Duration) - Method in class org.apache.beam.sdk.io.GenerateSequence
Specifies to generate at most a given number of elements per a given period.
withReadCommitted() - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Sets "isolation_level" to "read_committed" in Kafka consumer configuration.
withReadOperation(ReadOperation) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
withReadOptions(ReadOptions.TableReadOptions) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
withRecordClass(Class<T>) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
Sets a JAXB annotated class that can be populated using a record of the provided XML file.
withRecordClass(Class<T>) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
withRecordClass(Class<T>) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
Writes objects of the given class mapped to XML elements using JAXB.
withRecordElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
Sets name of the record element of the XML document.
withRecordElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
withRelativeError(double) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.GlobalSketch
Sets the relative error epsilon.
withRelativeError(double) - Method in class org.apache.beam.sdk.extensions.sketching.SketchFrequencies.PerKeySketch
Sets the relative error epsilon.
withRepresentativeCoder(Coder<IdT>) - Method in class org.apache.beam.sdk.transforms.Deduplicate.WithRepresentativeValues
Return a WithRepresentativeValues PTransform that is like this one, but with the specified id type coder.
withRepresentativeType(TypeDescriptor<IdT>) - Method in class org.apache.beam.sdk.transforms.Deduplicate.WithRepresentativeValues
Return a WithRepresentativeValues PTransform that is like this one, but with the specified id type descriptor.
withRepresentativeType(TypeDescriptor<IdT>) - Method in class org.apache.beam.sdk.transforms.Distinct.WithRepresentativeValues
Return a WithRepresentativeValues PTransform that is like this one, but with the specified output type descriptor.
withRepresentativeValueFn(SerializableFunction<T, IdT>) - Static method in class org.apache.beam.sdk.transforms.Deduplicate
Returns a deduplication transform that deduplicates values using the supplied representative value for up to 10 mins within the processing time domain.
withRepresentativeValueFn(SerializableFunction<T, IdT>) - Static method in class org.apache.beam.sdk.transforms.Distinct
Returns a Distinct<T, IdT> PTransform.
withRequestRecordsLimit(int) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specifies the maximum number of records in GetRecordsResult returned by GetRecords call which is limited by 10K records.
withResultOutputTag(TupleTag<PublishResult>) - Method in class org.apache.beam.sdk.io.aws.sns.SnsIO.Write
Tuple tag to store results.
withResults() - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
Returns JdbcIO.WriteVoid transform which can be used in Wait.on(PCollection[]) to wait until all data is written.
withResumeDelay(Duration) - Method in class org.apache.beam.sdk.transforms.DoFn.ProcessContinuation
Builder method to set the value of DoFn.ProcessContinuation.resumeDelay().
withRetained(boolean) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.Write
Whether or not the publish message should be retained by the messaging engine.
withRetryConfiguration(DynamoDBIO.RetryConfiguration) - Method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Write
Provides configuration to retry a failed request to publish a set of records to DynamoDB.
withRetryConfiguration(SnsIO.RetryConfiguration) - Method in class org.apache.beam.sdk.io.aws.sns.SnsIO.Write
Provides configuration to retry a failed request to publish a message to SNS.
withRetryConfiguration(DynamoDBIO.RetryConfiguration) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Write
Provides configuration to retry a failed request to publish a set of records to DynamoDb.
withRetryConfiguration(SnsIO.RetryConfiguration) - Method in class org.apache.beam.sdk.io.aws2.sns.SnsIO.Write
Provides configuration to retry a failed request to publish a message to SNS.
withRetryConfiguration(ElasticsearchIO.RetryConfiguration) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
Provides configuration to retry a failed batch call to Elasticsearch.
withRetryConfiguration(SolrIO.RetryConfiguration) - Method in class org.apache.beam.sdk.io.solr.SolrIO.Write
Provides configuration to retry a failed batch call to Solr.
withRetryStrategy(JdbcIO.RetryStrategy) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
See WriteVoid#withRetryStrategy(RetryStrategy).
withRetryStrategy(JdbcIO.RetryStrategy) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
When a SQL exception occurs, JdbcIO.Write uses this JdbcIO.RetryStrategy to determine if it will retry the statements.
withRootElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
Sets name of the root element of the XML document.
withRootElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
withRootElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Sink
 
withRootElement(String) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Write
Sets the enclosing root element for the generated XML files.
withRowFilter(RowFilter) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Returns a new BigtableIO.Read that will filter the rows read from Cloud Bigtable using the given row filter.
withRowMapper(JdbcIO.RowMapper<T>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
 
withRowMapper(JdbcIO.RowMapper<OutputT>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadAll
 
withRowRestriction(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
withRowRestriction(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
Read only rows which match the specified filter, which must be a SQL expression compatible with Google standard SQL.
withRowSchema(Schema) - Method in class org.apache.beam.sdk.transforms.Create.Values
Returns a Create.Values PTransform like this one that uses the given Schema to represent objects.
withRunnerDeterminedSharding() - Method in class org.apache.beam.sdk.io.WriteFiles
Returns a new WriteFiles that will write to the current FileBasedSink with runner-determined sharding.
withScan(Scan) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
Filters the rows read from HBase using the given* scan.
withScanRequestFn(SerializableFunction<Void, ScanRequest>) - Method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Read
Can't pass ScanRequest object directly from client since this object is not full serializable.
withScanRequestFn(SerializableFunction<Void, ScanRequest>) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
Can't pass ScanRequest object directly from client since this object is not full serializable.
withScanResponseMapperFn(SerializableFunction<ScanResponse, T>) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Read
 
withScanResultMapperFn(SerializableFunction<ScanResult, T>) - Method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Read
 
withSchema(Schema) - Static method in class org.apache.beam.sdk.extensions.sql.meta.provider.mongodb.MongoDbTable.DocumentToRow
 
withSchema(Schema) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Sets the the output schema.
withSchema(Schema) - Method in class org.apache.beam.sdk.io.AvroIO.Write
withSchema(String) - Method in class org.apache.beam.sdk.io.AvroSource
Reads files containing records that conform to the given schema.
withSchema(Schema) - Method in class org.apache.beam.sdk.io.AvroSource
withSchema(Class<X>) - Method in class org.apache.beam.sdk.io.AvroSource
Reads files containing records of the given class.
withSchema(TableSchema) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Uses the specified schema for rows to be written.
withSchema(ValueProvider<TableSchema>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
withSchema(Schema, TypeDescriptor<T>, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Method in class org.apache.beam.sdk.transforms.Create.TimestampedValues
Returns a Create.TimestampedValues PTransform like this one that uses the given Schema to represent objects.
withSchema(Schema, TypeDescriptor<T>, SerializableFunction<T, Row>, SerializableFunction<Row, T>) - Method in class org.apache.beam.sdk.transforms.Create.Values
Returns a Create.Values PTransform like this one that uses the given Schema to represent objects.
withSchema(Schema) - Static method in class org.apache.beam.sdk.transforms.JsonToRow
 
withSchema(Schema) - Static method in class org.apache.beam.sdk.values.Row
Creates a row builder with specified Row.getSchema().
withSchemaFromView(PCollectionView<Map<String, String>>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Allows the schemas for each table to be computed within the pipeline itself.
withSchemaReadySignal(PCollection) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies an optional input PCollection that can be used as the signal for Wait.OnSignal to indicate when the database schema is ready to be read.
withSchemaUpdateOptions(Set<BigQueryIO.Write.SchemaUpdateOption>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Allows the schema of the destination table to be updated as a side effect of the write.
withScrollKeepalive(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read
Provide a scroll keepalive.
withSelectedFields(List<String>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
withSelectedFields(ValueProvider<List<String>>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
Read only the specified fields (columns) from a BigQuery table.
withServerUri(String) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
Set up the MQTT broker URI.
withShard(int) - Method in class org.apache.beam.sdk.io.FileBasedSink.FileResult
 
withSharding(PTransform<PCollection<UserT>, PCollectionView<Integer>>) - Method in class org.apache.beam.sdk.io.FileIO.Write
Specifies a PTransform to use for computing the desired number of shards in each window.
withSharding(PTransform<PCollection<UserT>, PCollectionView<Integer>>) - Method in class org.apache.beam.sdk.io.WriteFiles
Returns a new WriteFiles that will write to the current FileBasedSink using the specified PTransform to compute the number of shards.
withShardingFunction(ShardingFunction<UserT, DestinationT>) - Method in class org.apache.beam.sdk.io.WriteFiles
Returns a new WriteFiles that will write to the current FileBasedSink using the specified sharding function to assign shard for inputs.
withShardNameTemplate(String) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Uses the given ShardNameTemplate for naming output files.
withShardNameTemplate(String) - Method in class org.apache.beam.sdk.io.AvroIO.Write
withShardNameTemplate(String) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Uses the given ShardNameTemplate for naming output files.
withShardNameTemplate(String) - Method in class org.apache.beam.sdk.io.TextIO.Write
withShardNameTemplate(String) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
Uses the given shard name template.
withShardTemplate(String) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
Sets the shard template.
withSideInput() - Method in class org.apache.beam.sdk.schemas.transforms.CoGroup.By
 
withSideInput(String, PCollectionView<?>) - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
Returns a new multi-output ParDo PTransform that's like this PTransform but with the specified additional side inputs.
withSideInput(String, PCollectionView<?>) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
Returns a new ParDo PTransform that's like this PTransform but with the specified additional side inputs.
withSideInputs(List<PCollectionView<?>>) - Method in class org.apache.beam.sdk.io.WriteFiles
 
withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.Combine.Globally
Returns a PTransform identical to this, but with the specified side inputs to use in CombineWithContext.CombineFnWithContext.
withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.Combine.Globally
Returns a PTransform identical to this, but with the specified side inputs to use in CombineWithContext.CombineFnWithContext.
withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
 
withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.Combine.GroupedValues
 
withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
Returns a PTransform identical to this, but with the specified side inputs to use in CombineWithContext.CombineFnWithContext.
withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.Combine.PerKey
Returns a PTransform identical to this, but with the specified side inputs to use in CombineWithContext.CombineFnWithContext.
withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
Returns a new multi-output ParDo PTransform that's like this PTransform but with the specified additional side inputs.
withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
 
withSideInputs(Map<String, PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.ParDo.MultiOutput
Returns a new multi-output ParDo PTransform that's like this PTransform but with the specified additional side inputs.
withSideInputs(PCollectionView<?>...) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
Returns a new ParDo PTransform that's like this PTransform but with the specified additional side inputs.
withSideInputs(Iterable<? extends PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
Returns a new ParDo PTransform that's like this PTransform but with the specified additional side inputs.
withSideInputs(Map<String, PCollectionView<?>>) - Method in class org.apache.beam.sdk.transforms.ParDo.SingleOutput
Returns a new ParDo PTransform that's like this PTransform but with the specified additional side inputs.
withSideInputStream(TranslationContext, JavaStream<WindowedValue<InputT>>, Map<String, PCollectionView<?>>) - Static method in class org.apache.beam.runners.gearpump.translators.utils.TranslatorUtils
 
withSingletonValues() - Method in class org.apache.beam.sdk.transforms.View.AsMap
Deprecated.
this method simply returns this AsMap unmodified
withSize(int) - Static method in class org.apache.beam.sdk.schemas.utils.AvroUtils.FixedBytesField
Create a AvroUtils.FixedBytesField with the specified size.
withSkew(Duration) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
 
withSnsClientProvider(SnsClientProvider) - Method in class org.apache.beam.sdk.io.aws2.sns.SnsIO.Write
Allows to specify custom SnsClientProvider.
withSnsClientProvider(AwsCredentialsProvider, String) - Method in class org.apache.beam.sdk.io.aws2.sns.SnsIO.Write
Specify AwsCredentialsProvider and region to be used to write to SNS.
withSnsClientProvider(AwsCredentialsProvider, String, URI) - Method in class org.apache.beam.sdk.io.aws2.sns.SnsIO.Write
Specify AwsCredentialsProvider and region to be used to write to SNS.
withSocketTimeout(Integer) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
If set, overwrites the default max retry timeout (30000ms) in the Elastic RestClient and the default socket timeout (30000ms) in the RequestConfig of the Elastic RestClient.
withSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
Specifies the Cloud Spanner configuration.
withSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
Specifies the Cloud Spanner configuration.
withSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
Specifies the Cloud Spanner configuration.
withSpannerConfig(SpannerConfig) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
Specifies the Cloud Spanner configuration.
withSparsePrecision(int) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.GloballyDistinct
Sets the sparse representation's precision sp.
withSparsePrecision(int) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.PerKeyDistinct
Sets the sparse representation's precision sp.
withSparseRepresentation(int) - Method in class org.apache.beam.sdk.extensions.sketching.ApproximateDistinct.ApproximateDistinctFn
Returns an ApproximateDistinct.ApproximateDistinctFn combiner with a new sparse representation's precision sp.
withSqsClientProvider(SqsClientProvider) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Read
Allows to specify custom SqsClientProvider.
withSqsClientProvider(AwsCredentialsProvider, String) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Read
Specify AwsCredentialsProvider and region to be used to read from SQS.
withSqsClientProvider(AwsCredentialsProvider, String, URI) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Read
Specify AwsCredentialsProvider and region to be used to write to SQS.
withSqsClientProvider(SqsClientProvider) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Write
Allows to specify custom SqsClientProvider.
withSqsClientProvider(AwsCredentialsProvider, String) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Write
Specify AwsCredentialsProvider and region to be used to write to SQS.
withSqsClientProvider(AwsCredentialsProvider, String, URI) - Method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Write
Specify AwsCredentialsProvider and region to be used to write to SQS.
withSSLEnabled(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
Enable ssl for connection.
withSSLEnabled(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
Enable ssl for connection.
withSSLInvalidHostNameAllowed(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
Enable invalidHostNameAllowed for ssl for connection.
withSSLInvalidHostNameAllowed(boolean) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
Enable invalidHostNameAllowed for ssl for connection.
withStartingDay(int, int, int) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
 
withStartingMonth(int, int) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
 
withStartingYear(int) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
 
withStartKey(ByteKey) - Method in class org.apache.beam.sdk.io.range.ByteKeyRange
Returns new ByteKeyRange like this one, but with the specified start key.
withStartReadTime(Instant) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Use timestamp to set up start offset.
withStatement(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
withStatement(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
 
withStatement(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
 
withStatementPreparator(JdbcIO.StatementPreparator) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Read
 
withStatementPreparator(JdbcIO.StatementPreparator) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.ReadRows
 
withStorageClient(BigQueryServices.StorageClient) - Method in class org.apache.beam.sdk.io.gcp.testing.FakeBigQueryServices
 
withStreamName(String) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specify reading from streamName.
withStreamName(String) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Write
Specify Kinesis stream name which will be used for writing, this name is required.
withSuffix(String) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Configures the filename suffix for written files.
withSuffix(String) - Method in class org.apache.beam.sdk.io.AvroIO.Write
withSuffix(String) - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
Sets the suffix.
withSuffix(String) - Method in class org.apache.beam.sdk.io.FileIO.Write
Specifies a common suffix to use for all generated filenames, if using the default file naming.
withSuffix(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.FileIO.Write
withSuffix(String) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Configures the filename suffix for written files.
withSuffix(String) - Method in class org.apache.beam.sdk.io.TextIO.Write
withSuffix(String) - Method in class org.apache.beam.sdk.io.TFRecordIO.Write
Writes to the file(s) with the given filename suffix.
withTable(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the Cassandra table where to read data.
withTable(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the Cassandra table where to read data.
withTable(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.ReadOperation
 
withTable(String) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
withTable(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
Sets the table name to read from.
withTable(String) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
Sets the table name to write to, the table should exist beforehand.
withTable(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.Write
withTable(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
 
withTableDescription(String) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Specifies the table description.
withTableId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Returns a new BigtableIO.Read that will read from the specified table.
withTableId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Read
Returns a new BigtableIO.Read that will read from the specified table.
withTableId(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
Returns a new BigtableIO.Write that will write to the specified table.
withTableId(String) - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
Returns a new BigtableIO.Write that will write to the specified table.
withTableId(String) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Read
Reads from the specified table.
withTableId(String) - Method in class org.apache.beam.sdk.io.hbase.HBaseIO.Write
Writes to the specified table.
withTableProvider(TableProvider) - Static method in class org.apache.beam.sdk.extensions.sql.impl.BeamSqlEnv
This method creates BeamSqlEnv using empty Pipeline Options.
withTableProvider(String, TableProvider) - Method in class org.apache.beam.sdk.extensions.sql.SqlTransform
 
withTableReference(TableReference) - Method in class org.apache.beam.sdk.io.gcp.bigquery.TableDestination
 
withTempDirectory(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Set the base directory used to generate temporary files.
withTempDirectory(ResourceId) - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Set the base directory used to generate temporary files.
withTempDirectory(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.AvroIO.Write
withTempDirectory(ResourceId) - Method in class org.apache.beam.sdk.io.AvroIO.Write
withTempDirectory(String) - Method in class org.apache.beam.sdk.io.FileIO.Write
Specifies a directory into which all temporary files will be placed.
withTempDirectory(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.FileIO.Write
withTempDirectory(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Set the base directory used to generate temporary files.
withTempDirectory(ResourceId) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Set the base directory used to generate temporary files.
withTempDirectory(ValueProvider<ResourceId>) - Method in class org.apache.beam.sdk.io.TextIO.Write
withTempDirectory(ResourceId) - Method in class org.apache.beam.sdk.io.TextIO.Write
withTemplateCompatibility() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Read
Use new template-compatible source implementation.
withTemplateCompatibility() - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.TypedRead
 
withTempLocation(String) - Method in class org.apache.beam.sdk.extensions.sorter.BufferedExternalSorter.Options
Sets the path to a temporary location where the sorter writes intermediate files.
withTerminationCondition(Watch.Growth.TerminationCondition<HCatalogIO.Read, ?>) - Method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Read
If specified, the poll function will stop polling after the termination condition has been satisfied.
withTerminationPerInput(Watch.Growth.TerminationCondition<InputT, ?>) - Method in class org.apache.beam.sdk.transforms.Watch.Growth
Specifies a Watch.Growth.TerminationCondition that will be independently used for every input.
withTestServices(BigQueryServices) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
 
withTikaConfigPath(String) - Method in class org.apache.beam.sdk.io.tika.TikaIO.ParseFiles
withTikaConfigPath(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.tika.TikaIO.ParseFiles
Like with(tikaConfigPath).
withTimeDomain(TimeDomain) - Method in class org.apache.beam.sdk.transforms.Deduplicate.KeyedValues
Returns a KeyedValues PTransform like this one but with the specified time domain.
withTimeDomain(TimeDomain) - Method in class org.apache.beam.sdk.transforms.Deduplicate.Values
Returns a Values PTransform like this one but with the specified time domain.
withTimeDomain(TimeDomain) - Method in class org.apache.beam.sdk.transforms.Deduplicate.WithRepresentativeValues
Returns a WithRepresentativeValues PTransform like this one but with the specified time domain.
withTimeout(int) - Method in class org.apache.beam.sdk.io.redis.RedisConnectionConfiguration
Define the Redis connection timeout.
withTimeout(int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Read
 
withTimeout(int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.ReadAll
 
withTimeout(int) - Method in class org.apache.beam.sdk.io.redis.RedisIO.Write
 
withTimePartitioning(TimePartitioning) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Allows newly created tables to include a TimePartitioning class.
withTimePartitioning(ValueProvider<TimePartitioning>) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
withTimestamp(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
withTimestamp(Timestamp) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
 
withTimestampAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Read
When reading from Cloud Pub/Sub where record timestamps are provided as Pub/Sub message attributes, specifies the name of the attribute that contains the timestamp.
withTimestampAttribute(String) - Method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
Writes to Pub/Sub and adds each record's timestamp to the published messages in an attribute with the specified name.
withTimestampBound(TimestampBound) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.CreateTransaction
 
withTimestampBound(TimestampBound) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
withTimestampBound(TimestampBound) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
 
withTimestampCombiner(TimestampCombiner) - Method in class org.apache.beam.sdk.transforms.windowing.Window
(Experimental) Override the default TimestampCombiner, to control the output timestamp of values output from a GroupByKey operation.
withTimestampCombiner(TimestampCombiner) - Method in class org.apache.beam.sdk.values.WindowingStrategy
 
withTimestampFn(SerializableFunction<Long, Instant>) - Method in class org.apache.beam.sdk.io.GenerateSequence
Specifies the function to use to assign timestamps to the elements.
withTimestampFn(SerializableFunction<KV<K, V>, Instant>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Deprecated.
withTimestampFn(SerializableFunction<KafkaRecord<K, V>, Instant>) - Static method in interface org.apache.beam.sdk.io.kafka.TimestampPolicyFactory
withTimestampFn(SerializableFunction<KinesisRecord, Instant>) - Method in class org.apache.beam.sdk.io.kinesis.WatermarkParameters
Specify the SerializableFunction to extract the event time from a KinesisRecord.
withTimestampFn2(SerializableFunction<KafkaRecord<K, V>, Instant>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Deprecated.
withTimestampPolicyFactory(TimestampPolicyFactory<K, V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Provide custom TimestampPolicyFactory to set event times and watermark for each partition.
WithTimestamps<T> - Class in org.apache.beam.sdk.transforms
A PTransform for assigning timestamps to all the elements of a PCollection.
withTimeZone(DateTimeZone) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.DaysWindows
 
withTimeZone(DateTimeZone) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.MonthsWindows
 
withTimeZone(DateTimeZone) - Method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows.YearsWindows
 
withTopic(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
Specify the JMS topic destination name where to receive messages from.
withTopic(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
Specify the JMS topic destination name where to send messages to.
withTopic(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Sets the topic to read from.
withTopic(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
Wrapper method over KafkaIO.WriteRecords.withTopic(String), used to keep the compatibility with old API based on KV type of element.
withTopic(String) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
Sets the default Kafka topic to write to.
withTopic(String) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
Set up the MQTT getTopic pattern.
withTopicArn(String) - Method in class org.apache.beam.sdk.io.aws2.sns.SnsIO.Write
Specify the SNS topic which will be used for writing, this name is mandatory.
withTopicName(String) - Method in class org.apache.beam.sdk.io.aws.sns.SnsIO.Write
Specify the SNS topic which will be used for writing, this name is mandatory.
withTopicPartitions(List<TopicPartition>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Sets a list of partitions to read from.
withTopics(List<String>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Sets a list of topics to read from.
withTransaction(PCollectionView<Transaction>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Read
 
withTransaction(PCollectionView<Transaction>) - Method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.ReadAll
 
withTrigger(Trigger) - Method in class org.apache.beam.sdk.values.WindowingStrategy
Returns a WindowingStrategy identical to this but with the trigger set to wildcardTrigger.
withTriggeringFrequency(Duration) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Choose the frequency at which file writes are triggered.
withTrustSelfSignedCerts(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
If Elasticsearch uses SSL/TLS then configure whether to trust self signed certs or not.
withType(Schema.FieldType) - Method in class org.apache.beam.sdk.schemas.Schema.Field
Returns a copy of the Field with the Schema.FieldType set.
withType(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.transforms.Create.TimestampedValues
Returns a Create.TimestampedValues PTransform like this one that uses the given TypeDescriptor<T> to determine the Coder to use to decode each of the objects into a value of type T.
withType(TypeDescriptor<T>) - Method in class org.apache.beam.sdk.transforms.Create.Values
Returns a Create.Values PTransform like this one that uses the given TypeDescriptor<T> to determine the Coder to use to decode each of the objects into a value of type T.
withTypeFn(ElasticsearchIO.Write.FieldValueExtractFn) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
Provide a function to extract the target type from the document allowing for dynamic document routing.
withUpToDateThreshold(Duration) - Method in class org.apache.beam.sdk.io.kinesis.KinesisIO.Read
Specifies how late records consumed by this source can be to still be considered on time.
withUri(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Read
 
withUri(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
 
withUri(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Read
Define the location of the MongoDB instances using an URI.
withUri(String) - Method in class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
Define the location of the MongoDB instances using an URI.
withUsePartialUpdate(boolean) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
Provide an instruction to control whether partial updates or inserts (default) are issued to Elasticsearch.
withUsername(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the username for authentication.
withUsername(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Read
Specify the username for authentication.
withUsername(String) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
Specify the username used for authentication.
withUsername(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
Specify the username for authentication.
withUsername(String) - Method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration
If Elasticsearch authentication is enabled, provide the username.
withUsername(String) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
 
withUsername(ValueProvider<String>) - Method in class org.apache.beam.sdk.io.jdbc.JdbcIO.DataSourceConfiguration
 
withUsername(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Read
Define the username to connect to the JMS broker (authenticated).
withUsername(String) - Method in class org.apache.beam.sdk.io.jms.JmsIO.Write
Define the username to connect to the JMS broker (authenticated).
withUsername(String) - Method in class org.apache.beam.sdk.io.mqtt.MqttIO.ConnectionConfiguration
 
withValidation() - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory.Builder
After creation we will validate that PipelineOptions conforms to all the validation criteria from <T>.
withValidation() - Method in class org.apache.beam.sdk.options.PipelineOptionsFactory
After creation we will validate that <T> conforms to all the validation criteria.
withValidationEventHandler(ValidationEventHandler) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Read
Sets the ValidationEventHandler to use with JAXB.
withValidationEventHandler(ValidationEventHandler) - Method in class org.apache.beam.sdk.io.xml.XmlIO.ReadFiles
withValueDeserializer(Class<? extends Deserializer<V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Sets a Kafka Deserializer to interpret value bytes read from Kafka.
withValueDeserializer(DeserializerProvider<V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
 
withValueDeserializerAndCoder(Class<? extends Deserializer<V>>, Coder<V>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Sets a Kafka Deserializer for interpreting value bytes read from Kafka along with a Coder for helping the Beam runner materialize value objects at runtime if necessary.
withValueField(String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.ByFields
 
withValueSerializer(Class<? extends Serializer<V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Write
Wrapper method over KafkaIO.WriteRecords.withValueSerializer(Class), used to keep the compatibility with old API based on KV type of element.
withValueSerializer(Class<? extends Serializer<V>>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
Sets a Serializer for serializing value to bytes.
withValueTranslation(SimpleFunction<?, V>) - Method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO.Read
Transforms the values read from the source using the given value translation function.
withWatermark(Instant) - Method in class org.apache.beam.sdk.transforms.Watch.Growth.PollResult
Returns a new Watch.Growth.PollResult like this one with the provided watermark.
withWatermarkFn(SerializableFunction<KV<K, V>, Instant>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Deprecated.
withWatermarkFn2(SerializableFunction<KafkaRecord<K, V>, Instant>) - Method in class org.apache.beam.sdk.io.kafka.KafkaIO.Read
Deprecated.
withWatermarkIdleDurationThreshold(Duration) - Method in class org.apache.beam.sdk.io.kinesis.WatermarkParameters
Specify the watermark idle duration to consider before advancing the watermark.
withWindowedWrites() - Method in class org.apache.beam.sdk.io.AvroIO.TypedWrite
Preserves windowing of input elements and writes them to files based on the element's window.
withWindowedWrites() - Method in class org.apache.beam.sdk.io.AvroIO.Write
withWindowedWrites() - Method in class org.apache.beam.sdk.io.DefaultFilenamePolicy.Params
Specify that writes are windowed.
withWindowedWrites() - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Preserves windowing of input elements and writes them to files based on the element's window.
withWindowedWrites() - Method in class org.apache.beam.sdk.io.TextIO.Write
withWindowedWrites() - Method in class org.apache.beam.sdk.io.WriteFiles
Returns a new WriteFiles that writes preserves windowing on it's input.
withWindowFn(WindowFn<?, ?>) - Method in class org.apache.beam.sdk.values.WindowingStrategy
Returns a WindowingStrategy identical to this but with the window function set to wildcardWindowFn.
withWritableByteChannelFactory(FileBasedSink.WritableByteChannelFactory) - Method in class org.apache.beam.sdk.io.TextIO.TypedWrite
Returns a transform for writing to text files like this one but that has the given FileBasedSink.WritableByteChannelFactory to be used by the FileBasedSink during output.
withWritableByteChannelFactory(FileBasedSink.WritableByteChannelFactory) - Method in class org.apache.beam.sdk.io.TextIO.Write
See TypedWrite#withWritableByteChannelFactory(WritableByteChannelFactory).
withWriteDisposition(BigQueryIO.Write.WriteDisposition) - Method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
Specifies what to do with existing data in the table, in case the table already exists.
withWriteRequestMapperFn(SerializableFunction<T, KV<String, WriteRequest>>) - Method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Write
 
withWriteRequestMapperFn(SerializableFunction<T, KV<String, WriteRequest>>) - Method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Write
 
withWriteResults() - Method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
Returns a BigtableIO.WriteWithResults that will emit a BigtableWriteResult for each batch of rows written.
witValueField(String) - Method in class org.apache.beam.sdk.schemas.transforms.Group.CombineFieldsByFields
Set the name of the value field in the resulting schema.
WordCount - Class in org.apache.beam.runners.spark.structuredstreaming.examples
Duplicated from beam-examples-java to avoid dependency.
WordCount() - Constructor for class org.apache.beam.runners.spark.structuredstreaming.examples.WordCount
 
WordCount.CountWords - Class in org.apache.beam.runners.spark.structuredstreaming.examples
A PTransform that converts a PCollection containing lines of text into a PCollection of formatted word counts.
WordCount.FormatAsTextFn - Class in org.apache.beam.runners.spark.structuredstreaming.examples
A SimpleFunction that converts a Word and Count into a printable string.
WordCount.WordCountOptions - Interface in org.apache.beam.runners.spark.structuredstreaming.examples
Options supported by WordCount.
WorkerHarnessContainerImageFactory() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowPipelineWorkerPoolOptions.WorkerHarnessContainerImageFactory
 
WorkerLogLevelOverrides() - Constructor for class org.apache.beam.runners.dataflow.options.DataflowWorkerLoggingOptions.WorkerLogLevelOverrides
Deprecated.
 
workerStatus(StreamObserver<BeamFnApi.WorkerStatusRequest>) - Method in class org.apache.beam.runners.fnexecution.status.BeamWorkerStatusGrpcService
 
wrap(Throwable) - Method in class org.apache.beam.sdk.testing.PAssert.PAssertionSite
 
wrap(String) - Method in class org.apache.beam.sdk.testing.PAssert.PAssertionSite
 
WrappedList(List<Object>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedList
 
WrappedMap(Map<Object, Object>) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedMap
 
WrappedRow(Row) - Constructor for class org.apache.beam.sdk.extensions.sql.impl.rel.BeamCalcRel.WrappedRow
 
wrapping(StreamObserver<V>) - Static method in class org.apache.beam.sdk.fn.stream.SynchronizedStreamObserver
Create a new SynchronizedStreamObserver which will delegate all calls to the underlying StreamObserver, synchronizing access to that observer.
wrapProcessContext(DoFn<InputT, ?>.ProcessContext) - Static method in class org.apache.beam.sdk.transforms.Contextful.Fn.Context
Convenience wrapper for creating a Contextful.Fn.Context from a DoFn.ProcessContext, to support the common case when a PTransform is invoking the closure from inside a DoFn.
WritableCoder<T extends org.apache.hadoop.io.Writable> - Class in org.apache.beam.sdk.io.hadoop
A WritableCoder is a Coder for a Java class that implements Writable.
WritableCoder(Class<T>) - Constructor for class org.apache.beam.sdk.io.hadoop.WritableCoder
 
WritableCoder.WritableCoderProviderRegistrar - Class in org.apache.beam.sdk.io.hadoop
A CoderProviderRegistrar which registers a CoderProvider which can handle writable types.
WritableCoderProviderRegistrar() - Constructor for class org.apache.beam.sdk.io.hadoop.WritableCoder.WritableCoderProviderRegistrar
 
write(int) - Method in class org.apache.beam.sdk.fn.stream.DataStreams.ElementDelimitedOutputStream
 
write(byte[], int, int) - Method in class org.apache.beam.sdk.fn.stream.DataStreams.ElementDelimitedOutputStream
 
write() - Static method in class org.apache.beam.sdk.io.amqp.AmqpIO
 
Write() - Constructor for class org.apache.beam.sdk.io.amqp.AmqpIO.Write
 
write(ElementT) - Method in class org.apache.beam.sdk.io.AvroIO.Sink
 
write(Class<T>) - Static method in class org.apache.beam.sdk.io.AvroIO
Writes a PCollection to an Avro file (or multiple Avro files matching a sharding pattern).
write() - Static method in class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO
 
Write() - Constructor for class org.apache.beam.sdk.io.aws.dynamodb.DynamoDBIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.aws.sns.SnsIO
 
Write() - Constructor for class org.apache.beam.sdk.io.aws.sns.SnsIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.aws.sqs.SqsIO
 
Write() - Constructor for class org.apache.beam.sdk.io.aws.sqs.SqsIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO
 
Write() - Constructor for class org.apache.beam.sdk.io.aws2.dynamodb.DynamoDBIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.aws2.sns.SnsIO
 
Write() - Constructor for class org.apache.beam.sdk.io.aws2.sns.SnsIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.aws2.sqs.SqsIO
 
Write() - Constructor for class org.apache.beam.sdk.io.aws2.sqs.SqsIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.cassandra.CassandraIO
Provide a CassandraIO.Write PTransform to write data to a Cassandra database.
Write() - Constructor for class org.apache.beam.sdk.io.cassandra.CassandraIO.Write
 
write(String, String) - Static method in class org.apache.beam.sdk.io.clickhouse.ClickHouseIO
 
Write() - Constructor for class org.apache.beam.sdk.io.clickhouse.ClickHouseIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO
 
Write() - Constructor for class org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write
 
write(OutputT) - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
Called for each value in the bundle.
write(ElementT) - Method in interface org.apache.beam.sdk.io.FileIO.Sink
Appends a single element to the file.
write() - Static method in class org.apache.beam.sdk.io.FileIO
Writes elements to files using a FileIO.Sink.
Write() - Constructor for class org.apache.beam.sdk.io.FileIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
A PTransform that writes a PCollection to a BigQuery table.
Write() - Constructor for class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO
Creates an uninitialized BigtableIO.Write.
Write() - Constructor for class org.apache.beam.sdk.io.gcp.bigtable.BigtableIO.Write
 
write() - Method in class org.apache.beam.sdk.io.gcp.datastore.DatastoreV1
Returns an empty DatastoreV1.Write builder.
Write() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.gcp.spanner.SpannerIO
Creates an uninitialized instance of SpannerIO.Write.
Write() - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.hadoop.format.HadoopFormatIO
Creates an Write.Builder for creation of Write Transformation.
write() - Static method in class org.apache.beam.sdk.io.hbase.HBaseIO
Creates an uninitialized HBaseIO.Write.
write() - Static method in class org.apache.beam.sdk.io.hcatalog.HCatalogIO
Write data to Hive.
Write() - Constructor for class org.apache.beam.sdk.io.hcatalog.HCatalogIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
Write data to a JDBC datasource.
write() - Static method in class org.apache.beam.sdk.io.jms.JmsIO
 
Write() - Constructor for class org.apache.beam.sdk.io.jms.JmsIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.kafka.KafkaIO
Creates an uninitialized KafkaIO.Write PTransform.
Write() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.kinesis.KinesisIO
A PTransform writing data to Kinesis.
Write() - Constructor for class org.apache.beam.sdk.io.kinesis.KinesisIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO
Write data to GridFS.
write(MongoDbGridFSIO.WriteFn<T>) - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO
 
Write() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.Write
 
write(T, OutputStream) - Method in interface org.apache.beam.sdk.io.mongodb.MongoDbGridFSIO.WriteFn
Output the object to the given OutputStream.
write() - Static method in class org.apache.beam.sdk.io.mongodb.MongoDbIO
Write data to MongoDB.
Write() - Constructor for class org.apache.beam.sdk.io.mongodb.MongoDbIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.mqtt.MqttIO
 
Write() - Constructor for class org.apache.beam.sdk.io.mqtt.MqttIO.Write
 
write(GenericRecord) - Method in class org.apache.beam.sdk.io.parquet.ParquetIO.Sink
 
write() - Static method in class org.apache.beam.sdk.io.redis.RedisIO
Write data to a Redis server.
Write() - Constructor for class org.apache.beam.sdk.io.redis.RedisIO.Write
 
write() - Static method in class org.apache.beam.sdk.io.solr.SolrIO
 
Write() - Constructor for class org.apache.beam.sdk.io.solr.SolrIO.Write
 
write(String) - Method in class org.apache.beam.sdk.io.TextIO.Sink
 
write() - Static method in class org.apache.beam.sdk.io.TextIO
A PTransform that writes a PCollection to a text file (or multiple text files matching a sharding pattern), with each element of the input collection encoded into its own line.
write(byte[]) - Method in class org.apache.beam.sdk.io.TFRecordIO.Sink
 
write() - Static method in class org.apache.beam.sdk.io.TFRecordIO
A PTransform that writes a PCollection to TFRecord file (or multiple TFRecord files matching a sharding pattern), with each element of the input collection encoded into its own record.
Write() - Constructor for class org.apache.beam.sdk.io.TFRecordIO.Write
 
write(T) - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.Sink
 
write(T) - Method in class org.apache.beam.sdk.io.thrift.ThriftIO.ThriftWriter
 
write(T) - Method in class org.apache.beam.sdk.io.xml.XmlIO.Sink
 
write() - Static method in class org.apache.beam.sdk.io.xml.XmlIO
Writes all elements in the input PCollection to a single XML file using XmlIO.sink(java.lang.Class<T>).
Write() - Constructor for class org.apache.beam.sdk.io.xml.XmlIO.Write
 
write(T) - Method in interface org.apache.beam.sdk.state.ValueState
Set the value.
writeAvros(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
Returns A PTransform that writes binary encoded Avro messages of a given type to a Google Cloud Pub/Sub stream.
WriteBuilder() - Constructor for class org.apache.beam.sdk.io.gcp.pubsub.ExternalWrite.WriteBuilder
 
writeCompressed(WritableByteChannel) - Method in enum org.apache.beam.sdk.io.Compression
 
writeCustomType() - Static method in class org.apache.beam.sdk.io.AvroIO
A PTransform that writes a PCollection to an avro file (or multiple avro files matching a sharding pattern), with each element of the input collection encoded into its own record of type OutputT.
writeCustomType() - Static method in class org.apache.beam.sdk.io.TextIO
A PTransform that writes a PCollection to a text file (or multiple text files matching a sharding pattern), with each element of the input collection encoded into its own line.
writeCustomTypeToGenericRecords() - Static method in class org.apache.beam.sdk.io.AvroIO
Similar to AvroIO.writeCustomType(), but specialized for the case where the output type is GenericRecord.
writeDefaultJobName(JarOutputStream, String) - Static method in class org.apache.beam.runners.fnexecution.jobsubmission.PortablePipelineJarUtils
 
writeDynamic() - Static method in class org.apache.beam.sdk.io.FileIO
Writes elements to files using a FileIO.Sink and grouping the elements using "dynamic destinations".
writeExternal(ObjectOutput) - Method in class org.apache.beam.sdk.io.hadoop.SerializableConfiguration
 
WriteFiles<UserT,DestinationT,OutputT> - Class in org.apache.beam.sdk.io
A PTransform that writes to a FileBasedSink.
WriteFiles() - Constructor for class org.apache.beam.sdk.io.WriteFiles
 
WriteFilesResult<DestinationT> - Class in org.apache.beam.sdk.io
The result of a WriteFiles transform.
writeFooter() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
Writes footer at the end of output files.
writeGenericRecords(Schema) - Static method in class org.apache.beam.sdk.io.AvroIO
Writes Avro records of the specified schema.
writeGenericRecords(String) - Static method in class org.apache.beam.sdk.io.AvroIO
Writes Avro records of the specified schema.
WriteGrouped(SpannerIO.Write) - Constructor for class org.apache.beam.sdk.io.gcp.spanner.SpannerIO.WriteGrouped
 
writeHeader() - Method in class org.apache.beam.sdk.io.FileBasedSink.Writer
Writes header at the beginning of output files.
writeMessages() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
Returns A PTransform that writes to a Google Cloud Pub/Sub stream.
writeMetrics(MetricQueryResults) - Method in interface org.apache.beam.sdk.metrics.MetricsSink
 
WriteOperation(FileBasedSink<?, DestinationT, OutputT>) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
Constructs a WriteOperation using the default strategy for generating a temporary directory from the base output filename.
WriteOperation(FileBasedSink<?, DestinationT, OutputT>, ResourceId) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.WriteOperation
Create a new WriteOperation.
writeProtos(Class<T>) - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
Returns A PTransform that writes binary encoded protobuf messages of a given type to a Google Cloud Pub/Sub stream.
Writer(FileBasedSink.WriteOperation<DestinationT, OutputT>, String) - Constructor for class org.apache.beam.sdk.io.FileBasedSink.Writer
Construct a new FileBasedSink.Writer that will produce files of the given MIME type.
writeRecords() - Static method in class org.apache.beam.sdk.io.kafka.KafkaIO
Creates an uninitialized KafkaIO.WriteRecords PTransform.
WriteRecords() - Constructor for class org.apache.beam.sdk.io.kafka.KafkaIO.WriteRecords
 
WriteResult - Class in org.apache.beam.sdk.io.gcp.bigquery
The result of a BigQueryIO.Write transform.
writeStrings() - Static method in class org.apache.beam.sdk.io.gcp.pubsub.PubsubIO
Returns A PTransform that writes UTF-8 encoded strings to a Google Cloud Pub/Sub stream.
writeTableRows() - Static method in class org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO
A PTransform that writes a PCollection containing TableRows to a BigQuery table.
writeTo(OutputStream, int, int) - Method in class org.apache.beam.runners.dataflow.util.RandomAccessData
Writes length bytes starting at offset from the backing data store to the specified output stream.
writeToPort(String, BeamFnApi.RemoteGrpcPort) - Static method in class org.apache.beam.sdk.fn.data.RemoteGrpcPortWrite
Create a RemoteGrpcPortWrite which writes the RunnerApi.PCollection with the provided Pipeline id to the provided BeamFnApi.RemoteGrpcPort.
writeVoid() - Static method in class org.apache.beam.sdk.io.jdbc.JdbcIO
 
WriteVoid() - Constructor for class org.apache.beam.sdk.io.jdbc.JdbcIO.WriteVoid
 
WS - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
WS - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 

X

XmlIO - Class in org.apache.beam.sdk.io.xml
Transforms for reading and writing XML files using JAXB mappers.
XmlIO() - Constructor for class org.apache.beam.sdk.io.xml.XmlIO
 
XmlIO.Read<T> - Class in org.apache.beam.sdk.io.xml
Implementation of XmlIO.read().
XmlIO.Read.CompressionType - Enum in org.apache.beam.sdk.io.xml
Deprecated.
Use Compression instead.
XmlIO.ReadFiles<T> - Class in org.apache.beam.sdk.io.xml
Implementation of XmlIO.readFiles().
XmlIO.Sink<T> - Class in org.apache.beam.sdk.io.xml
XmlIO.Write<T> - Class in org.apache.beam.sdk.io.xml
Implementation of XmlIO.write().
XmlSource<T> - Class in org.apache.beam.sdk.io.xml
Implementation of XmlIO.read().

Y

years(int) - Static method in class org.apache.beam.sdk.transforms.windowing.CalendarWindows
Returns a WindowFn that windows elements into periods measured by years.

Z

ZETASQL_CAST_OP - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLCastFunctionImpl
 
ZETASQL_FUNCTION_TO_CALCITE_SQL_OPERATOR - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlStdOperatorMappingTable
 
ZETASQL_FUNCTION_TO_CALCITE_SQL_OPERATOR_REWRITER - Static variable in class org.apache.beam.sdk.extensions.sql.zetasql.SqlStdOperatorMappingTable
 
ZetaSQLCastFunctionImpl - Class in org.apache.beam.sdk.extensions.sql.zetasql
ZetaSQLCastFunctionImpl.
ZetaSQLCastFunctionImpl() - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLCastFunctionImpl
 
ZetaSQLPlannerImpl - Class in org.apache.beam.sdk.extensions.sql.zetasql
ZetaSQLPlannerImpl.
ZetaSQLPlannerImpl(FrameworkConfig) - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLPlannerImpl
 
ZetaSQLQueryPlanner - Class in org.apache.beam.sdk.extensions.sql.zetasql
ZetaSQLQueryPlanner.
ZetaSQLQueryPlanner(FrameworkConfig) - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
 
ZetaSQLQueryPlanner(JdbcConnection, RuleSet[]) - Constructor for class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSQLQueryPlanner
 
zetaSqlTypeToCalciteType(ZetaSQLType.TypeKind) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlUtils
 
ZetaSqlUtils - Class in org.apache.beam.sdk.extensions.sql.zetasql
Utility methods for ZetaSQL related operations.
zetaSqlValueToJavaObject(Value, Schema.FieldType, boolean) - Static method in class org.apache.beam.sdk.extensions.sql.zetasql.ZetaSqlUtils
 

_

_ATN - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
_ATN - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
_decisionToDFA - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
_decisionToDFA - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
_serializedATN - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
_serializedATN - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
_sharedContextCache - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationLexer
 
_sharedContextCache - Static variable in class org.apache.beam.sdk.schemas.parser.generated.FieldSpecifierNotationParser
 
A B C D E F G H I J K L M N O P Q R S T U V W X Y Z _