PCollectionView<T> view
boolean streaming
DataflowPipelineJob job
PCollectionView<T> view
Coder<T> dataCoder
java.util.Map<K,V> properties
PCollectionView<T> view
org.apache.beam.runners.core.construction.SerializablePipelineOptions options
int intTag
org.apache.beam.runners.core.construction.SerializablePipelineOptions options
org.apache.beam.runners.core.metrics.MetricsContainerStepMap metricsContainers
ReferenceCountingExecutableStageContextFactory.Creator creator
SerializableFunction<InputT,OutputT> isReleaseSynchronous
long count
org.apache.beam.runners.core.metrics.DistributionData distributionData
org.apache.beam.runners.core.metrics.GaugeData gaugeData
java.lang.String stepName
java.lang.String metricsKey
java.util.Map<K,V> counters
java.util.Map<K,V> distributions
java.util.Map<K,V> gauges
com.hazelcast.map.IMap<K,V> accumulator
java.util.Map<K,V> mNamedAggregators
NamedAggregators value
int num
UnboundedSource<OutputT,CheckpointMarkT extends UnboundedSource.CheckpointMark> source
Duration maxReadTime
int numInitialSplits
long maxNumRecords
int sourceId
double readerCacheInterval
int splitId
BoundedSource<T> source
org.apache.beam.runners.core.construction.SerializablePipelineOptions options
int numPartitions
long bundleSize
java.lang.String stepName
MetricsContainerStepMapAccumulator metricsAccum
MicrobatchSource<T,CheckpointMarkT extends UnboundedSource.CheckpointMark> microbatchSource
org.apache.beam.runners.core.construction.SerializablePipelineOptions options
org.apache.spark.Partitioner partitioner
org.apache.beam.runners.core.metrics.MetricsContainerStepMap value
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.collect.Table<R,C,V> state
java.util.Collection<E> serTimers
java.util.Map<K,V> mNamedAggregators
NamedAggregators value
org.apache.beam.runners.core.metrics.MetricsContainerStepMap value
MetricsContainerStepMapAccumulator metricsAccum
java.lang.String stepName
DoFn<InputT,OutputT> doFn
WindowingStrategy<T,W extends BoundedWindow> windowingStrategy
java.util.Map<K,V> sideInputs
org.apache.beam.runners.core.construction.SerializablePipelineOptions serializableOptions
java.util.List<E> additionalOutputTags
TupleTag<V> mainOutputTag
Coder<T> inputCoder
java.util.Map<K,V> outputCoderMap
SideInputBroadcast broadcastStateData
DoFnSchemaInformation doFnSchemaInformation
java.util.Map<K,V> sideInputMapping
WindowingStrategy<T,W extends BoundedWindow> windowingStrategy
org.apache.beam.runners.core.StateInternalsFactory<K> stateInternalsFactory
org.apache.beam.runners.core.SystemReduceFn<K,InputT,AccumT,OutputT,W extends BoundedWindow> reduceFn
org.apache.beam.runners.core.construction.SerializablePipelineOptions options
java.util.Map<K,V> bcast
java.util.Map<K,V> coder
byte[] value
java.util.HashMap<K,V> sideInputDataSets
java.util.Set<E> leaves
protected java.lang.Object readResolve() throws java.io.ObjectStreamException
java.io.ObjectStreamException
protected java.lang.Object readResolve() throws java.io.ObjectStreamException
java.io.ObjectStreamException
byte[] windowFnBytes
java.lang.String serializedOptions
protected java.lang.Object readResolve() throws java.io.ObjectStreamException
java.io.ObjectStreamException
byte[] keyCoderBytes
byte[] wvCoderBytes
protected java.lang.Object readResolve() throws java.io.ObjectStreamException
java.io.ObjectStreamException
byte[] keyCoderBytes
byte[] wvCoderBytes
protected java.lang.Object readResolve() throws java.io.ObjectStreamException
java.io.ObjectStreamException
org.apache.beam.runners.core.StepContext stepcontext
byte[] doFnwithExBytes
byte[] coderBytes
java.util.Map<K,V> outputCodersBytes
byte[] windowBytes
java.util.Map<K,V> sideInputBytes
java.lang.String serializedOptions
java.util.List<E> serializedSideOutputs
java.util.Map<K,V> serializedOutputMap
protected java.lang.Object readResolve() throws java.io.ObjectStreamException
java.io.ObjectStreamException
org.apache.beam.runners.core.SystemReduceFn<K,InputT,AccumT,OutputT,W extends BoundedWindow> reduceFn
java.lang.String serializedOptions
byte[] windowBytes
boolean impulseEmitted
protected java.lang.Object readResolve() throws java.io.ObjectStreamException
java.io.ObjectStreamException
byte[] keyCoderBytes
byte[] wvCoderBytes
int tag
protected java.lang.Object readResolve() throws java.io.ObjectStreamException
java.io.ObjectStreamException
byte[] coderBytes
java.lang.Class<T> type
org.apache.beam.sdk.coders.AvroCoder.SerializableSchemaSupplier schemaSupplier
TypeDescriptor<T> typeDescriptor
java.util.List<E> nonDeterministicReasons
org.apache.beam.sdk.util.EmptyOnDeserializationThreadLocal<T> decoder
org.apache.beam.sdk.util.EmptyOnDeserializationThreadLocal<T> encoder
org.apache.beam.sdk.util.EmptyOnDeserializationThreadLocal<T> writer
org.apache.beam.sdk.util.EmptyOnDeserializationThreadLocal<T> reader
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.base.Supplier<T> reflectData
CannotProvideCoderException.ReasonCode reason
Coder<T> coder
DelegateCoder.CodingFunction<InputT,OutputT> toFn
DelegateCoder.CodingFunction<InputT,OutputT> fromFn
TypeDescriptor<T> typeDescriptor
java.lang.Class<T> type
Coder<T> keyCoder
VarIntCoder shardNumberCoder
DelegateCoder<T,IntermediateT> delegateCoder
java.lang.Class<T> clazz
GcsPath gcsPath
java.lang.String bucket
java.lang.String object
java.lang.Class<T> inputClass
ObjectMapper customMapper
InferableFunction<InputT,OutputT> exceptionHandler
java.lang.Class<T> outputClass
ObjectMapper customMapper
InferableFunction<InputT,OutputT> exceptionHandler
java.lang.Object leftNullValue
java.lang.Object rightNullValue
java.lang.Object nullValue
java.lang.Object nullValue
PCollectionView<T> contextSideInput
java.util.List<E> featureList
java.util.List<E> featureList
PCollectionView<T> contextSideInput
java.util.List<E> featureList
java.util.List<E> featureList
private void readObject(java.io.ObjectInputStream ois) throws java.lang.ClassNotFoundException, java.io.IOException
java.lang.ClassNotFoundException
java.io.IOException
private void writeObject(java.io.ObjectOutputStream oos) throws java.io.IOException
java.io.IOException
java.lang.Class<T> protoMessageClass
Message
type to be coded.java.util.Set<E> extensionHostClasses
ProtoCoder
. The extensions from these
classes will be included in the ExtensionRegistry
used during encoding and decoding.private void readObject(java.io.ObjectInputStream ois) throws java.io.IOException
java.io.IOException
private void writeObject(java.io.ObjectOutputStream oos) throws java.io.IOException
java.io.IOException
private java.lang.Object readResolve()
org.apache.beam.sdk.extensions.protobuf.ProtoDynamicMessageSchema.Context<T> context
double compression
java.lang.String tempLocation
int memoryMB
ExternalSorter.Options.SorterType sorterType
java.lang.String tempLocation
int memoryMB
ExternalSorter.Options.SorterType sorterType
long memoryMB
BufferedExternalSorter.Options sorterOptions
int id
java.lang.String name
java.lang.String countryOfResidence
int id
int customerId
boolean unknown
java.lang.Double rowCount
java.lang.Double rate
Combine.CombineFn<InputT,AccumT,OutputT> combineFn
CEPOperator operator
java.util.List<E> operands
java.lang.String alpha
int fieldIndex
Schema.TypeName typeName
java.lang.String outTableName
CEPOperation opr
CEPFieldRef fieldRef
Schema.FieldType fieldType
CEPKind cepKind
Schema mySchema
java.lang.String patternVar
CEPCall patternCondition
Quantifier quant
int fIndex
boolean dir
boolean nullFirst
java.lang.String patternVar
java.lang.String repr
int start
int end
boolean isReluctant
org.apache.beam.sdk.extensions.sql.impl.nfa.NFA.State startState
java.util.ArrayList<E> currentRuns
Schema upstreamSchema
BeamRelNode beamRelNode
boolean all
BeamSetOperatorRelBase.OpType opType
java.util.List<E> fieldsIndices
java.util.List<E> orientation
java.util.List<E> nullsFirst
BeamSqlSeekableTable seekableTable
Schema lkpSchema
int factColOffset
Schema joinSubsetType
Schema outputSchema
java.util.List<E> factJoinIdx
java.lang.String leftTag
java.lang.String rightTag
BeamSetOperatorRelBase.OpType opType
boolean all
java.lang.String stepName
boolean isSample
SerializableFunction<InputT,OutputT> decimalConverter
boolean isSample
SerializableFunction<InputT,OutputT> decimalConverter
java.util.List<E> indexes
int index
Schema schema
java.lang.String projectId
java.lang.String instanceId
java.lang.String tableId
java.lang.String emulatorHost
boolean useFlatSchema
java.util.Map<K,V> columnsMapping
CSVFormat csvFormat
java.lang.Class<T> protoClass
java.lang.String bootstrapServers
java.util.List<E> topics
java.util.List<E> topicPartitions
java.util.Map<K,V> configUpdates
BeamTableStatistics rowCountStatistics
int numberOfRecordsForRate
java.util.regex.Pattern locationPattern
java.lang.String dbCollection
java.lang.String dbName
java.lang.String dbUri
Schema schema
java.util.List<E> rows
Table table
java.util.List<E> rows
long tableProviderInstanceId
java.util.List<E> timestampedRows
int timestampField
BeamTableStatistics statistics
PTransform<InputT extends PInput,OutputT extends POutput> readConverter
PTransform<InputT extends PInput,OutputT extends POutput> writeConverter
java.lang.String filePattern
BeamTableStatistics rowCountStatistics
AvroIO.TypedWrite<UserT,DestinationT,OutputT> inner
boolean genericRecords
private java.lang.Object readResolve() throws java.io.ObjectStreamException
java.io.ObjectStreamException
org.apache.beam.sdk.io.AvroSource.Mode<T> mode
UnboundedSource<OutputT,CheckpointMarkT extends UnboundedSource.CheckpointMark> source
long maxNumRecords
Duration maxReadTime
java.lang.String path
FileBasedSource<T> sourceDelegate
CompressedSource.DecompressingChannelFactory channelFactory
DefaultFilenamePolicy.Params params
ValueProvider<T> baseFilename
java.lang.String shardTemplate
boolean explicitTemplate
java.lang.String suffix
FileBasedSink.DynamicDestinations<UserT,DestinationT,OutputT> dynamicDestinations
FileBasedSink.WritableByteChannelFactory writableByteChannelFactory
FileBasedSink.WritableByteChannelFactory
that is used to wrap the raw data output to the
underlying channel. The default is to not compress the output using Compression.UNCOMPRESSED
.ValueProvider<T> tempDirectoryProvider
FileBasedSink<UserT,DestinationT,OutputT> sink
ValueProvider<T> tempDirectory
boolean windowedWrites
ValueProvider<T> fileOrPatternSpec
EmptyMatchTreatment emptyMatchTreatment
MatchResult.Metadata singleFileMetadata
FileBasedSource.Mode mode
long startOffset
long endOffset
long minBundleSize
BoundedSource<T> source
UnboundedSource<OutputT,CheckpointMarkT extends UnboundedSource.CheckpointMark> source
long desiredBundleSizeBytes
SerializableFunction<InputT,OutputT> createSource
Coder<T> coder
TextIO.TypedWrite<UserT,DestinationT> inner
FileBasedSink.WriteOperation<DestinationT,OutputT> writeOperation
java.lang.String accessKey
java.lang.String secretKey
Regions region
java.lang.String serviceEndpoint
java.lang.String awsCredentialsProviderSerialized
java.lang.String region
java.net.URI serviceEndpoint
WatermarkPolicyFactory.CustomWatermarkPolicy watermarkPolicy
WatermarkParameters watermarkParameters
int backendVersion
ElasticsearchIO.Read spec
java.lang.String shardPreference
java.lang.Integer numSlices
java.lang.Integer sliceId
java.lang.Long estimatedByteSize
BigQueryIO.TypedRead<T> inner
private void readObject(java.io.ObjectInputStream in) throws java.lang.ClassNotFoundException, java.io.IOException
java.lang.ClassNotFoundException
java.io.IOException
java.lang.String stepUuid
ValueProvider<T> queryProvider
java.lang.Boolean flattenResults
java.lang.Boolean useLegacySql
BigQueryIO.TypedRead.QueryPriority priority
java.lang.String location
java.lang.String queryTempDataset
java.lang.String kmsKey
com.google.cloud.bigquery.storage.v1.ReadSession readSession
com.google.cloud.bigquery.storage.v1.ReadStream readStream
java.lang.String jsonTableSchema
SerializableFunction<InputT,OutputT> parseFn
Coder<T> outputCoder
BigQueryServices bqServices
private void readObject(java.io.ObjectInputStream in) throws java.lang.ClassNotFoundException, java.io.IOException
java.lang.ClassNotFoundException
java.io.IOException
ValueProvider<T> tableReferenceProvider
BigQueryIO.Write.CreateDisposition createDisposition
BigQueryServices bqServices
DynamicDestinations<T,DestinationT> dynamicDestinations
java.lang.String kmsKey
DynamicDestinations<T,DestinationT> dynamicDestinations
SerializableFunction<InputT,OutputT> formatFunction
BigQueryServices bigQueryServices
BigQueryIO.Write.CreateDisposition createDisposition
DynamicDestinations<T,DestinationT> dynamicDestinations
InsertRetryPolicy retryPolicy
boolean extendedErrorInfo
boolean skipInvalidRows
boolean ignoreUnknownValues
boolean ignoreInsertIds
java.lang.String kmsKey
Coder<T> elementCoder
SerializableFunction<InputT,OutputT> toTableRow
SerializableFunction<InputT,OutputT> toFailsafeTableRow
BigQueryServices bigQueryServices
InsertRetryPolicy retryPolicy
boolean extendedErrorInfo
boolean skipInvalidRows
boolean ignoreUnknownValues
boolean ignoreInsertIds
Coder<T> elementCoder
SerializableFunction<InputT,OutputT> toTableRow
SerializableFunction<InputT,OutputT> toFailsafeTableRow
java.lang.String tableSpec
java.lang.String tableDescription
java.lang.String jsonTimePartitioning
java.lang.String jsonClustering
java.util.Map<K,V> columnFamilyMapping
java.util.Map<K,V> columnFamilyMapping
org.apache.beam.sdk.io.gcp.bigtable.CellValueParser cellValueParser
org.apache.beam.sdk.io.gcp.bigtable.BigtableConfig bigtableConfig
Schema schema
Schema schema
java.util.Map<K,V> columnsMapping
Schema dataSchema
java.lang.String location
java.lang.String kind
java.lang.String projectId
java.lang.String keyField
Schema schema
java.lang.String keyField
java.util.function.Supplier<T> keySupplier
java.lang.String kind
java.lang.String keyField
ValueProvider<T> sourceFhirStore
ValueProvider<T> destinationFhirStore
ValueProvider<T> deidConfig
HealthcareApiClient client
ValueProvider<T> destinationFhirStore
java.lang.String deidConfigJson
ValueProvider<T> fhirStore
ValueProvider<T> fhirStore
ValueProvider<T> exportGcsUriPrefix
HealthcareApiClient client
ValueProvider<T> exportGcsUriPrefix
ValueProvider<T> fhirStore
ValueProvider<T> deadLetterGcsPath
FhirIO.Import.ContentStructure contentStructure
ValueProvider<T> tempGcsPath
ValueProvider<T> fhirStore
ValueProvider<T> hl7v2Stores
ValueProvider<T> filter
Duration initialSplitDuration
Counter failedMessageGets
Counter successfulHL7v2MessageGets
HealthcareApiClient client
int statusCode
java.lang.String projectId
java.lang.String projectId
java.lang.String subscriptionName
java.lang.String path
org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription.Type type
java.lang.String project
java.lang.String subscription
org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic.Type type
java.lang.String project
java.lang.String topic
int maxPublishBatchByteSize
int maxPublishBatchSize
PubsubClient.PubsubClientFactory pubsubFactory
ValueProvider<T> topic
java.lang.String timestampAttribute
java.lang.String idAttribute
int numShards
int publishBatchSize
int publishBatchBytes
Duration maxLatency
org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink.RecordIdMethod recordIdMethod
PubsubUnboundedSink.idAttribute
is non-null).com.google.api.client.util.Clock clock
PubsubClient.PubsubClientFactory pubsubFactory
ValueProvider<T> project
PubsubUnboundedSource.topic
was given.ValueProvider<T> topic
PubsubUnboundedSource.subscription
must be given. Otherwise
PubsubUnboundedSource.subscription
must be null.ValueProvider<T> subscription
PubsubUnboundedSource.topic
must be given. Otherwise
PubsubUnboundedSource.topic
must be null.
If no subscription is given a random one will be created when the transorm is applied. This field will be update with that subscription's path. The created subscription is never deleted.
java.lang.String timestampAttribute
java.lang.String idAttribute
boolean needsAttributes
boolean needsMessageId
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.collect.ImmutableList<E> mutations
SpannerIO.Write spec
java.lang.String expectedChecksum
java.lang.String actualChecksum
BigqueryClient bigqueryClient
BigQueryServices.JobService jobService
BigQueryServices.DatasetService datasetService
BigQueryServices.StorageClient storageClient
java.util.List<E> items
java.util.Map<K,V> insertErrors
int numFailuresExpected
int numFailures
FakeDatasetService datasetService
public void readExternal(java.io.ObjectInput in) throws java.io.IOException, java.lang.ClassNotFoundException
java.io.IOException
java.lang.ClassNotFoundException
public void writeExternal(java.io.ObjectOutput out) throws java.io.IOException
java.io.IOException
java.lang.Class<T> type
SerializableConfiguration conf
Coder<T> keyCoder
Coder<T> valueCoder
SimpleFunction<InputT,OutputT> keyTranslationFunction
SimpleFunction<InputT,OutputT> valueTranslationFunction
HadoopFormatIO.SerializableSplit inputSplit
long boundedSourceEstimatedSize
private void readObject(java.io.ObjectInputStream in) throws java.io.IOException
java.io.IOException
private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException
java.io.IOException
InputSplit inputSplit
PTransform<InputT extends PInput,OutputT extends POutput> configTransform
ExternalSynchronization externalSynchronization
boolean withPartitioning
java.lang.String locksDir
org.apache.beam.sdk.io.hadoop.format.HDFSSynchronization.ThrowingFunction<T1,T2,X extends java.lang.Exception> fileSystemFactory
private java.lang.Object writeReplace()
Configuration configuration
java.lang.String tableId
org.apache.hadoop.hbase.client.Scan scan
private java.lang.Object writeReplace()
Configuration configuration
java.lang.String tableId
JdbcIO.DataSourceConfiguration config
JdbcIO.DataSourceProviderFromDataSourceConfiguration config
JdbcIO.WriteVoid<T> inner
java.lang.String jmsMessageID
long jmsTimestamp
java.lang.String jmsCorrelationID
javax.jms.Destination jmsReplyTo
javax.jms.Destination jmsDestination
int jmsDeliveryMode
boolean jmsRedelivered
java.lang.String jmsType
long jmsExpiration
int jmsPriority
java.util.Map<K,V> properties
java.lang.String text
SerializableFunction<InputT,OutputT> schemaRegistryClientProviderFn
java.lang.String schemaRegistryUrl
java.lang.String subject
java.lang.Integer version
java.lang.String topic
int partition
long nextOffset
long watermarkMillis
KafkaIO.ReadSourceDescriptors<K,V> readSourceDescriptors
KafkaIO.Read<K,V> read
TopicPartition topicPartition
WatermarkPolicyFactory.CustomWatermarkPolicy watermarkPolicy
WatermarkParameters watermarkParameters
MongoDbGridFSIO.Read<T> spec
java.util.List<E> objectIds
java.lang.String routingKey
byte[] body
java.lang.String contentType
java.lang.String contentEncoding
java.util.Map<K,V> headers
java.lang.Integer deliveryMode
java.lang.Integer priority
java.lang.String correlationId
java.lang.String replyTo
java.lang.String expiration
java.lang.String messageId
java.util.Date timestamp
java.lang.String type
java.lang.String userId
java.lang.String appId
java.lang.String clusterId
org.apache.beam.vendor.grpc.v1p26p0.com.google.protobuf.ByteString value
long from
long to
SnowflakeIO.DataSourceConfiguration config
ValueProvider<T> stagingBucketDir
java.lang.String tmpDirName
ValueProvider<T> quoteChar
SnowflakeDataType dataType
java.lang.String name
boolean isNullable
SnowflakeColumn[] columns
int precision
int scale
java.lang.Long size
java.lang.Long length
java.lang.Class<T> type
org.apache.thrift.protocol.TProtocolFactory protocolFactory
java.util.Map<K,V> typedefs
java.lang.String fileLocation
java.lang.String content
org.apache.tika.metadata.Metadata metadata
java.lang.String[] metadataNames
org.apache.beam.sdk.util.SerializableThrowable error
java.lang.Class<T> jaxbClass
org.apache.beam.sdk.util.EmptyOnDeserializationThreadLocal<T> jaxbMarshaller
org.apache.beam.sdk.util.EmptyOnDeserializationThreadLocal<T> jaxbUnmarshaller
org.apache.beam.sdk.io.xml.XmlIO.MappingConfiguration<T> configuration
MetricName name
ValueProvider<T> value
SerializableFunction<InputT,OutputT> translator
java.lang.Class<T> klass
java.lang.String methodName
java.lang.String propertyName
java.lang.Object defaultValue
java.lang.Long optionsId
java.lang.Object value
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.collect.BiMap<K,V> fieldIndices
java.util.Map<K,V> encodingPositions
java.util.List<E> fields
int hashCode
java.util.UUID uuid
Schema.Options options
java.util.Map<K,V> options
Schema schema
TypeDescriptor<T> typeDescriptor
SerializableFunction<InputT,OutputT> toRowFunction
SerializableFunction<InputT,OutputT> fromRowFunction
java.util.Map<K,V> cachedProviders
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.collect.BiMap<K,V> enumValues
java.util.List<E> values
int value
int byteArraySize
Schema schema
Schema oneOfSchema
EnumerationType enumerationType
byte[] schemaProtoRepresentation
java.lang.String identifier
Schema.FieldType argumentType
java.lang.Object argument
Schema.FieldType fieldType
java.util.List<E> newFields
org.apache.beam.sdk.schemas.transforms.Cast.Narrowing.Fold fold
org.apache.beam.sdk.schemas.transforms.Cast.Widening.Fold fold
org.apache.beam.sdk.schemas.transforms.CoGroup.JoinArguments joinArgs
org.apache.beam.sdk.schemas.transforms.CoGroup.JoinArguments joinArgs
java.lang.String keyFieldName
FieldAccessDescriptor fieldsToDrop
RowSelector rowSelector
java.util.List<E> filters
org.apache.beam.sdk.schemas.transforms.SchemaAggregateFn.Inner schemaAggregateFn
Combine.CombineFn<InputT,AccumT,OutputT> combineFn
FieldAccessDescriptor lhs
FieldAccessDescriptor rhs
org.apache.beam.sdk.schemas.transforms.Join.JoinType joinType
Join.FieldsEqual.Impl predicate
java.util.List<E> renames
SchemaCoder<T> outputSchemaCoder
Schema.FieldType unboxedType
Schema inputSchema
FieldAccessDescriptor fieldAccessDescriptor
boolean optimized
java.lang.String expectedChecksum
java.lang.String actualChecksum
SerializableFunction<InputT,OutputT> checkerFn
org.apache.beam.sdk.testing.PAssert.AssertionWindows rewindowingStrategy
SimpleFunction<InputT,OutputT> paneExtractor
PAssert.PAssertionSite site
SerializableFunction<InputT,OutputT> checkerFn
org.apache.beam.sdk.testing.PAssert.AssertionWindows rewindowingStrategy
SimpleFunction<InputT,OutputT> paneExtractor
PAssert.PAssertionSite site
SerializableFunction<InputT,OutputT> checkerFn
PAssert.PAssertionSite site
java.lang.String message
java.lang.StackTraceElement[] creationStackTrace
SerializableMatcher<T> matcher
boolean isSuccess
PAssert.PAssertionSite site
org.apache.beam.sdk.util.SerializableThrowable throwable
TimestampedValue.TimestampedValueCoder<T> elementCoder
java.util.Comparator<T> compareFn
int numQuantiles
int bufferSize
int numBuffers
long maxNumElements
java.util.TreeSet<E> heap
long minHash
long sampleSize
long sampleSize
java.lang.Double maximumEstimationError
long sampleSize
java.lang.Double maximumEstimationError
CombineFnBase.GlobalCombineFn<InputT,AccumT,OutputT> fn
DisplayData.ItemSpec<T> fnDisplayData
boolean insertDefault
int fanout
java.util.List<E> sideInputs
CombineFnBase.GlobalCombineFn<InputT,AccumT,OutputT> fn
DisplayData.ItemSpec<T> fnDisplayData
boolean insertDefault
int fanout
CombineFnBase.GlobalCombineFn<InputT,AccumT,OutputT> fn
DisplayData.ItemSpec<T> fnDisplayData
java.util.List<E> sideInputs
SerializableFunction<InputT,OutputT> combiner
int bufferSize
CombineFnBase.GlobalCombineFn<InputT,AccumT,OutputT> fn
DisplayData.ItemSpec<T> fnDisplayData
boolean fewKeys
java.util.List<E> sideInputs
CombineFnBase.GlobalCombineFn<InputT,AccumT,OutputT> fn
DisplayData.ItemSpec<T> fnDisplayData
SerializableFunction<InputT,OutputT> hotKeyFanout
java.util.Map<K,V> valuesMap
java.util.List<E> combineFns
java.util.List<E> combineInputCoders
java.util.List<E> extractInputFns
java.util.List<E> outputTags
int combineFnCount
java.util.List<E> extractInputFns
java.util.List<E> combineInputCoders
java.util.List<E> combineFnWithContexts
java.util.List<E> outputTags
int combineFnCount
java.lang.Object closure
Requirements requirements
ValueProvider<T> provider
Coder<T> coder
TimeDomain timeDomain
Duration duration
TimeDomain timeDomain
Duration duration
SerializableFunction<InputT,OutputT> fn
TypeDescriptor<T> type
Coder<T> coder
TimeDomain timeDomain
Duration duration
SerializableFunction<InputT,OutputT> fn
TypeDescriptor<T> representativeType
ProcessFunction<InputT,OutputT> predicate
java.lang.String predicateDescription
Contextful<ClosureT> fn
Contextful<ClosureT> fn
ProcessFunction<InputT,OutputT> exceptionHandler
boolean fewKeys
long batchSize
Duration maxBufferingDuration
ProcessFunction<InputT,OutputT> fn
Contextful<ClosureT> fn
Contextful<ClosureT> fn
ProcessFunction<InputT,OutputT> exceptionHandler
java.util.Map<K,V> sideInputs
TupleTag<V> mainOutputTag
TupleTagList additionalOutputTags
DisplayData.ItemSpec<T> fnDisplayData
DoFn<InputT,OutputT> fn
java.util.Map<K,V> sideInputs
DoFn<InputT,OutputT> fn
DisplayData.ItemSpec<T> fnDisplayData
private void readObject(java.io.ObjectInputStream oos)
private void writeObject(java.io.ObjectOutputStream oos)
java.util.regex.Pattern pattern
java.util.regex.Pattern pattern
int group
java.util.regex.Pattern pattern
java.util.regex.Pattern pattern
int keyGroup
int valueGroup
java.util.regex.Pattern pattern
java.lang.String groupName
java.util.regex.Pattern pattern
java.lang.String keyGroupName
java.lang.String valueGroupName
java.util.regex.Pattern pattern
int group
java.util.regex.Pattern pattern
int keyGroup
int valueGroup
java.util.regex.Pattern pattern
java.lang.String groupName
java.util.regex.Pattern pattern
java.lang.String keyGroupName
java.lang.String valueGroupName
java.util.regex.Pattern pattern
java.lang.String replacement
java.util.regex.Pattern pattern
java.lang.String replacement
java.util.regex.Pattern pattern
boolean outputEmpty
java.util.Collection<E> sideInputs
java.lang.Integer numBuckets
int sampleSize
Top.TopCombineFn<T,ComparatorT extends java.util.Comparator<T> & java.io.Serializable> topCombineFn
java.util.Random rand
SerializableFunction<InputT,OutputT> fn
int count
java.util.Comparator<T> compareFn
java.lang.Object defaultValue
boolean hasDefault
PCollectionView<T> view
SerializableFunction<InputT,OutputT> fn
SerializableFunction<InputT,OutputT> fn
Duration allowedTimestampSkew
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.collect.ImmutableMap<K,V> entries
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.collect.ImmutableList<E> components
CoGbkResultSchema schema
UnionCoder unionCoder
TupleTagList tupleTagList
java.util.HashMap<K,V> tagMap
java.util.List<E> elementCoders
int countElems
java.util.List<E> timestampTransforms
Trigger.OnceTrigger earlyTrigger
Trigger.OnceTrigger lateTrigger
int number
DateTime startDate
DateTimeZone timeZone
int number
int dayOfMonth
DateTime startDate
DateTimeZone timeZone
int number
int monthOfYear
int dayOfMonth
DateTime startDate
DateTimeZone timeZone
WindowFn<T,W extends BoundedWindow> givenWindowFn
java.lang.String reason
java.lang.String cause
WindowFn<T,W extends BoundedWindow> originalWindowFn
Trigger repeatedTrigger
Duration gapDuration
java.util.List<E> subTriggers
Window<T> original
WindowingStrategy<T,W extends BoundedWindow> updatedStrategy
Duration maximumLookback
java.lang.Object key
java.lang.Object value
PCollectionViews.TypeDescriptorSupplier<T> typeDescriptorSupplier
PCollectionViews.TypeDescriptorSupplier<T> typeDescriptorSupplier
PCollectionViews.TypeDescriptorSupplier<T> keyTypeDescriptorSupplier
PCollectionViews.TypeDescriptorSupplier<T> valueTypeDescriptorSupplier
PCollectionViews.TypeDescriptorSupplier<T> keyTypeDescriptorSupplier
PCollectionViews.TypeDescriptorSupplier<T> valueTypeDescriptorSupplier
TupleTag<V> tag
WindowMappingFn<TargetWindowT extends BoundedWindow> windowMappingFn
WindowingStrategy<T,W extends BoundedWindow> windowingStrategy
Coder<T> coder
ViewFn<PrimitiveViewT,ViewT> viewFn
ViewFn
for this view.byte[] encodedDefaultValue
Coder<T> valueCoder
boolean hasDefault
PCollectionViews.TypeDescriptorSupplier<T> typeDescriptorSupplier
Schema schema
java.util.List<E> values
java.lang.Object key
int shardNumber
java.lang.String id
boolean generated
java.util.List<E> tupleTags
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.reflect.TypeToken<T> token
Coder<T> valueCoder
ByteArrayCoder idCoder
WindowFn<T,W extends BoundedWindow> windowFn
Trigger trigger
WindowingStrategy.AccumulationMode mode
Duration allowedLateness
Window.ClosingBehavior closingBehavior
Window.OnTimeBehavior onTimeBehavior
TimestampCombiner timestampCombiner
java.lang.String environmentId
boolean triggerSpecified
boolean modeSpecified
boolean allowedLatenessSpecified
boolean timestampCombinerSpecified