PCollectionView<T> view
com.datatorrent.api.DAG dag
com.datatorrent.api.Attribute.AttributeMap launchAttributes
java.util.Properties configProperties
java.util.HashMap<K,V> env
java.lang.String cmd
PCollectionView<T> view
DataflowPipelineJob job
PCollectionView<T> view
Coder<T> dataCoder
ParDo.SingleOutput<InputT,OutputT> original
DoFn<InputT,OutputT> doFn
WindowingStrategy<T,W extends BoundedWindow> windowingStrategy
java.lang.Iterable<T> sideInputViews
Coder<T> inputCoder
long mainOutput
java.util.Map<K,V> outputMap
PCollectionView<T> view
org.apache.beam.runners.core.metrics.MetricsContainerStepMap metricsContainers
java.util.Map<K,V> mNamedAggregators
private void readObject(java.io.ObjectInputStream ois) throws java.io.IOException, java.lang.ClassNotFoundException
java.io.IOException
java.lang.ClassNotFoundException
private void writeObject(java.io.ObjectOutputStream oos) throws java.io.IOException
java.io.IOException
int num
UnboundedSource<OutputT,CheckpointMarkT extends UnboundedSource.CheckpointMark> source
Duration maxReadTime
int numInitialSplits
long maxNumRecords
int sourceId
double readerCacheInterval
int splitId
BoundedSource<T> source
org.apache.beam.runners.spark.translation.SparkRuntimeContext runtimeContext
int numPartitions
java.lang.String stepName
org.apache.spark.Accumulator<T> metricsAccum
MicrobatchSource<T,CheckpointMarkT extends UnboundedSource.CheckpointMark> microbatchSource
org.apache.beam.runners.spark.translation.SparkRuntimeContext runtimeContext
org.apache.spark.Partitioner partitioner
byte[] value
PTransform<InputT extends PInput,OutputT extends POutput> transform
java.lang.Class<T> type
org.apache.beam.sdk.coders.AvroCoder.SerializableSchemaSupplier schemaSupplier
TypeDescriptor<T> typeDescriptor
java.util.List<E> nonDeterministicReasons
org.apache.beam.sdk.util.EmptyOnDeserializationThreadLocal<T> decoder
org.apache.beam.sdk.util.EmptyOnDeserializationThreadLocal<T> encoder
org.apache.beam.sdk.util.EmptyOnDeserializationThreadLocal<T> writer
org.apache.beam.sdk.util.EmptyOnDeserializationThreadLocal<T> reader
Supplier<T> reflectData
CannotProvideCoderException.ReasonCode reason
Coder<T> coder
DelegateCoder.CodingFunction<InputT,OutputT> toFn
DelegateCoder.CodingFunction<InputT,OutputT> fromFn
TypeDescriptor<T> typeDescriptor
java.lang.Class<T> type
DelegateCoder<T,IntermediateT> delegateCoder
java.lang.Class<T> clazz
org.apache.beam.sdk.util.gcsfs.GcsPath gcsPath
java.lang.Class<T> inputClass
ObjectMapper customMapper
java.lang.Class<T> outputClass
ObjectMapper customMapper
java.lang.Class<T> protoMessageClass
Message
type to be coded.java.util.Set<E> extensionHostClasses
ProtoCoder
. The extensions from these
classes will be included in the ExtensionRegistry
used during encoding and decoding.java.lang.String tempLocation
int memoryMB
java.lang.String tempLocation
int memoryMB
long memoryMB
BufferedExternalSorter.Options sorterOptions
private java.lang.Object readResolve() throws java.io.ObjectStreamException
java.io.ObjectStreamException
java.lang.String readSchemaString
java.lang.String fileSchemaString
java.lang.Class<T> type
java.lang.String codec
byte[] syncMarker
UnboundedSource<OutputT,CheckpointMarkT extends UnboundedSource.CheckpointMark> source
long maxNumRecords
Duration maxReadTime
BoundedSource<T> adaptedSource
FileBasedSource<T> sourceDelegate
CompressedSource.DecompressingChannelFactory channelFactory
ValueProvider<T> prefix
java.lang.String shardTemplate
java.lang.String suffix
FileBasedSink.WritableByteChannelFactory writableByteChannelFactory
FileBasedSink.WritableByteChannelFactory
that is used to wrap the raw data output to the
underlying channel. The default is to not compress the output using
FileBasedSink.CompressionType.UNCOMPRESSED
.FileBasedSink.FilenamePolicy filenamePolicy
ValueProvider<T> baseOutputDirectoryProvider
FileBasedSink<T> sink
ValueProvider<T> tempDirectory
boolean windowedWrites
ValueProvider<T> fileOrPatternSpec
MatchResult.Metadata singleFileMetadata
FileBasedSource.Mode mode
long startOffset
long endOffset
long minBundleSize
BoundedSource<T> source
UnboundedSource<OutputT,CheckpointMarkT extends UnboundedSource.CheckpointMark> source
FileBasedSink<T> sink
FileBasedSink.WriteOperation<T> writeOperation
PTransform<InputT extends PInput,OutputT extends POutput> computeNumShards
ValueProvider<T> numShardsProvider
boolean windowedWrites
int maxNumWritersPerBundle
BigQueryIO.Write.CreateDisposition createDisposition
org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices bqServices
DynamicDestinations<T,DestinationT> dynamicDestinations
org.apache.beam.sdk.io.gcp.bigquery.DynamicDestinations.SideInputAccessor sideInputAccessor
DynamicDestinations<T,DestinationT> dynamicDestinations
SerializableFunction<InputT,OutputT> formatFunction
org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices bigQueryServices
BigQueryIO.Write.CreateDisposition createDisposition
DynamicDestinations<T,DestinationT> dynamicDestinations
InsertRetryPolicy retryPolicy
org.apache.beam.sdk.io.gcp.bigquery.BigQueryServices bigQueryServices
InsertRetryPolicy retryPolicy
java.lang.String tableSpec
java.lang.String tableDescription
java.lang.String filename
java.lang.Long fileByteSize
java.lang.Object destination
java.lang.String projectId
java.lang.String subscriptionName
java.lang.String path
org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubSubscription.Type type
java.lang.String project
java.lang.String subscription
org.apache.beam.sdk.io.gcp.pubsub.PubsubIO.PubsubTopic.Type type
java.lang.String project
java.lang.String topic
PubsubClient.PubsubClientFactory pubsubFactory
ValueProvider<T> topic
java.lang.String timestampAttribute
java.lang.String idAttribute
int numShards
int publishBatchSize
int publishBatchBytes
Duration maxLatency
org.apache.beam.sdk.io.gcp.pubsub.PubsubUnboundedSink.RecordIdMethod recordIdMethod
PubsubUnboundedSink.idAttribute
is non-null).com.google.api.client.util.Clock clock
PubsubClient.PubsubClientFactory pubsubFactory
ValueProvider<T> project
PubsubUnboundedSource.topic
was given.ValueProvider<T> topic
PubsubUnboundedSource.subscription
must be given.
Otherwise PubsubUnboundedSource.subscription
must be null.ValueProvider<T> subscription
PubsubUnboundedSource.topic
must be given.
Otherwise PubsubUnboundedSource.topic
must be null.
If no subscription is given a random one will be created when the transorm is applied. This field will be update with that subscription's path. The created subscription is never deleted.
java.lang.String timestampAttribute
java.lang.String idAttribute
boolean needsAttributes
ImmutableList<E> mutations
SpannerIO.Write spec
public void readExternal(java.io.ObjectInput in) throws java.io.IOException, java.lang.ClassNotFoundException
java.io.IOException
java.lang.ClassNotFoundException
public void writeExternal(java.io.ObjectOutput out) throws java.io.IOException
java.io.IOException
java.lang.Class<T> type
SerializableConfiguration conf
Coder<T> keyCoder
Coder<T> valueCoder
SimpleFunction<InputT,OutputT> keyTranslationFunction
SimpleFunction<InputT,OutputT> valueTranslationFunction
HadoopInputFormatIO.SerializableSplit inputSplit
long boundedSourceEstimatedSize
private void readObject(java.io.ObjectInputStream in) throws java.io.IOException, java.lang.ClassNotFoundException
java.io.IOException
java.lang.ClassNotFoundException
private void writeObject(java.io.ObjectOutputStream out) throws java.io.IOException
java.io.IOException
org.apache.hadoop.mapreduce.InputSplit inputSplit
SerializableConfiguration serializableConfiguration
java.lang.String tableId
org.apache.beam.sdk.io.hbase.SerializableScan serializableScan
java.lang.String tableId
SerializableConfiguration serializableConfiguration
JmsIO.Read spec
java.lang.String jmsMessageID
long jmsTimestamp
java.lang.String jmsCorrelationID
javax.jms.Destination jmsReplyTo
javax.jms.Destination jmsDestination
int jmsDeliveryMode
boolean jmsRedelivered
java.lang.String jmsType
long jmsExpiration
int jmsPriority
java.util.Map<K,V> properties
java.lang.String text
java.lang.String topic
int partition
long nextOffset
KafkaIO.Read<K,V> read
MongoDbGridFSIO.Read<T> spec
java.util.List<E> objectIds
com.google.protobuf.ByteString value
java.lang.Class<T> jaxbClass
org.apache.beam.sdk.util.EmptyOnDeserializationThreadLocal<T> jaxbMarshaller
org.apache.beam.sdk.util.EmptyOnDeserializationThreadLocal<T> jaxbUnmarshaller
XmlIO.Read<T> spec
ValueProvider<T> value
SerializableFunction<InputT,OutputT> translator
java.lang.Class<T> klass
java.lang.String methodName
java.lang.String propertyName
java.lang.Object defaultValue
java.lang.Long optionsId
java.lang.Object value
java.lang.String expectedChecksum
java.lang.String actualChecksum
org.apache.beam.sdk.util.ShardedFile shardedFile
SerializableFunction<InputT,OutputT> checkerFn
org.apache.beam.sdk.testing.PAssert.AssertionWindows rewindowingStrategy
SimpleFunction<InputT,OutputT> paneExtractor
PAssert.PAssertionSite site
SerializableFunction<InputT,OutputT> checkerFn
org.apache.beam.sdk.testing.PAssert.AssertionWindows rewindowingStrategy
SimpleFunction<InputT,OutputT> paneExtractor
PAssert.PAssertionSite site
SerializableFunction<InputT,OutputT> checkerFn
PAssert.PAssertionSite site
java.lang.String message
java.lang.StackTraceElement[] creationStackTrace
SerializableMatcher<T> matcher
boolean isSuccess
PAssert.PAssertionSite site
java.lang.String message
java.util.Comparator<T> compareFn
int numQuantiles
int bufferSize
int numBuffers
long maxNumElements
java.util.PriorityQueue<E> heap
long sampleSize
CombineFnBase.GlobalCombineFn<InputT,AccumT,OutputT> fn
DisplayData.ItemSpec<T> fnDisplayData
boolean insertDefault
int fanout
java.util.List<E> sideInputs
CombineFnBase.GlobalCombineFn<InputT,AccumT,OutputT> fn
DisplayData.ItemSpec<T> fnDisplayData
boolean insertDefault
int fanout
CombineFnBase.GlobalCombineFn<InputT,AccumT,OutputT> fn
DisplayData.ItemSpec<T> fnDisplayData
java.util.List<E> sideInputs
SerializableFunction<InputT,OutputT> combiner
int bufferSize
CombineFnBase.GlobalCombineFn<InputT,AccumT,OutputT> fn
DisplayData.ItemSpec<T> fnDisplayData
boolean fewKeys
java.util.List<E> sideInputs
CombineFnBase.GlobalCombineFn<InputT,AccumT,OutputT> fn
DisplayData.ItemSpec<T> fnDisplayData
SerializableFunction<InputT,OutputT> hotKeyFanout
java.util.Map<K,V> valuesMap
java.util.List<E> combineFns
java.util.List<E> extractInputFns
java.util.List<E> outputTags
int combineFnCount
java.util.List<E> extractInputFns
java.util.List<E> combineFnWithContexts
java.util.List<E> outputTags
int combineFnCount
SerializableFunction<InputT,OutputT> fn
TypeDescriptor<T> representativeType
SerializableFunction<InputT,OutputT> predicate
java.lang.String predicateDescription
SimpleFunction<InputT,OutputT> fn
FlatMapElements.into(TypeDescriptor)
, until the fn is specified using FlatMapElements.via(SerializableFunction)
.DisplayData.ItemSpec<T> fnClassDisplayData
boolean fewKeys
long batchSize
SimpleFunction<InputT,OutputT> fn
MapElements.into(TypeDescriptor)
, until the fn is specified using MapElements.via(SerializableFunction)
.DisplayData.ItemSpec<T> fnClassDisplayData
java.util.List<E> sideInputs
TupleTag<V> mainOutputTag
TupleTagList additionalOutputTags
DisplayData.ItemSpec<T> fnDisplayData
DoFn<InputT,OutputT> fn
java.util.List<E> sideInputs
DoFn<InputT,OutputT> fn
DisplayData.ItemSpec<T> fnDisplayData
private void readObject(java.io.ObjectInputStream oos)
private void writeObject(java.io.ObjectOutputStream oos)
java.util.regex.Pattern pattern
java.util.regex.Pattern pattern
int group
java.util.regex.Pattern pattern
java.util.regex.Pattern pattern
int keyGroup
int valueGroup
java.util.regex.Pattern pattern
java.lang.String groupName
java.util.regex.Pattern pattern
java.lang.String keyGroupName
java.lang.String valueGroupName
java.util.regex.Pattern pattern
int group
java.util.regex.Pattern pattern
int keyGroup
int valueGroup
java.util.regex.Pattern pattern
java.lang.String groupName
java.util.regex.Pattern pattern
java.lang.String keyGroupName
java.lang.String valueGroupName
java.util.regex.Pattern pattern
java.lang.String replacement
java.util.regex.Pattern pattern
java.lang.String replacement
java.util.regex.Pattern pattern
boolean outputEmpty
int sampleSize
Top.TopCombineFn<T,ComparatorT extends java.util.Comparator<T> & java.io.Serializable> topCombineFn
java.util.Random rand
SerializableFunction<InputT,OutputT> fn
int count
java.util.Comparator<T> compareFn
java.lang.Object defaultValue
boolean hasDefault
PCollectionView<T> view
SerializableFunction<InputT,OutputT> fn
SerializableFunction<InputT,OutputT> fn
Duration allowedTimestampSkew
ImmutableMap<K,V> entries
ImmutableList<E> components
CoGbkResultSchema schema
UnionCoder unionCoder
TupleTagList tupleTagList
java.util.HashMap<K,V> tagMap
java.util.List<E> elementCoders
long from
long to
int countElems
java.util.List<E> timestampTransforms
Trigger.OnceTrigger earlyTrigger
Trigger.OnceTrigger lateTrigger
int number
DateTime startDate
DateTimeZone timeZone
int number
int dayOfMonth
DateTime startDate
DateTimeZone timeZone
int number
int monthOfYear
int dayOfMonth
DateTime startDate
DateTimeZone timeZone
WindowFn<T,W extends BoundedWindow> givenWindowFn
java.lang.String reason
java.lang.String cause
WindowFn<T,W extends BoundedWindow> originalWindowFn
Trigger repeatedTrigger
Duration gapDuration
java.util.List<E> subTriggers
Window<T> original
WindowingStrategy<T,W extends BoundedWindow> updatedStrategy
Duration maximumLookback
java.lang.Object key
java.lang.Object value
TupleTag<V> tag
WindowMappingFn<TargetWindowT extends BoundedWindow> windowMappingFn
WindowingStrategy<T,W extends BoundedWindow> windowingStrategy
Coder<T> coder
ViewFn<PrimitiveViewT,ViewT> viewFn
ViewFn
for this view.java.lang.String id
boolean generated
java.util.List<E> tupleTags
Coder<T> valueCoder
ByteArrayCoder idCoder
WindowFn<T,W extends BoundedWindow> windowFn
Trigger trigger
WindowingStrategy.AccumulationMode mode
Duration allowedLateness
Window.ClosingBehavior closingBehavior
Window.OnTimeBehavior onTimeBehavior
TimestampCombiner timestampCombiner
boolean triggerSpecified
boolean modeSpecified
boolean allowedLatenessSpecified
boolean timestampCombinerSpecified