All Classes Interface Summary Class Summary Enum Summary Exception Summary
| Class |
Description |
| BatchStatefulParDoOverrides |
PTransformOverrideFactories that expands to correctly implement
stateful ParDo using window-unaware BatchViewOverrides.GroupByKeyAndSortValuesOnly to linearize
processing per key.
|
| BatchStatefulParDoOverrides.BatchStatefulDoFn<K,V,OutputT> |
A key-preserving DoFn that explodes an iterable that has been grouped by key and
window.
|
| CloudObject |
A representation of an arbitrary Java object to be instantiated by Dataflow workers.
|
| CloudObjects |
|
| CloudObjectTranslator<T> |
A translator that takes an object and creates a CloudObject which can be converted back
to the original object.
|
| CoderCloudObjectTranslatorRegistrar |
Coder authors have the ability to automatically have their Coder registered with
the Dataflow Runner by creating a ServiceLoader entry and a concrete implementation of
this interface.
|
| CreateDataflowView<ElemT,ViewT> |
|
| CustomSources |
A helper class for supporting sources defined as Source.
|
| DataflowClient |
Wrapper around the generated Dataflow client to provide common functionality.
|
| DataflowGroupByKey<K,V> |
Specialized implementation of GroupByKey for translating Redistribute transform into
Dataflow service protos.
|
| DataflowGroupByKey.Registrar |
Registers DataflowGroupByKey.DataflowGroupByKeyTranslator.
|
| DataflowJobAlreadyExistsException |
An exception that is thrown if the unique job name constraint of the Dataflow service is broken
because an existing job with the same job name is currently active.
|
| DataflowJobAlreadyUpdatedException |
An exception that is thrown if the existing job has already been updated within the Dataflow
service and is no longer able to be updated.
|
| DataflowJobException |
|
| DataflowPipelineDebugOptions |
Internal.
|
| DataflowPipelineDebugOptions.DataflowClientFactory |
Returns the default Dataflow client built from the passed in PipelineOptions.
|
| DataflowPipelineDebugOptions.StagerFactory |
|
| DataflowPipelineDebugOptions.UnboundedReaderMaxReadTimeFactory |
|
| DataflowPipelineJob |
A DataflowPipelineJob represents a job submitted to Dataflow using DataflowRunner.
|
| DataflowPipelineOptions |
|
| DataflowPipelineOptions.FlexResourceSchedulingGoal |
Set of available Flexible Resource Scheduling goals.
|
| DataflowPipelineOptions.StagingLocationFactory |
Returns a default staging location under GcpOptions.getGcpTempLocation().
|
| DataflowPipelineRegistrar |
Contains the PipelineOptionsRegistrar and PipelineRunnerRegistrar for the DataflowRunner.
|
| DataflowPipelineRegistrar.Options |
|
| DataflowPipelineRegistrar.Runner |
|
| DataflowPipelineTranslator |
|
| DataflowPipelineTranslator.JobSpecification |
The result of a job translation.
|
| DataflowPipelineWorkerPoolOptions |
Options that are used to configure the Dataflow pipeline worker pool.
|
| DataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType |
Type of autoscaling algorithm to use.
|
| DataflowProfilingOptions |
Options for controlling profiling of pipeline execution.
|
| DataflowProfilingOptions.DataflowProfilingAgentConfiguration |
Configuration the for profiling agent.
|
| DataflowRunner |
A PipelineRunner that executes the operations in the pipeline by first translating them
to the Dataflow representation using the DataflowPipelineTranslator and then submitting
them to a Dataflow service for execution.
|
| DataflowRunner.DataflowTransformTranslator |
|
| DataflowRunner.StreamingPCollectionViewWriterFn<T> |
A marker DoFn for writing the contents of a PCollection to a streaming PCollectionView backend implementation.
|
| DataflowRunnerHooks |
An instance of this class can be passed to the DataflowRunner to add user defined hooks
to be invoked at various times during pipeline execution.
|
| DataflowRunnerInfo |
|
| DataflowServiceException |
Signals there was an error retrieving information about a job from the Cloud Dataflow Service.
|
| DataflowStreamingPipelineOptions |
[Internal] Options for configuring StreamingDataflowWorker.
|
| DataflowStreamingPipelineOptions.EnableWindmillServiceDirectPathFactory |
EnableStreamingEngine defaults to false unless one of the experiment is set.
|
| DataflowStreamingPipelineOptions.GlobalConfigRefreshPeriodFactory |
Read global get config request period from system property
'windmill.global_config_refresh_period'.
|
| DataflowStreamingPipelineOptions.HarnessUpdateReportingPeriodFactory |
Read counter reporting period from system property 'windmill.harness_update_reporting_period'.
|
| DataflowStreamingPipelineOptions.LocalWindmillHostportFactory |
Factory for creating local Windmill address.
|
| DataflowStreamingPipelineOptions.MaxStackTraceDepthToReportFactory |
Read 'MaxStackTraceToReport' from system property 'windmill.max_stack_trace_to_report' or
Integer.MAX_VALUE if unspecified.
|
| DataflowStreamingPipelineOptions.PeriodicStatusPageDirectoryFactory |
Read 'PeriodicStatusPageOutputDirector' from system property
'windmill.periodic_status_page_directory' or null if unspecified.
|
| DataflowStreamingPipelineOptions.WindmillServiceStreamingRpcBatchLimitFactory |
Factory for setting value of WindmillServiceStreamingRpcBatchLimit based on environment.
|
| DataflowTemplateJob |
|
| DataflowTransport |
Helpers for cloud communication.
|
| DataflowWorkerHarnessOptions |
Options that are used exclusively within the Dataflow worker harness.
|
| DataflowWorkerLoggingOptions |
Deprecated.
|
| DataflowWorkerLoggingOptions.Level |
The set of log levels that can be used on the Dataflow worker.
|
| DataflowWorkerLoggingOptions.WorkerLogLevelOverrides |
Defines a log level override for a specific class, package, or name.
|
| DefaultCoderCloudObjectTranslatorRegistrar |
|
| DefaultGcpRegionFactory |
Factory for a default value for Google Cloud region according to
https://cloud.google.com/compute/docs/gcloud-compute/#default-properties.
|
| GcsStager |
Utility class for staging files to GCS.
|
| GroupIntoBatchesOverride |
|
| IsmFormat |
An Ism file is a prefix encoded composite key value file broken into shards.
|
| IsmFormat.Footer |
The footer stores the relevant information required to locate the index and bloom filter.
|
| IsmFormat.FooterCoder |
|
| IsmFormat.IsmRecord<V> |
A record containing a composite key and either a value or metadata.
|
| IsmFormat.IsmRecordCoder<V> |
|
| IsmFormat.IsmShard |
A shard descriptor containing shard id, the data block offset, and the index offset for the
given shard.
|
| IsmFormat.IsmShardCoder |
|
| IsmFormat.KeyPrefix |
The prefix used before each key which contains the number of shared and unshared bytes from the
previous key that was read.
|
| IsmFormat.KeyPrefixCoder |
|
| IsmFormat.MetadataKeyCoder<K> |
A coder for metadata key component.
|
| MonitoringUtil |
A helper class for monitoring jobs submitted to the service.
|
| MonitoringUtil.JobMessagesHandler |
An interface that can be used for defining callbacks to receive a list of JobMessages
containing monitoring information.
|
| MonitoringUtil.LoggingHandler |
A handler that logs monitoring messages.
|
| MonitoringUtil.TimeStampComparator |
Comparator for sorting rows in increasing order based on timestamp.
|
| OutputReference |
A representation used by Steps to reference the
output of other Steps.
|
| PackageUtil |
Helper routines for packages.
|
| PackageUtil.StagedFile |
|
| PrimitiveParDoSingleFactory<InputT,OutputT> |
|
| PrimitiveParDoSingleFactory.ParDoSingle<InputT,OutputT> |
A single-output primitive ParDo.
|
| PrimitiveParDoSingleFactory.PayloadTranslator |
|
| PrimitiveParDoSingleFactory.Registrar |
|
| PropertyNames |
Constant property names used by the SDK in CloudWorkflow specifications.
|
| RandomAccessData |
An elastic-sized byte array which allows you to manipulate it as a stream, or access it directly.
|
| RandomAccessData.RandomAccessDataCoder |
A Coder which encodes the valid parts of this stream.
|
| RandomAccessData.UnsignedLexicographicalComparator |
A Comparator that compares two byte arrays lexicographically.
|
| RowCoderCloudObjectTranslator |
Translator for row coders.
|
| SchemaCoderCloudObjectTranslator |
Translator for Schema coders.
|
| Stager |
Interface for staging files needed for running a Dataflow pipeline.
|
| Structs |
A collection of static methods for manipulating datastructure representations transferred via the
Dataflow API.
|
| TestDataflowPipelineOptions |
A set of options used to configure the TestPipeline.
|
| TestDataflowRunner |
|
| TimeUtil |
A helper class for converting between Dataflow API and SDK time representations.
|
| TransformTranslator<TransformT extends org.apache.beam.sdk.transforms.PTransform> |
A TransformTranslator knows how to translate a particular subclass of PTransform
for the Cloud Dataflow service.
|
| TransformTranslator.StepTranslationContext |
|
| TransformTranslator.TranslationContext |
The interface provided to registered callbacks for interacting with the DataflowRunner,
including reading and writing the values of PCollections and side inputs.
|