| Class | Description |
|---|---|
| ApproximateProgress |
Obsolete in favor of ApproximateReportedProgress and ApproximateSplitRequest.
|
| ApproximateReportedProgress |
A progress measurement of a WorkItem by a worker.
|
| ApproximateSplitRequest |
A suggestion by the service to the worker to dynamically split the WorkItem.
|
| AutoscalingEvent |
A structured message reporting an autoscaling decision made by the Dataflow service.
|
| AutoscalingSettings |
Settings for WorkerPool autoscaling.
|
| Base2Exponent |
Exponential buckets where the growth factor between buckets is `2**(2**-scale)`.
|
| BigQueryIODetails |
Metadata for a BigQuery connector used by the job.
|
| BigTableIODetails |
Metadata for a Cloud Bigtable connector used by the job.
|
| BoundedTrie |
The message type used for encoding metrics of type bounded trie.
|
| BoundedTrieNode |
A single node in a BoundedTrie.
|
| BucketOptions |
`BucketOptions` describes the bucket boundaries used in the histogram.
|
| ComponentSource |
Description of an interstitial value between transforms in an execution stage.
|
| ComponentTransform |
Description of a transform executed as part of an execution stage.
|
| ComputationTopology |
All configuration data for a particular Computation.
|
| ConcatPosition |
A position that encapsulates an inner position and an index for the inner position.
|
| ContainerSpec |
Container Spec.
|
| CounterMetadata |
CounterMetadata includes all static non-name non-value counter attributes.
|
| CounterStructuredName |
Identifies a counter within a per-job namespace.
|
| CounterStructuredNameAndMetadata |
A single message which encapsulates structured name and metadata for a given counter.
|
| CounterUpdate |
An update to a Counter sent from a worker.
|
| CPUTime |
Modeled after information exposed by /proc/stat.
|
| CreateJobFromTemplateRequest |
A request to create a Cloud Dataflow job from a template.
|
| CustomSourceLocation |
Identifies the location of a custom souce.
|
| DataDiskAssignment |
Data disk assignment for a given VM instance.
|
| DataflowGaugeValue |
The gauge value of a metric.
|
| DataflowHistogramValue |
Summary statistics for a population of values.
|
| DataflowPackage |
The packages that must be installed in order for a worker to run the steps of the Cloud Dataflow
job that will be assigned to its worker pool.
|
| DataSamplingConfig |
Configuration options for sampling elements.
|
| DataSamplingReport |
Contains per-worker telemetry about the data sampling feature.
|
| DatastoreIODetails |
Metadata for a Datastore connector used by the job.
|
| DebugOptions |
Describes any options that have an effect on the debugging of pipelines.
|
| DeleteSnapshotResponse |
Response from deleting a snapshot.
|
| DerivedSource |
Specification of one of the bundles produced as a result of splitting a Source (e.g.
|
| Disk |
Describes the data disk used by a workflow job.
|
| DisplayData |
Data provided with a pipeline or transform to provide descriptive info.
|
| DistributionUpdate |
A metric value representing a distribution.
|
| DynamicSourceSplit |
When a task splits using WorkItemStatus.dynamic_source_split, this message describes the two
parts of the split relative to the description of the current task's input.
|
| Environment |
Describes the environment in which a Dataflow Job runs.
|
| ExecutionStageState |
A message describing the state of a particular execution stage.
|
| ExecutionStageSummary |
Description of the composing transforms, names/ids, and input/outputs of a stage of execution.
|
| FailedLocation |
Indicates which [regional endpoint] (https://cloud.google.com/dataflow/docs/concepts/regional-
endpoints) failed to respond to a request for data.
|
| FileIODetails |
Metadata for a File connector used by the job.
|
| FlattenInstruction |
An instruction that copies its inputs (zero or more) to its (single) output.
|
| FlexTemplateRuntimeEnvironment |
The environment values to be set at runtime for flex template.
|
| FloatingPointList |
A metric value representing a list of floating point numbers.
|
| FloatingPointMean |
A representation of a floating point mean metric contribution.
|
| GetDebugConfigRequest |
Request to get updated debug configuration for component.
|
| GetDebugConfigResponse |
Response to a get debug configuration request.
|
| GetTemplateResponse |
The response to a GetTemplate request.
|
| GetWorkerStacktracesRequest |
Request to get worker stacktraces from debug capture.
|
| GetWorkerStacktracesResponse |
Response to get worker stacktraces from debug capture.
|
| GPUUsage |
Information about the GPU usage on the worker.
|
| GPUUtilization |
Utilization details about the GPU.
|
| Histogram |
Histogram of value counts for a distribution.
|
| HotKeyDebuggingInfo |
Information useful for debugging a hot key detection.
|
| HotKeyDetection |
Proto describing a hot key detected on a given WorkItem.
|
| HotKeyInfo |
Information about a hot key.
|
| InstructionInput |
An input of an instruction, as a reference to an output of a producer instruction.
|
| InstructionOutput |
An output of an instruction.
|
| IntegerGauge |
A metric value representing temporal values of a variable.
|
| IntegerList |
A metric value representing a list of integers.
|
| IntegerMean |
A representation of an integer mean metric contribution.
|
| Job |
Defines a job to be run by the Cloud Dataflow service.
|
| JobExecutionDetails |
Information about the execution of a job.
|
| JobExecutionInfo |
Additional information about how a Cloud Dataflow job will be executed that isn't contained in
the submitted job.
|
| JobExecutionStageInfo |
Contains information about how a particular google.dataflow.v1beta3.Step will be executed.
|
| JobMessage |
A particular message pertaining to a Dataflow job.
|
| JobMetadata |
Metadata available primarily for filtering jobs.
|
| JobMetrics |
JobMetrics contains a collection of metrics describing the detailed progress of a Dataflow job.
|
| KeyRangeDataDiskAssignment |
Data disk assignment information for a specific key-range of a sharded computation.
|
| KeyRangeLocation |
Location information for a specific key-range of a sharded computation.
|
| LaunchFlexTemplateParameter |
Launch FlexTemplate Parameter.
|
| LaunchFlexTemplateRequest |
A request to launch a Cloud Dataflow job from a FlexTemplate.
|
| LaunchFlexTemplateResponse |
Response to the request to launch a job from Flex Template.
|
| LaunchTemplateParameters |
Parameters to provide to the template being launched.
|
| LaunchTemplateResponse |
Response to the request to launch a template.
|
| LeaseWorkItemRequest |
Request to lease WorkItems.
|
| LeaseWorkItemResponse |
Response to a request to lease WorkItems.
|
| Linear |
Linear buckets with the following boundaries for indices in 0 to n-1.
|
| ListJobMessagesResponse |
Response to a request to list job messages.
|
| ListJobsResponse |
Response to a request to list Cloud Dataflow jobs in a project.
|
| ListSnapshotsResponse |
List of snapshots.
|
| MapTask |
MapTask consists of an ordered set of instructions, each of which describes one particular low-
level operation for the worker to perform in order to accomplish the MapTask's WorkItem.
|
| MemInfo |
Information about the memory usage of a worker or a container within a worker.
|
| MetricShortId |
The metric short id is returned to the user alongside an offset into ReportWorkItemStatusRequest
|
| MetricStructuredName |
Identifies a metric, by describing the source which generated the metric.
|
| MetricUpdate |
Describes the state of a metric.
|
| MetricValue |
The value of a metric along with its name and labels.
|
| MountedDataDisk |
Describes mounted data disk.
|
| MultiOutputInfo |
Information about an output of a multi-output DoFn.
|
| NameAndKind |
Basic metadata about a counter.
|
| OutlierStats |
Statistics for the underflow and overflow bucket.
|
| ParallelInstruction |
Describes a particular operation comprising a MapTask.
|
| Parameter |
Structured data associated with this message.
|
| ParameterMetadata |
Metadata for a specific parameter.
|
| ParameterMetadataEnumOption |
ParameterMetadataEnumOption specifies the option shown in the enum form.
|
| ParDoInstruction |
An instruction that does a ParDo operation.
|
| PartialGroupByKeyInstruction |
An instruction that does a partial group-by-key.
|
| PerStepNamespaceMetrics |
Metrics for a particular unfused step and namespace.
|
| PerWorkerMetrics |
Per worker metrics.
|
| PipelineDescription |
A descriptive representation of submitted pipeline as well as the executed form.
|
| Point |
A point in the timeseries.
|
| Position |
Position defines a position within a collection of data.
|
| ProgressTimeseries |
Information about the progress of some component of job execution.
|
| PubSubIODetails |
Metadata for a Pub/Sub connector used by the job.
|
| PubsubLocation |
Identifies a pubsub location to use for transferring data into or out of a streaming Dataflow
job.
|
| PubsubSnapshotMetadata |
Represents a Pubsub snapshot.
|
| ReadInstruction |
An instruction that reads records.
|
| ReportedParallelism |
Represents the level of parallelism in a WorkItem's input, reported by the worker.
|
| ReportWorkItemStatusRequest |
Request to report the status of WorkItems.
|
| ReportWorkItemStatusResponse |
Response from a request to report the status of WorkItems.
|
| ResourceUtilizationReport |
Worker metrics exported from workers.
|
| ResourceUtilizationReportResponse |
Service-side response to WorkerMessage reporting resource utilization.
|
| RuntimeEnvironment |
The environment values to set at runtime.
|
| RuntimeMetadata |
RuntimeMetadata describing a runtime environment.
|
| RuntimeUpdatableParams |
Additional job parameters that can only be updated during runtime using the projects.jobs.update
method.
|
| Sdk |
A structured representation of an SDK.
|
| SdkBug |
A bug found in the Dataflow SDK.
|
| SdkHarnessContainerImage |
Defines an SDK harness container for executing Dataflow pipelines.
|
| SDKInfo |
SDK Information.
|
| SdkVersion |
The version of the SDK used to run the job.
|
| SendDebugCaptureRequest |
Request to send encoded debug information.
|
| SendDebugCaptureResponse |
Response to a send capture request.
|
| SendWorkerMessagesRequest |
A request for sending worker messages to the service.
|
| SendWorkerMessagesResponse |
The response to the worker messages.
|
| SeqMapTask |
Describes a particular function to invoke.
|
| SeqMapTaskOutputInfo |
Information about an output of a SeqMapTask.
|
| ServiceResources |
Resources used by the Dataflow Service to run the job.
|
| ShellTask |
A task which consists of a shell command for the worker to execute.
|
| SideInputInfo |
Information about a side input of a DoFn or an input of a SeqDoFn.
|
| Sink |
A sink that records can be encoded and written to.
|
| Snapshot |
Represents a snapshot of a job.
|
| SnapshotJobRequest |
Request to create a snapshot of a job.
|
| Source |
A source that records can be read and decoded from.
|
| SourceFork |
DEPRECATED in favor of DynamicSourceSplit.
|
| SourceGetMetadataRequest |
A request to compute the SourceMetadata of a Source.
|
| SourceGetMetadataResponse |
The result of a SourceGetMetadataOperation.
|
| SourceMetadata |
Metadata about a Source useful for automatically optimizing and tuning the pipeline, etc.
|
| SourceOperationRequest |
A work item that represents the different operations that can be performed on a user-defined
Source specification.
|
| SourceOperationResponse |
The result of a SourceOperationRequest, specified in ReportWorkItemStatusRequest.source_operation
when the work item is completed.
|
| SourceSplitOptions |
Hints for splitting a Source into bundles (parts for parallel processing) using
SourceSplitRequest.
|
| SourceSplitRequest |
Represents the operation to split a high-level Source specification into bundles (parts for
parallel processing).
|
| SourceSplitResponse |
The response to a SourceSplitRequest.
|
| SourceSplitShard |
DEPRECATED in favor of DerivedSource.
|
| SpannerIODetails |
Metadata for a Spanner connector used by the job.
|
| SplitInt64 |
A representation of an int64, n, that is immune to precision loss when encoded in JSON.
|
| Stack |
A structuredstacktrace for a process running on the worker.
|
| StageExecutionDetails |
Information about the workers and work items within a stage.
|
| StageSource |
Description of an input or output of an execution stage.
|
| StageSummary |
Information about a particular execution stage of a job.
|
| StateFamilyConfig |
State family configuration.
|
| Status |
The `Status` type defines a logical error model that is suitable for different programming
environments, including REST APIs and RPC APIs.
|
| Step |
Defines a particular step within a Cloud Dataflow job.
|
| Straggler |
Information for a straggler.
|
| StragglerDebuggingInfo |
Information useful for debugging a straggler.
|
| StragglerInfo |
Information useful for straggler identification and debugging.
|
| StragglerSummary |
Summarized straggler identification details.
|
| StreamingApplianceSnapshotConfig |
Streaming appliance snapshot configuration.
|
| StreamingComputationConfig |
Configuration information for a single streaming computation.
|
| StreamingComputationRanges |
Describes full or partial data disk assignment information of the computation ranges.
|
| StreamingComputationTask |
A task which describes what action should be performed for the specified streaming computation
ranges.
|
| StreamingConfigTask |
A task that carries configuration information for streaming computations.
|
| StreamingOperationalLimits |
Operational limits imposed on streaming jobs by the backend.
|
| StreamingScalingReport |
Contains per-user worker telemetry used in streaming autoscaling.
|
| StreamingScalingReportResponse |
Contains per-user-worker streaming scaling recommendation from the backend.
|
| StreamingSetupTask |
A task which initializes part of a streaming Dataflow job.
|
| StreamingSideInputLocation |
Identifies the location of a streaming side input.
|
| StreamingStageLocation |
Identifies the location of a streaming computation stage, for stage-to-stage communication.
|
| StreamingStragglerInfo |
Information useful for streaming straggler identification and debugging.
|
| StreamLocation |
Describes a stream of data, either as input to be processed or as output of a streaming Dataflow
job.
|
| StringList |
A metric value representing a list of strings.
|
| StructuredMessage |
A rich message format, including a human readable string, a key for identifying the message, and
structured data associated with the message for programmatic consumption.
|
| TaskRunnerSettings |
Taskrunner configuration settings.
|
| TemplateMetadata |
Metadata describing a template.
|
| TopologyConfig |
Global topology of the streaming Dataflow job, including all computations and their sharded
locations.
|
| TransformSummary |
Description of the type, names/ids, and input/outputs for a transform.
|
| WorkerDetails |
Information about a worker
|
| WorkerHealthReport |
WorkerHealthReport contains information about the health of a worker.
|
| WorkerHealthReportResponse |
WorkerHealthReportResponse contains information returned to the worker in response to a health
ping.
|
| WorkerLifecycleEvent |
A report of an event in a worker's lifecycle.
|
| WorkerMessage |
WorkerMessage provides information to the backend about a worker.
|
| WorkerMessageCode |
A message code is used to report status and error messages to the service.
|
| WorkerMessageResponse |
A worker_message response allows the server to pass information to the sender.
|
| WorkerPool |
Describes one particular pool of Cloud Dataflow workers to be instantiated by the Cloud Dataflow
service in order to perform the computations required by a job.
|
| WorkerSettings |
Provides data to pass through to the worker harness.
|
| WorkerShutdownNotice |
Shutdown notification from workers.
|
| WorkerShutdownNoticeResponse |
Service-side response to WorkerMessage issuing shutdown notice.
|
| WorkerThreadScalingReport |
Contains information about the thread scaling information of a worker.
|
| WorkerThreadScalingReportResponse |
Contains the thread scaling recommendation for a worker from the backend.
|
| WorkItem |
WorkItem represents basic information about a WorkItem to be executed in the cloud.
|
| WorkItemDetails |
Information about an individual work item execution.
|
| WorkItemServiceState |
The Dataflow service's idea of the current state of a WorkItem being processed by a worker.
|
| WorkItemStatus |
Conveys a worker's progress through the work described by a WorkItem.
|
| WriteInstruction |
An instruction that writes records.
|
Copyright © 2011–2025 Google. All rights reserved.