dataflow/v1b3 library
Dataflow API - v1b3
Manages Google Cloud Dataflow projects on Google Cloud Platform.
For more information, see cloud.google.com/dataflow
Create an instance of DataflowApi to access these resources:
Classes
- ApproximateProgress
- Obsolete in favor of ApproximateReportedProgress and ApproximateSplitRequest.
- ApproximateReportedProgress
- A progress measurement of a WorkItem by a worker.
- ApproximateSplitRequest
- A suggestion by the service to the worker to dynamically split the WorkItem.
- AutoscalingEvent
- A structured message reporting an autoscaling decision made by the Dataflow service.
- AutoscalingSettings
- Settings for WorkerPool autoscaling.
- BigQueryIODetails
- Metadata for a BigQuery connector used by the job.
- BigTableIODetails
- Metadata for a Cloud Bigtable connector used by the job.
- ComponentSource
- Description of an interstitial value between transforms in an execution stage.
- ComponentTransform
- Description of a transform executed as part of an execution stage.
- ComputationTopology
- All configuration data for a particular Computation.
- ConcatPosition
- A position that encapsulates an inner position and an index for the inner position.
- ContainerSpec
- Container Spec.
- CounterMetadata
- CounterMetadata includes all static non-name non-value counter attributes.
- CounterStructuredName
- Identifies a counter within a per-job namespace.
- CounterStructuredNameAndMetadata
- A single message which encapsulates structured name and metadata for a given counter.
- CounterUpdate
- An update to a Counter sent from a worker.
- CPUTime
- Modeled after information exposed by /proc/stat.
- CreateJobFromTemplateRequest
- A request to create a Cloud Dataflow job from a template.
- CustomSourceLocation
- Identifies the location of a custom souce.
- DataDiskAssignment
- Data disk assignment for a given VM instance.
- DataflowApi
- Manages Google Cloud Dataflow projects on Google Cloud Platform.
- DataSamplingConfig
- Configuration options for sampling elements.
- DataSamplingReport
- Contains per-worker telemetry about the data sampling feature.
- DatastoreIODetails
- Metadata for a Datastore connector used by the job.
- DebugOptions
- Describes any options that have an effect on the debugging of pipelines.
- DerivedSource
- Specification of one of the bundles produced as a result of splitting a Source (e.g. when executing a SourceSplitRequest, or when splitting an active task using WorkItemStatus.dynamic_source_split), relative to the source being split.
- Disk
- Describes the data disk used by a workflow job.
- DisplayData
- Data provided with a pipeline or transform to provide descriptive info.
- DistributionUpdate
- A metric value representing a distribution.
- DynamicSourceSplit
- When a task splits using WorkItemStatus.dynamic_source_split, this message describes the two parts of the split relative to the description of the current task's input.
- Environment
- Describes the environment in which a Dataflow Job runs.
- ExecutionStageState
- A message describing the state of a particular execution stage.
- ExecutionStageSummary
- Description of the composing transforms, names/ids, and input/outputs of a stage of execution.
- FailedLocation
- Indicates which regional endpoint failed to respond to a request for data.
- FileIODetails
- Metadata for a File connector used by the job.
- FlattenInstruction
- An instruction that copies its inputs (zero or more) to its (single) output.
- FlexTemplateRuntimeEnvironment
- The environment values to be set at runtime for flex template.
- FloatingPointList
- A metric value representing a list of floating point numbers.
- FloatingPointMean
- A representation of a floating point mean metric contribution.
- GetDebugConfigRequest
- Request to get updated debug configuration for component.
- GetDebugConfigResponse
- Response to a get debug configuration request.
- GetTemplateResponse
- The response to a GetTemplate request.
- Histogram
- Histogram of value counts for a distribution.
- HotKeyDebuggingInfo
- Information useful for debugging a hot key detection.
- HotKeyDetection
- Proto describing a hot key detected on a given WorkItem.
- HotKeyInfo
- Information about a hot key.
- InstructionInput
- An input of an instruction, as a reference to an output of a producer instruction.
- InstructionOutput
- An output of an instruction.
- IntegerGauge
- A metric value representing temporal values of a variable.
- IntegerList
- A metric value representing a list of integers.
- IntegerMean
- A representation of an integer mean metric contribution.
- Job
- Defines a job to be run by the Cloud Dataflow service.
- JobExecutionDetails
- Information about the execution of a job.
- JobExecutionInfo
- Additional information about how a Cloud Dataflow job will be executed that isn't contained in the submitted job.
- JobExecutionStageInfo
- Contains information about how a particular google.dataflow.v1beta3.Step will be executed.
- JobMessage
- A particular message pertaining to a Dataflow job.
- JobMetadata
- Metadata available primarily for filtering jobs.
- JobMetrics
- JobMetrics contains a collection of metrics describing the detailed progress of a Dataflow job.
- KeyRangeDataDiskAssignment
- Data disk assignment information for a specific key-range of a sharded computation.
- KeyRangeLocation
- Location information for a specific key-range of a sharded computation.
- LaunchFlexTemplateParameter
- Launch FlexTemplate Parameter.
- LaunchFlexTemplateRequest
- A request to launch a Cloud Dataflow job from a FlexTemplate.
- LaunchFlexTemplateResponse
- Response to the request to launch a job from Flex Template.
- LaunchTemplateParameters
- Parameters to provide to the template being launched.
- LaunchTemplateResponse
- Response to the request to launch a template.
- LeaseWorkItemRequest
- Request to lease WorkItems.
- LeaseWorkItemResponse
- Response to a request to lease WorkItems.
- ListJobMessagesResponse
- Response to a request to list job messages.
- ListJobsResponse
- Response to a request to list Cloud Dataflow jobs in a project.
- ListSnapshotsResponse
- List of snapshots.
- MapTask
- MapTask consists of an ordered set of instructions, each of which describes one particular low-level operation for the worker to perform in order to accomplish the MapTask's WorkItem.
- MemInfo
- Information about the memory usage of a worker or a container within a worker.
- MetricShortId
- The metric short id is returned to the user alongside an offset into ReportWorkItemStatusRequest
- MetricStructuredName
- Identifies a metric, by describing the source which generated the metric.
- MetricUpdate
- Describes the state of a metric.
- MountedDataDisk
- Describes mounted data disk.
- MultiOutputInfo
- Information about an output of a multi-output DoFn.
- NameAndKind
- Basic metadata about a counter.
- Package
- The packages that must be installed in order for a worker to run the steps of the Cloud Dataflow job that will be assigned to its worker pool.
- ParallelInstruction
- Describes a particular operation comprising a MapTask.
- Parameter
- Structured data associated with this message.
- ParameterMetadata
- Metadata for a specific parameter.
- ParameterMetadataEnumOption
- ParameterMetadataEnumOption specifies the option shown in the enum form.
- ParDoInstruction
- An instruction that does a ParDo operation.
- PartialGroupByKeyInstruction
- An instruction that does a partial group-by-key.
- PipelineDescription
- A descriptive representation of submitted pipeline as well as the executed form.
- Point
- A point in the timeseries.
- Position
- Position defines a position within a collection of data.
- ProgressTimeseries
- Information about the progress of some component of job execution.
- ProjectsJobsDebugResource
- ProjectsJobsMessagesResource
- ProjectsJobsResource
- ProjectsJobsWorkItemsResource
- ProjectsLocationsFlexTemplatesResource
- ProjectsLocationsJobsDebugResource
- ProjectsLocationsJobsMessagesResource
- ProjectsLocationsJobsResource
- ProjectsLocationsJobsSnapshotsResource
- ProjectsLocationsJobsStagesResource
- ProjectsLocationsJobsWorkItemsResource
- ProjectsLocationsResource
- ProjectsLocationsSnapshotsResource
- ProjectsLocationsTemplatesResource
- ProjectsResource
- ProjectsSnapshotsResource
- ProjectsTemplatesResource
- PubSubIODetails
- Metadata for a Pub/Sub connector used by the job.
- PubsubLocation
- Identifies a pubsub location to use for transferring data into or out of a streaming Dataflow job.
- PubsubSnapshotMetadata
- Represents a Pubsub snapshot.
- ReadInstruction
- An instruction that reads records.
- ReportedParallelism
- Represents the level of parallelism in a WorkItem's input, reported by the worker.
- ReportWorkItemStatusRequest
- Request to report the status of WorkItems.
- ReportWorkItemStatusResponse
- Response from a request to report the status of WorkItems.
- ResourceUtilizationReport
- Worker metrics exported from workers.
- RuntimeEnvironment
- The environment values to set at runtime.
- RuntimeMetadata
- RuntimeMetadata describing a runtime environment.
- RuntimeUpdatableParams
- Additional job parameters that can only be updated during runtime using the projects.jobs.update method.
- SdkBug
- A bug found in the Dataflow SDK.
- SdkHarnessContainerImage
- Defines an SDK harness container for executing Dataflow pipelines.
- SDKInfo
- SDK Information.
- SdkVersion
- The version of the SDK used to run the job.
- SendDebugCaptureRequest
- Request to send encoded debug information.
- SendWorkerMessagesRequest
- A request for sending worker messages to the service.
- SendWorkerMessagesResponse
- The response to the worker messages.
- SeqMapTask
- Describes a particular function to invoke.
- SeqMapTaskOutputInfo
- Information about an output of a SeqMapTask.
- ShellTask
- A task which consists of a shell command for the worker to execute.
- SideInputInfo
- Information about a side input of a DoFn or an input of a SeqDoFn.
- Sink
- A sink that records can be encoded and written to.
- Snapshot
- Represents a snapshot of a job.
- SnapshotJobRequest
- Request to create a snapshot of a job.
- Source
- A source that records can be read and decoded from.
- SourceFork
- DEPRECATED in favor of DynamicSourceSplit.
- SourceGetMetadataRequest
- A request to compute the SourceMetadata of a Source.
- SourceGetMetadataResponse
- The result of a SourceGetMetadataOperation.
- SourceMetadata
- Metadata about a Source useful for automatically optimizing and tuning the pipeline, etc.
- SourceOperationRequest
- A work item that represents the different operations that can be performed on a user-defined Source specification.
- SourceOperationResponse
- The result of a SourceOperationRequest, specified in ReportWorkItemStatusRequest.source_operation when the work item is completed.
- SourceSplitOptions
- Hints for splitting a Source into bundles (parts for parallel processing) using SourceSplitRequest.
- SourceSplitRequest
- Represents the operation to split a high-level Source specification into bundles (parts for parallel processing).
- SourceSplitResponse
- The response to a SourceSplitRequest.
- SourceSplitShard
- DEPRECATED in favor of DerivedSource.
- SpannerIODetails
- Metadata for a Spanner connector used by the job.
- SplitInt64
- A representation of an int64, n, that is immune to precision loss when encoded in JSON.
- StageExecutionDetails
- Information about the workers and work items within a stage.
- StageSource
- Description of an input or output of an execution stage.
- StageSummary
- Information about a particular execution stage of a job.
- StateFamilyConfig
- State family configuration.
- Step
- Defines a particular step within a Cloud Dataflow job.
- Straggler
- Information for a straggler.
- StragglerDebuggingInfo
- Information useful for debugging a straggler.
- StragglerInfo
- Information useful for straggler identification and debugging.
- StragglerSummary
- Summarized straggler identification details.
- StreamingApplianceSnapshotConfig
- Streaming appliance snapshot configuration.
- StreamingComputationConfig
- Configuration information for a single streaming computation.
- StreamingComputationRanges
- Describes full or partial data disk assignment information of the computation ranges.
- StreamingComputationTask
- A task which describes what action should be performed for the specified streaming computation ranges.
- StreamingConfigTask
- A task that carries configuration information for streaming computations.
- StreamingSetupTask
- A task which initializes part of a streaming Dataflow job.
- StreamingSideInputLocation
- Identifies the location of a streaming side input.
- StreamingStageLocation
- Identifies the location of a streaming computation stage, for stage-to-stage communication.
- StreamingStragglerInfo
- Information useful for streaming straggler identification and debugging.
- StreamLocation
- Describes a stream of data, either as input to be processed or as output of a streaming Dataflow job.
- StringList
- A metric value representing a list of strings.
- StructuredMessage
- A rich message format, including a human readable string, a key for identifying the message, and structured data associated with the message for programmatic consumption.
- TaskRunnerSettings
- Taskrunner configuration settings.
- TemplateMetadata
- Metadata describing a template.
- TopologyConfig
- Global topology of the streaming Dataflow job, including all computations and their sharded locations.
- TransformSummary
- Description of the type, names/ids, and input/outputs for a transform.
- WorkerDetails
- Information about a worker
- WorkerHealthReport
- WorkerHealthReport contains information about the health of a worker.
- WorkerHealthReportResponse
- WorkerHealthReportResponse contains information returned to the worker in response to a health ping.
- WorkerLifecycleEvent
- A report of an event in a worker's lifecycle.
- WorkerMessage
- WorkerMessage provides information to the backend about a worker.
- WorkerMessageCode
- A message code is used to report status and error messages to the service.
- WorkerMessageResponse
- A worker_message response allows the server to pass information to the sender.
- WorkerPool
- Describes one particular pool of Cloud Dataflow workers to be instantiated by the Cloud Dataflow service in order to perform the computations required by a job.
- WorkerSettings
- Provides data to pass through to the worker harness.
- WorkerShutdownNotice
- Shutdown notification from workers.
- WorkerThreadScalingReport
- Contains information about the thread scaling information of a worker.
- WorkerThreadScalingReportResponse
- Contains the thread scaling recommendation for a worker from the backend.
- WorkItem
- WorkItem represents basic information about a WorkItem to be executed in the cloud.
- WorkItemDetails
- Information about an individual work item execution.
- WorkItemServiceState
- The Dataflow service's idea of the current state of a WorkItem being processed by a worker.
- WorkItemStatus
- Conveys a worker's progress through the work described by a WorkItem.
- WriteInstruction
- An instruction that writes records.
Typedefs
- DeleteSnapshotResponse = $Empty
- Response from deleting a snapshot.
- ResourceUtilizationReportResponse = $Empty
- Service-side response to WorkerMessage reporting resource utilization.
- SendDebugCaptureResponse = $Empty
- Response to a send capture request.
- Status = $Status
-
The
Status
type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. - WorkerShutdownNoticeResponse = $Empty
- Service-side response to WorkerMessage issuing shutdown notice.
Exceptions / Errors
- ApiRequestError
- Represents a general error reported by the API endpoint.
- DetailedApiRequestError
- Represents a specific error reported by the API endpoint.