public interface DataflowPipelineOptions extends org.apache.beam.sdk.options.PipelineOptions, org.apache.beam.sdk.extensions.gcp.options.GcpOptions, org.apache.beam.sdk.options.ApplicationNameOptions, DataflowPipelineDebugOptions, DataflowPipelineWorkerPoolOptions, org.apache.beam.sdk.io.gcp.bigquery.BigQueryOptions, org.apache.beam.sdk.extensions.gcp.options.GcsOptions, org.apache.beam.sdk.options.StreamingOptions, CloudDebuggerOptions, DataflowWorkerLoggingOptions, DataflowProfilingOptions, org.apache.beam.sdk.io.gcp.pubsub.PubsubOptions
DataflowRunner.| Modifier and Type | Interface and Description |
|---|---|
static class |
DataflowPipelineOptions.FlexResourceSchedulingGoal
Set of available Flexible Resource Scheduling goals.
|
static class |
DataflowPipelineOptions.StagingLocationFactory
Returns a default staging location under
GcpOptions.getGcpTempLocation(). |
DataflowPipelineDebugOptions.DataflowClientFactory, DataflowPipelineDebugOptions.StagerFactoryDataflowPipelineWorkerPoolOptions.AutoscalingAlgorithmType, DataflowPipelineWorkerPoolOptions.WorkerHarnessContainerImageFactoryorg.apache.beam.sdk.extensions.gcp.options.GcsOptions.ExecutorServiceFactory, org.apache.beam.sdk.extensions.gcp.options.GcsOptions.PathValidatorFactoryDataflowWorkerLoggingOptions.Level, DataflowWorkerLoggingOptions.WorkerLogLevelOverridesDataflowProfilingOptions.DataflowProfilingAgentConfigurationorg.apache.beam.sdk.extensions.gcp.options.GcpOptions.DefaultProjectFactory, org.apache.beam.sdk.extensions.gcp.options.GcpOptions.EnableStreamingEngineFactory, org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpTempLocationFactory, org.apache.beam.sdk.extensions.gcp.options.GcpOptions.GcpUserCredentialsFactory| Modifier and Type | Method and Description |
|---|---|
java.lang.String |
getCreateFromSnapshot()
If set, the snapshot from which the job should be created.
|
java.lang.String |
getDataflowWorkerJar() |
DataflowPipelineOptions.FlexResourceSchedulingGoal |
getFlexRSGoal()
This option controls Flexible Resource Scheduling mode.
|
java.util.Map<java.lang.String,java.lang.String> |
getLabels()
Labels that will be applied to the billing records for this job.
|
java.lang.String |
getPipelineUrl()
The URL of the staged portable pipeline.
|
java.lang.String |
getProject() |
java.lang.String |
getRegion()
The Google Compute Engine region for
creating Dataflow jobs.
|
java.lang.String |
getServiceAccount()
Run the job as a specific service account, instead of the default GCE robot.
|
java.lang.String |
getStagingLocation()
GCS path for staging local files, e.g.
|
java.lang.String |
getTemplateLocation()
Where the runner should generate a template file.
|
boolean |
isUpdate()
Whether to update the currently running pipeline with the same name as this one.
|
void |
setCreateFromSnapshot(java.lang.String value) |
void |
setDataflowWorkerJar(java.lang.String dataflowWorkerJar) |
void |
setFlexRSGoal(DataflowPipelineOptions.FlexResourceSchedulingGoal goal) |
void |
setLabels(java.util.Map<java.lang.String,java.lang.String> labels) |
void |
setPipelineUrl(java.lang.String urlString) |
void |
setProject(java.lang.String value) |
void |
setRegion(java.lang.String region) |
void |
setServiceAccount(java.lang.String value) |
void |
setStagingLocation(java.lang.String value) |
void |
setTemplateLocation(java.lang.String value) |
void |
setUpdate(boolean value) |
getApiRootUrl, getDataflowClient, getDataflowEndpoint, getDataflowJobFile, getDumpHeapOnOOM, getNumberOfWorkerHarnessThreads, getOverrideWindmillBinary, getSaveHeapDumpsToGcsPath, getStager, getStagerClass, getTransformNameMapping, getWindmillServiceEndpoint, getWindmillServicePort, setApiRootUrl, setDataflowClient, setDataflowEndpoint, setDataflowJobFile, setDumpHeapOnOOM, setNumberOfWorkerHarnessThreads, setOverrideWindmillBinary, setSaveHeapDumpsToGcsPath, setStager, setStagerClass, setTransformNameMapping, setWindmillServiceEndpoint, setWindmillServicePortaddExperiment, getExperiments, hasExperiment, setExperimentsgetAutoscalingAlgorithm, getDiskSizeGb, getFilesToStage, getMaxNumWorkers, getMinCpuPlatform, getNetwork, getNumWorkers, getSubnetwork, getUsePublicIps, getWorkerDiskType, getWorkerHarnessContainerImage, getWorkerMachineType, getZone, setAutoscalingAlgorithm, setDiskSizeGb, setFilesToStage, setMaxNumWorkers, setMinCpuPlatform, setNetwork, setNumWorkers, setSubnetwork, setUsePublicIps, setWorkerDiskType, setWorkerHarnessContainerImage, setWorkerMachineType, setZonegetHTTPWriteTimeout, getInsertBundleParallelism, getMaxStreamingBatchSize, getMaxStreamingRowsToBatch, getNumStreamingKeys, getTempDatasetId, setHTTPWriteTimeout, setInsertBundleParallelism, setMaxStreamingBatchSize, setMaxStreamingRowsToBatch, setNumStreamingKeys, setTempDatasetIdgetExecutorService, getGcsEndpoint, getGcsPerformanceMetrics, getGcsUploadBufferSizeBytes, getGcsUtil, getPathValidator, getPathValidatorClass, setExecutorService, setGcsEndpoint, setGcsPerformanceMetrics, setGcsUploadBufferSizeBytes, setGcsUtil, setPathValidator, setPathValidatorClassgetDebuggee, getEnableCloudDebugger, getMaxConditionCost, setDebuggee, setEnableCloudDebugger, setMaxConditionCostgetDefaultWorkerLogLevel, getWorkerLogLevelOverrides, getWorkerSystemErrMessageLevel, getWorkerSystemOutMessageLevel, setDefaultWorkerLogLevel, setWorkerLogLevelOverrides, setWorkerSystemErrMessageLevel, setWorkerSystemOutMessageLevelgetProfilingAgentConfiguration, getSaveProfilesToGcs, setProfilingAgentConfiguration, setSaveProfilesToGcsgetPubsubRootUrl, setPubsubRootUrlgetCredentialFactoryClass, getDataflowKmsKey, getGcpCredential, getGcpTempLocation, getZone, isEnableStreamingEngine, setCredentialFactoryClass, setDataflowKmsKey, setEnableStreamingEngine, setGcpCredential, setGcpTempLocation, setZone@Validation.Required @Default.InstanceFactory(value=org.apache.beam.sdk.extensions.gcp.options.GcpOptions.DefaultProjectFactory.class) java.lang.String getProject()
getProject in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptionsvoid setProject(java.lang.String value)
setProject in interface org.apache.beam.sdk.extensions.gcp.options.GcpOptions@Default.InstanceFactory(value=DataflowPipelineOptions.StagingLocationFactory.class) java.lang.String getStagingLocation()
Must be a valid Cloud Storage URL, beginning with the prefix "gs://"
If getStagingLocation() is not set, it will default to GcpOptions.getGcpTempLocation(). GcpOptions.getGcpTempLocation() must be a valid GCS
path.
void setStagingLocation(java.lang.String value)
boolean isUpdate()
void setUpdate(boolean value)
@Hidden @Experimental java.lang.String getCreateFromSnapshot()
void setCreateFromSnapshot(java.lang.String value)
java.lang.String getTemplateLocation()
void setTemplateLocation(java.lang.String value)
@Hidden @Experimental java.lang.String getServiceAccount()
void setServiceAccount(java.lang.String value)
@Hidden @Experimental java.lang.String getRegion()
NOTE: The Cloud Dataflow now also supports the region flag.
void setRegion(java.lang.String region)
java.util.Map<java.lang.String,java.lang.String> getLabels()
void setLabels(java.util.Map<java.lang.String,java.lang.String> labels)
java.lang.String getPipelineUrl()
void setPipelineUrl(java.lang.String urlString)
java.lang.String getDataflowWorkerJar()
void setDataflowWorkerJar(java.lang.String dataflowWorkerJar)
@Default.Enum(value="UNSPECIFIED") DataflowPipelineOptions.FlexResourceSchedulingGoal getFlexRSGoal()
void setFlexRSGoal(DataflowPipelineOptions.FlexResourceSchedulingGoal goal)